node               13 arch/alpha/include/asm/topology.h 	int node;
node               18 arch/alpha/include/asm/topology.h 	node = alpha_mv.cpuid_to_nid(cpu);
node               21 arch/alpha/include/asm/topology.h 	BUG_ON(node < 0);
node               24 arch/alpha/include/asm/topology.h 	return node;
node               29 arch/alpha/include/asm/topology.h static const struct cpumask *cpumask_of_node(int node)
node               33 arch/alpha/include/asm/topology.h 	if (node == NUMA_NO_NODE)
node               36 arch/alpha/include/asm/topology.h 	cpumask_clear(&node_to_cpumask_map[node]);
node               39 arch/alpha/include/asm/topology.h 		if (cpu_to_node(cpu) == node)
node               40 arch/alpha/include/asm/topology.h 			cpumask_set_cpu(cpu, node_to_cpumask_map[node]);
node               43 arch/alpha/include/asm/topology.h 	return &node_to_cpumask_map[node];
node              362 arch/alpha/kernel/core_marvel.c marvel_io7_present(gct6_node *node)
node              366 arch/alpha/kernel/core_marvel.c 	if (node->type != GCT_TYPE_HOSE ||
node              367 arch/alpha/kernel/core_marvel.c 	    node->subtype != GCT_SUBTYPE_IO_PORT_MODULE) 
node              370 arch/alpha/kernel/core_marvel.c 	pe = (node->id >> 8) & 0xff;
node               14 arch/alpha/kernel/gct.c gct6_find_nodes(gct6_node *node, gct6_search_struct *search)
node               20 arch/alpha/kernel/gct.c 	if (node->magic != GCT_NODE_MAGIC) {
node               29 arch/alpha/kernel/gct.c 		if (node->type != wanted->type)
node               31 arch/alpha/kernel/gct.c 		if (node->subtype != wanted->subtype)
node               36 arch/alpha/kernel/gct.c 			wanted->callout(node);
node               40 arch/alpha/kernel/gct.c 	if (node->next) 
node               41 arch/alpha/kernel/gct.c 		status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search);
node               44 arch/alpha/kernel/gct.c 	if (node->child) 
node               45 arch/alpha/kernel/gct.c 		status |= gct6_find_nodes(GCT_NODE_PTR(node->child), search);
node              308 arch/alpha/kernel/pci.c 	list_for_each_entry(child_bus, &b->children, node)
node              317 arch/alpha/kernel/pci.c 	list_for_each_entry(b, &pci_root_buses, node)
node              108 arch/arc/plat-hsdk/platform.c 	int node, ret;
node              111 arch/arc/plat-hsdk/platform.c 	node = fdt_path_offset(fdt, path);
node              112 arch/arc/plat-hsdk/platform.c 	if (node < 0)
node              115 arch/arc/plat-hsdk/platform.c 	prop = fdt_getprop(fdt, node, "dma-coherent", &ret);
node              124 arch/arc/plat-hsdk/platform.c 		ret = fdt_delprop(fdt, node, "dma-coherent");
node              128 arch/arc/plat-hsdk/platform.c 		ret = fdt_setprop(fdt, node, "dma-coherent", NULL, 0);
node               46 arch/arm/common/dmabounce.c 	struct list_head node;
node              158 arch/arm/common/dmabounce.c 	list_add(&buf->node, &device_info->safe_buffers);
node              173 arch/arm/common/dmabounce.c 	list_for_each_entry(b, &device_info->safe_buffers, node)
node              193 arch/arm/common/dmabounce.c 	list_del(&buf->node);
node               42 arch/arm/include/asm/mach/pci.h 	struct list_head node;
node               11 arch/arm/include/asm/traps.h 	struct list_head node;
node               63 arch/arm/kernel/bios32.c 	list_for_each_entry(bus, &pci_root_buses, node)
node              516 arch/arm/kernel/bios32.c 			list_add(&sys->node, head);
node              537 arch/arm/kernel/bios32.c 	list_for_each_entry(sys, &head, node) {
node              553 arch/arm/kernel/bios32.c 			list_for_each_entry(child, &bus->children, node)
node               35 arch/arm/kernel/devtree.c static int __init set_smp_ops_by_method(struct device_node *node)
node               40 arch/arm/kernel/devtree.c 	if (of_property_read_string(node, "enable-method", &method))
node               52 arch/arm/kernel/devtree.c static inline int set_smp_ops_by_method(struct device_node *node)
node              406 arch/arm/kernel/traps.c 	list_add(&hook->node, &undef_hook);
node              415 arch/arm/kernel/traps.c 	list_del(&hook->node);
node              427 arch/arm/kernel/traps.c 	list_for_each_entry(hook, &undef_hook, node)
node               97 arch/arm/mach-actions/platsmp.c 	struct device_node *node;
node               99 arch/arm/mach-actions/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "actions,s500-timer");
node              100 arch/arm/mach-actions/platsmp.c 	if (!node) {
node              105 arch/arm/mach-actions/platsmp.c 	timer_base_addr = of_iomap(node, 0);
node              111 arch/arm/mach-actions/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "actions,s500-sps");
node              112 arch/arm/mach-actions/platsmp.c 	if (!node) {
node              117 arch/arm/mach-actions/platsmp.c 	sps_base_addr = of_iomap(node, 0);
node              124 arch/arm/mach-actions/platsmp.c 		node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu");
node              125 arch/arm/mach-actions/platsmp.c 		if (!node) {
node              130 arch/arm/mach-actions/platsmp.c 		scu_base_addr = of_iomap(node, 0);
node              576 arch/arm/mach-at91/pm.c 	struct device_node *node;
node              579 arch/arm/mach-at91/pm.c 	for_each_compatible_node(node, NULL, "mmio-sram") {
node              580 arch/arm/mach-at91/pm.c 		pdev = of_find_device_by_node(node);
node              582 arch/arm/mach-at91/pm.c 			of_node_put(node);
node               45 arch/arm/mach-bcm/bcm_kona_smc.c 	struct device_node *node;
node               52 arch/arm/mach-bcm/bcm_kona_smc.c 	node = of_find_matching_node(NULL, bcm_kona_smc_ids);
node               53 arch/arm/mach-bcm/bcm_kona_smc.c 	if (!node)
node               56 arch/arm/mach-bcm/bcm_kona_smc.c 	prop_val = of_get_address(node, 0, &prop_size, NULL);
node               41 arch/arm/mach-exynos/exynos.c 	struct device_node *node;
node               43 arch/arm/mach-exynos/exynos.c 	for_each_compatible_node(node, NULL, "samsung,exynos4210-sysram") {
node               44 arch/arm/mach-exynos/exynos.c 		if (!of_device_is_available(node))
node               46 arch/arm/mach-exynos/exynos.c 		sysram_base_addr = of_iomap(node, 0);
node               47 arch/arm/mach-exynos/exynos.c 		sysram_base_phys = of_translate_address(node,
node               48 arch/arm/mach-exynos/exynos.c 					   of_get_address(node, 0, NULL, NULL));
node               52 arch/arm/mach-exynos/exynos.c 	for_each_compatible_node(node, NULL, "samsung,exynos4210-sysram-ns") {
node               53 arch/arm/mach-exynos/exynos.c 		if (!of_device_is_available(node))
node               55 arch/arm/mach-exynos/exynos.c 		sysram_ns_base_addr = of_iomap(node, 0);
node               60 arch/arm/mach-exynos/exynos.c static int __init exynos_fdt_map_chipid(unsigned long node, const char *uname,
node               67 arch/arm/mach-exynos/exynos.c 	if (!of_flat_dt_is_compatible(node, "samsung,exynos4210-chipid"))
node               70 arch/arm/mach-exynos/exynos.c 	reg = of_get_flat_dt_prop(node, "reg", &len);
node              237 arch/arm/mach-exynos/mcpm-exynos.c 	struct device_node *node;
node              241 arch/arm/mach-exynos/mcpm-exynos.c 	node = of_find_matching_node(NULL, exynos_dt_mcpm_match);
node              242 arch/arm/mach-exynos/mcpm-exynos.c 	if (!node)
node              244 arch/arm/mach-exynos/mcpm-exynos.c 	of_node_put(node);
node              249 arch/arm/mach-exynos/mcpm-exynos.c 	node = of_find_compatible_node(NULL, NULL,
node              251 arch/arm/mach-exynos/mcpm-exynos.c 	if (!node)
node              254 arch/arm/mach-exynos/mcpm-exynos.c 	ns_sram_base_addr = of_iomap(node, 0);
node              255 arch/arm/mach-exynos/mcpm-exynos.c 	of_node_put(node);
node              189 arch/arm/mach-exynos/suspend.c static int __init exynos_pmu_irq_init(struct device_node *node,
node              195 arch/arm/mach-exynos/suspend.c 		pr_err("%pOF: no parent, giving up\n", node);
node              201 arch/arm/mach-exynos/suspend.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              205 arch/arm/mach-exynos/suspend.c 	pmu_base_addr = of_iomap(node, 0);
node              208 arch/arm/mach-exynos/suspend.c 		pr_err("%pOF: failed to find exynos pmu register\n", node);
node              213 arch/arm/mach-exynos/suspend.c 					  node, &exynos_pmu_domain_ops,
node              225 arch/arm/mach-exynos/suspend.c 	of_node_clear_flag(node, OF_POPULATED);
node              145 arch/arm/mach-hisi/hotplug.c 	struct device_node *node;
node              147 arch/arm/mach-hisi/hotplug.c 	node = of_find_compatible_node(NULL, NULL, "hisilicon,sysctrl");
node              148 arch/arm/mach-hisi/hotplug.c 	if (!node) {
node              153 arch/arm/mach-hisi/hotplug.c 	ctrl_base = of_iomap(node, 0);
node              154 arch/arm/mach-hisi/hotplug.c 	of_node_put(node);
node              153 arch/arm/mach-hisi/platsmp.c 	struct device_node *node;
node              159 arch/arm/mach-hisi/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "hisilicon,hip01-sysctrl");
node              160 arch/arm/mach-hisi/platsmp.c 	if (WARN_ON(!node))
node              162 arch/arm/mach-hisi/platsmp.c 	ctrl_base = of_iomap(node, 0);
node              228 arch/arm/mach-imx/gpc.c static int __init imx_gpc_init(struct device_node *node,
node              235 arch/arm/mach-imx/gpc.c 		pr_err("%pOF: no parent, giving up\n", node);
node              241 arch/arm/mach-imx/gpc.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              245 arch/arm/mach-imx/gpc.c 	gpc_base = of_iomap(node, 0);
node              250 arch/arm/mach-imx/gpc.c 					  node, &imx_gpc_domain_ops,
node              265 arch/arm/mach-imx/gpc.c 	of_node_clear_flag(node, OF_POPULATED);
node              104 arch/arm/mach-imx/mmdc.c 	struct hlist_node node;
node              214 arch/arm/mach-imx/mmdc.c static int mmdc_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              216 arch/arm/mach-imx/mmdc.c 	struct mmdc_pmu *pmu_mmdc = hlist_entry_safe(node, struct mmdc_pmu, node);
node              463 arch/arm/mach-imx/mmdc.c 	cpuhp_state_remove_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node);
node              512 arch/arm/mach-imx/mmdc.c 	cpuhp_state_add_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node);
node              523 arch/arm/mach-imx/mmdc.c 	cpuhp_state_remove_instance_nocalls(cpuhp_mmdc_state, &pmu_mmdc->node);
node              272 arch/arm/mach-imx/pm-imx5.c 	struct device_node *node;
node              281 arch/arm/mach-imx/pm-imx5.c 	node = of_find_compatible_node(NULL, NULL, "mmio-sram");
node              282 arch/arm/mach-imx/pm-imx5.c 	if (!node) {
node              287 arch/arm/mach-imx/pm-imx5.c 	pdev = of_find_device_by_node(node);
node              316 arch/arm/mach-imx/pm-imx5.c 	of_node_put(node);
node              439 arch/arm/mach-imx/pm-imx6.c 	struct device_node *node;
node              443 arch/arm/mach-imx/pm-imx6.c 	node = of_find_compatible_node(NULL, NULL, compat);
node              444 arch/arm/mach-imx/pm-imx6.c 	if (!node)
node              447 arch/arm/mach-imx/pm-imx6.c 	ret = of_address_to_resource(node, 0, &res);
node              457 arch/arm/mach-imx/pm-imx6.c 	of_node_put(node);
node              464 arch/arm/mach-imx/pm-imx6.c 	struct device_node *node;
node              479 arch/arm/mach-imx/pm-imx6.c 	node = of_find_compatible_node(NULL, NULL, "mmio-sram");
node              480 arch/arm/mach-imx/pm-imx6.c 	if (!node) {
node              485 arch/arm/mach-imx/pm-imx6.c 	pdev = of_find_device_by_node(node);
node              584 arch/arm/mach-imx/pm-imx6.c 	of_node_put(node);
node               64 arch/arm/mach-meson/platsmp.c 	static struct device_node *node;
node               67 arch/arm/mach-meson/platsmp.c 	node = of_find_compatible_node(NULL, NULL, sram_compatible);
node               68 arch/arm/mach-meson/platsmp.c 	if (!node) {
node               73 arch/arm/mach-meson/platsmp.c 	sram_base = of_iomap(node, 0);
node               87 arch/arm/mach-meson/platsmp.c 	node = of_find_compatible_node(NULL, NULL, scu_compatible);
node               88 arch/arm/mach-meson/platsmp.c 	if (!node) {
node               93 arch/arm/mach-meson/platsmp.c 	scu_base = of_iomap(node, 0);
node               33 arch/arm/mach-mmp/sram.c 	struct list_head node;
node               48 arch/arm/mach-mmp/sram.c 	list_for_each_entry(info, &sram_bank_list, node)
node               54 arch/arm/mach-mmp/sram.c 	if (&info->node == &sram_bank_list)
node              107 arch/arm/mach-mmp/sram.c 	list_add(&info->node, &sram_bank_list);
node              134 arch/arm/mach-mmp/sram.c 	list_del(&info->node);
node               69 arch/arm/mach-mvebu/board-v7.c static int __init mvebu_scan_mem(unsigned long node, const char *uname,
node               72 arch/arm/mach-mvebu/board-v7.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node               79 arch/arm/mach-mvebu/board-v7.c 	reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l);
node               81 arch/arm/mach-mvebu/board-v7.c 		reg = of_get_flat_dt_prop(node, "reg", &l);
node              117 arch/arm/mach-mvebu/platsmp.c 	struct device_node *node;
node              136 arch/arm/mach-mvebu/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "marvell,bootrom");
node              137 arch/arm/mach-mvebu/platsmp.c 	if (!node)
node              140 arch/arm/mach-mvebu/platsmp.c 	err = of_address_to_resource(node, 0, &res);
node              141 arch/arm/mach-mvebu/platsmp.c 	of_node_put(node);
node              800 arch/arm/mach-omap1/clock.c 	if (clk->node.next || clk->node.prev)
node              809 arch/arm/mach-omap1/clock.c 	list_add(&clk->node, &clocks);
node              825 arch/arm/mach-omap1/clock.c 	list_del(&clk->node);
node              834 arch/arm/mach-omap1/clock.c 	list_for_each_entry(clkp, &clocks, node)
node              854 arch/arm/mach-omap1/clock.c 	list_for_each_entry(c, &clocks, node) {
node              873 arch/arm/mach-omap1/clock.c 	list_for_each_entry(c, &clocks, node)
node              889 arch/arm/mach-omap1/clock.c 	list_for_each_entry(c, &clocks, node)
node              941 arch/arm/mach-omap1/clock.c 	list_for_each_entry(ck, &clocks, node) {
node              977 arch/arm/mach-omap1/clock.c 	list_for_each_entry(c, &clocks, node) {
node             1022 arch/arm/mach-omap1/clock.c 	list_for_each_entry(c, &clocks, node)
node              139 arch/arm/mach-omap1/clock.h 	struct list_head	node;
node               55 arch/arm/mach-omap2/clockdomain.c 	list_for_each_entry(temp_clkdm, &clkdm_list, node) {
node               92 arch/arm/mach-omap2/clockdomain.c 	list_add(&clkdm->node, &clkdm_list);
node              485 arch/arm/mach-omap2/clockdomain.c 	list_for_each_entry(clkdm, &clkdm_list, node) {
node              520 arch/arm/mach-omap2/clockdomain.c 	list_for_each_entry(temp_clkdm, &clkdm_list, node) {
node              553 arch/arm/mach-omap2/clockdomain.c 	list_for_each_entry(clkdm, &clkdm_list, node) {
node              140 arch/arm/mach-omap2/clockdomain.h 	struct list_head node;
node              172 arch/arm/mach-omap2/display.c 	struct device_node *node;
node              211 arch/arm/mach-omap2/display.c 	node = of_find_node_by_name(NULL, "omap4_padconf_global");
node              212 arch/arm/mach-omap2/display.c 	if (node)
node              213 arch/arm/mach-omap2/display.c 		omap4_dsi_mux_syscon = syscon_node_to_regmap(node);
node              228 arch/arm/mach-omap2/display.c 	struct device_node *node;
node              232 arch/arm/mach-omap2/display.c 		node = of_find_compatible_node(NULL, NULL,
node              234 arch/arm/mach-omap2/display.c 		if (node)
node              235 arch/arm/mach-omap2/display.c 			return node;
node              244 arch/arm/mach-omap2/display.c 	struct device_node *node;
node              249 arch/arm/mach-omap2/display.c 	node = omapdss_find_dss_of_node();
node              250 arch/arm/mach-omap2/display.c 	if (!node)
node              253 arch/arm/mach-omap2/display.c 	if (!of_device_is_available(node)) {
node              254 arch/arm/mach-omap2/display.c 		of_node_put(node);
node              258 arch/arm/mach-omap2/display.c 	pdev = of_find_device_by_node(node);
node              265 arch/arm/mach-omap2/display.c 	r = of_platform_populate(node, NULL, NULL, &pdev->dev);
node              545 arch/arm/mach-omap2/omap-wakeupgen.c static int __init wakeupgen_init(struct device_node *node,
node              554 arch/arm/mach-omap2/omap-wakeupgen.c 		pr_err("%pOF: no parent, giving up\n", node);
node              560 arch/arm/mach-omap2/omap-wakeupgen.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              570 arch/arm/mach-omap2/omap-wakeupgen.c 	wakeupgen_base = of_iomap(node, 0);
node              588 arch/arm/mach-omap2/omap-wakeupgen.c 					  node, &wakeupgen_domain_ops,
node              137 arch/arm/mach-omap2/omap_device.c 	struct device_node *node = pdev->dev.of_node;
node              143 arch/arm/mach-omap2/omap_device.c 	oh_cnt = of_property_count_strings(node, "ti,hwmods");
node              150 arch/arm/mach-omap2/omap_device.c 	ret = of_property_read_string_index(node, "ti,hwmods", 0, &oh_name);
node              157 arch/arm/mach-omap2/omap_device.c 	    !omap_hwmod_parse_module_range(NULL, node, &res))
node              167 arch/arm/mach-omap2/omap_device.c 		of_property_read_string_index(node, "ti,hwmods", i, &oh_name);
node              201 arch/arm/mach-omap2/omap_hwmod.c 	struct device_node	*node;
node              748 arch/arm/mach-omap2/omap_hwmod.c 	provider->node = np;
node              825 arch/arm/mach-omap2/omap_hwmod.c 				clkspec.np = provider->node;
node              834 arch/arm/mach-omap2/omap_hwmod.c 					 clkspec.args[0], provider->node);
node              907 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(os, &oh->slave_ports, node) {
node             1019 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(os, &oh->slave_ports, node) {
node             1074 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(os, &oh->slave_ports, node) {
node             1159 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(os, &oh->slave_ports, node) {
node             1388 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(temp_oh, &omap_hwmod_list, node) {
node             2453 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(os, &oh->slave_ports, node) {
node             2653 arch/arm/mach-omap2/omap_hwmod.c 	list_add_tail(&oh->node, &omap_hwmod_list);
node             2686 arch/arm/mach-omap2/omap_hwmod.c 	list_add(&oi->node, &oi->slave->slave_ports);
node             3058 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(temp_oh, &omap_hwmod_list, node) {
node             4014 arch/arm/mach-omap2/omap_hwmod.c 	list_for_each_entry(temp_oh, &omap_hwmod_list, node) {
node              245 arch/arm/mach-omap2/omap_hwmod.h 	struct list_head		node;
node              597 arch/arm/mach-omap2/omap_hwmod.h 	struct list_head		node;
node             2822 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c 	struct device_node *node;
node             2828 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c 	node = of_get_child_by_name(bus, dev_name);
node             2829 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c 	available = of_device_is_available(node);
node             2830 arch/arm/mach-omap2/omap_hwmod_3xxx_data.c 	of_node_put(node);
node               61 arch/arm/mach-omap2/pm34xx.c 	struct list_head node;
node              314 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node)
node              317 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              330 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              367 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              385 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              396 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              417 arch/arm/mach-omap2/pm34xx.c 	list_add(&pwrst->node, &pwrst_list);
node              561 arch/arm/mach-omap2/pm34xx.c 	list_for_each_entry_safe(pwrst, tmp, &pwrst_list, node) {
node              562 arch/arm/mach-omap2/pm34xx.c 		list_del(&pwrst->node);
node               34 arch/arm/mach-omap2/pm44xx.c 	struct list_head node;
node               59 arch/arm/mach-omap2/pm44xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node               65 arch/arm/mach-omap2/pm44xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node               82 arch/arm/mach-omap2/pm44xx.c 	list_for_each_entry(pwrst, &pwrst_list, node) {
node              154 arch/arm/mach-omap2/pm44xx.c 	list_add(&pwrst->node, &pwrst_list);
node               71 arch/arm/mach-omap2/powerdomain.c 	list_for_each_entry(temp_pwrdm, &pwrdm_list, node) {
node              122 arch/arm/mach-omap2/powerdomain.c 	list_add(&pwrdm->node, &pwrdm_list);
node              372 arch/arm/mach-omap2/powerdomain.c 	list_for_each_entry(temp_p, &pwrdm_list, node)
node              446 arch/arm/mach-omap2/powerdomain.c 	list_for_each_entry(temp_pwrdm, &pwrdm_list, node) {
node              124 arch/arm/mach-omap2/powerdomain.h 	struct list_head node;
node              246 arch/arm/mach-omap2/voltage.c 	list_for_each_entry(voltdm, &voltdm_list, node) {
node              281 arch/arm/mach-omap2/voltage.c 	list_for_each_entry(temp_voltdm, &voltdm_list, node) {
node              296 arch/arm/mach-omap2/voltage.c 	list_add(&voltdm->node, &voltdm_list);
node               65 arch/arm/mach-omap2/voltage.h 	struct list_head node;
node              143 arch/arm/mach-pxa/irq.c pxa_init_irq_common(struct device_node *node, int irq_nr,
node              149 arch/arm/mach-pxa/irq.c 	pxa_irq_domain = irq_domain_add_legacy(node, irq_nr,
node              235 arch/arm/mach-pxa/irq.c 	struct device_node *node;
node              239 arch/arm/mach-pxa/irq.c 	node = of_find_matching_node(NULL, intc_ids);
node              240 arch/arm/mach-pxa/irq.c 	if (!node) {
node              245 arch/arm/mach-pxa/irq.c 	ret = of_property_read_u32(node, "marvell,intc-nr-irqs",
node              252 arch/arm/mach-pxa/irq.c 	ret = of_address_to_resource(node, 0, &res);
node              259 arch/arm/mach-pxa/irq.c 	if (of_find_property(node, "marvell,intc-priority", NULL))
node              268 arch/arm/mach-pxa/irq.c 	pxa_init_irq_common(node, pxa_internal_irq_nr, fn);
node              156 arch/arm/mach-pxa/pxa25x.c pxa25x_dt_init_irq(struct device_node *node, struct device_node *parent)
node              238 arch/arm/mach-pxa/pxa27x.c pxa27x_dt_init_irq(struct device_node *node, struct device_node *parent)
node              361 arch/arm/mach-pxa/pxa3xx.c pxa3xx_dt_init_irq(struct device_node *node, struct device_node *parent)
node               55 arch/arm/mach-qcom/platsmp.c 	struct device_node *node;
node               58 arch/arm/mach-qcom/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "qcom,gcc-msm8660");
node               59 arch/arm/mach-qcom/platsmp.c 	if (!node) {
node               64 arch/arm/mach-qcom/platsmp.c 	base = of_iomap(node, 0);
node               65 arch/arm/mach-qcom/platsmp.c 	of_node_put(node);
node              166 arch/arm/mach-rockchip/platsmp.c static int __init rockchip_smp_prepare_sram(struct device_node *node)
node              174 arch/arm/mach-rockchip/platsmp.c 	ret = of_address_to_resource(node, 0, &res);
node              177 arch/arm/mach-rockchip/platsmp.c 		       __func__, node);
node              210 arch/arm/mach-rockchip/platsmp.c 	struct device_node *node;
node              219 arch/arm/mach-rockchip/platsmp.c 	node = of_find_node_by_path("/cpus");
node              221 arch/arm/mach-rockchip/platsmp.c 	pmu = syscon_regmap_lookup_by_phandle(node, "rockchip,pmu");
node              222 arch/arm/mach-rockchip/platsmp.c 	of_node_put(node);
node              232 arch/arm/mach-rockchip/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-pmu");
node              233 arch/arm/mach-rockchip/platsmp.c 	if (!node) {
node              238 arch/arm/mach-rockchip/platsmp.c 	pmu_base = of_iomap(node, 0);
node              239 arch/arm/mach-rockchip/platsmp.c 	of_node_put(node);
node              260 arch/arm/mach-rockchip/platsmp.c 	struct device_node *node;
node              263 arch/arm/mach-rockchip/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-smp-sram");
node              264 arch/arm/mach-rockchip/platsmp.c 	if (!node) {
node              269 arch/arm/mach-rockchip/platsmp.c 	sram_base_addr = of_iomap(node, 0);
node              272 arch/arm/mach-rockchip/platsmp.c 		of_node_put(node);
node              277 arch/arm/mach-rockchip/platsmp.c 		of_node_put(node);
node              282 arch/arm/mach-rockchip/platsmp.c 		if (rockchip_smp_prepare_sram(node)) {
node              283 arch/arm/mach-rockchip/platsmp.c 			of_node_put(node);
node              290 arch/arm/mach-rockchip/platsmp.c 		of_node_put(node);
node              291 arch/arm/mach-rockchip/platsmp.c 		node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu");
node              292 arch/arm/mach-rockchip/platsmp.c 		if (!node) {
node              297 arch/arm/mach-rockchip/platsmp.c 		scu_base_addr = of_iomap(node, 0);
node              300 arch/arm/mach-rockchip/platsmp.c 			of_node_put(node);
node              319 arch/arm/mach-rockchip/platsmp.c 	of_node_put(node);
node               21 arch/arm/mach-s5pv210/s5pv210.c static int __init s5pv210_fdt_map_sys(unsigned long node, const char *uname,
node               28 arch/arm/mach-s5pv210/s5pv210.c 	if (!of_flat_dt_is_compatible(node, "samsung,s5pv210-clock"))
node               31 arch/arm/mach-s5pv210/s5pv210.c 	reg = of_get_flat_dt_prop(node, "reg", &len);
node              131 arch/arm/mach-shmobile/setup-rcar-gen2.c static int __init rcar_gen2_scan_mem(unsigned long node, const char *uname,
node              134 arch/arm/mach-shmobile/setup-rcar-gen2.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              144 arch/arm/mach-shmobile/setup-rcar-gen2.c 	reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l);
node              146 arch/arm/mach-shmobile/setup-rcar-gen2.c 		reg = of_get_flat_dt_prop(node, "reg", &l);
node               90 arch/arm/mach-sunxi/mc_smp.c 	struct device_node *node;
node               94 arch/arm/mach-sunxi/mc_smp.c 	node = of_cpu_device_node_get(cpu);
node               97 arch/arm/mach-sunxi/mc_smp.c 	if (!node)
node               98 arch/arm/mach-sunxi/mc_smp.c 		node = of_get_cpu_node(cpu, NULL);
node              100 arch/arm/mach-sunxi/mc_smp.c 	if (!node) {
node              111 arch/arm/mach-sunxi/mc_smp.c 	is_compatible = of_device_is_compatible(node, "arm,cortex-a15");
node              112 arch/arm/mach-sunxi/mc_smp.c 	of_node_put(node);
node              783 arch/arm/mach-sunxi/mc_smp.c 	struct device_node *node;
node              792 arch/arm/mach-sunxi/mc_smp.c 	node = of_cpu_device_node_get(0);
node              793 arch/arm/mach-sunxi/mc_smp.c 	if (!node)
node              805 arch/arm/mach-sunxi/mc_smp.c 		ret = of_property_match_string(node, "enable-method",
node              813 arch/arm/mach-sunxi/mc_smp.c 	of_node_put(node);
node               44 arch/arm/mach-sunxi/platsmp.c 	struct device_node *node;
node               46 arch/arm/mach-sunxi/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "allwinner,sun6i-a31-prcm");
node               47 arch/arm/mach-sunxi/platsmp.c 	if (!node) {
node               52 arch/arm/mach-sunxi/platsmp.c 	prcm_membase = of_iomap(node, 0);
node               53 arch/arm/mach-sunxi/platsmp.c 	of_node_put(node);
node               59 arch/arm/mach-sunxi/platsmp.c 	node = of_find_compatible_node(NULL, NULL,
node               61 arch/arm/mach-sunxi/platsmp.c 	if (!node) {
node               66 arch/arm/mach-sunxi/platsmp.c 	cpucfg_membase = of_iomap(node, 0);
node               67 arch/arm/mach-sunxi/platsmp.c 	of_node_put(node);
node              129 arch/arm/mach-sunxi/platsmp.c 	struct device_node *node;
node              131 arch/arm/mach-sunxi/platsmp.c 	node = of_find_compatible_node(NULL, NULL, "allwinner,sun8i-a23-prcm");
node              132 arch/arm/mach-sunxi/platsmp.c 	if (!node) {
node              137 arch/arm/mach-sunxi/platsmp.c 	prcm_membase = of_iomap(node, 0);
node              138 arch/arm/mach-sunxi/platsmp.c 	of_node_put(node);
node              144 arch/arm/mach-sunxi/platsmp.c 	node = of_find_compatible_node(NULL, NULL,
node              146 arch/arm/mach-sunxi/platsmp.c 	if (!node) {
node              151 arch/arm/mach-sunxi/platsmp.c 	cpucfg_membase = of_iomap(node, 0);
node              152 arch/arm/mach-sunxi/platsmp.c 	of_node_put(node);
node              135 arch/arm/mach-vexpress/dcscb.c 	struct device_node *node;
node              142 arch/arm/mach-vexpress/dcscb.c 	node = of_find_compatible_node(NULL, NULL, "arm,rtsm,dcscb");
node              143 arch/arm/mach-vexpress/dcscb.c 	if (!node)
node              145 arch/arm/mach-vexpress/dcscb.c 	dcscb_base = of_iomap(node, 0);
node              368 arch/arm/mm/cache-feroceon-l2.c 	struct device_node *node;
node              376 arch/arm/mm/cache-feroceon-l2.c 	node = of_find_matching_node(NULL, feroceon_ids);
node              377 arch/arm/mm/cache-feroceon-l2.c 	if (node && of_device_is_compatible(node, "marvell,kirkwood-cache")) {
node              378 arch/arm/mm/cache-feroceon-l2.c 		base = of_iomap(node, 0);
node              288 arch/arm/mm/cache-tauros2.c 	struct device_node *node;
node              292 arch/arm/mm/cache-tauros2.c 	node = of_find_matching_node(NULL, tauros2_ids);
node              293 arch/arm/mm/cache-tauros2.c 	if (!node) {
node              296 arch/arm/mm/cache-tauros2.c 		ret = of_property_read_u32(node, "marvell,tauros2-cache-features", &f);
node               44 arch/arm/plat-pxa/ssp.c 	list_for_each_entry(ssp, &ssp_list, node) {
node               54 arch/arm/plat-pxa/ssp.c 	if (&ssp->node == &ssp_list)
node               68 arch/arm/plat-pxa/ssp.c 	list_for_each_entry(ssp, &ssp_list, node) {
node               78 arch/arm/plat-pxa/ssp.c 	if (&ssp->node == &ssp_list)
node              173 arch/arm/plat-pxa/ssp.c 	list_add(&ssp->node, &ssp_list);
node              190 arch/arm/plat-pxa/ssp.c 	list_del(&ssp->node);
node              207 arch/arm/xen/enlighten.c static int __init fdt_find_hyper_node(unsigned long node, const char *uname,
node              216 arch/arm/xen/enlighten.c 	if (of_flat_dt_is_compatible(node, hyper_node.compat))
node              219 arch/arm/xen/enlighten.c 	s = of_get_flat_dt_prop(node, "compatible", &len);
node              232 arch/arm/xen/enlighten.c 		if ((of_get_flat_dt_subnode_by_name(node, "uefi") > 0) &&
node               77 arch/arm64/include/asm/debug-monitors.h 	struct list_head node;
node               88 arch/arm64/include/asm/debug-monitors.h 	struct list_head node;
node               23 arch/arm64/include/asm/numa.h const struct cpumask *cpumask_of_node(int node);
node               26 arch/arm64/include/asm/numa.h static inline const struct cpumask *cpumask_of_node(int node)
node               28 arch/arm64/include/asm/numa.h 	return node_to_cpumask_map[node];
node               17 arch/arm64/include/asm/traps.h 	struct list_head node;
node               18 arch/arm64/include/asm/vmap_stack.h static inline unsigned long *arch_alloc_vmap_stack(size_t stack_size, int node)
node               24 arch/arm64/include/asm/vmap_stack.h 				    THREADINFO_GFP, PAGE_KERNEL, 0, node,
node               64 arch/arm64/kernel/acpi.c static int __init dt_scan_depth1_nodes(unsigned long node,
node               79 arch/arm64/kernel/acpi.c 	    of_flat_dt_is_compatible(node, "xen,xen"))
node               52 arch/arm64/kernel/acpi_numa.c 	int cpu, pxm, node;
node               65 arch/arm64/kernel/acpi_numa.c 	node = pxm_to_node(pxm);
node               77 arch/arm64/kernel/acpi_numa.c 	acpi_early_node_map[cpu] = node;
node               79 arch/arm64/kernel/acpi_numa.c 		cpu_logical_map(cpu), node);
node               94 arch/arm64/kernel/acpi_numa.c 	int pxm, node;
node              110 arch/arm64/kernel/acpi_numa.c 	node = acpi_map_pxm_to_node(pxm);
node              112 arch/arm64/kernel/acpi_numa.c 	if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) {
node              118 arch/arm64/kernel/acpi_numa.c 	node_set(node, numa_nodes_parsed);
node               52 arch/arm64/kernel/armv8_deprecated.c 	struct list_head node;
node              126 arch/arm64/kernel/armv8_deprecated.c 	list_for_each_entry(insn, &insn_emulation, node) {
node              197 arch/arm64/kernel/armv8_deprecated.c 	list_add(&insn->node, &insn_emulation);
node              243 arch/arm64/kernel/armv8_deprecated.c 	list_for_each_entry(insn, &insn_emulation, node) {
node              160 arch/arm64/kernel/debug-monitors.c static void register_debug_hook(struct list_head *node, struct list_head *list)
node              163 arch/arm64/kernel/debug-monitors.c 	list_add_rcu(node, list);
node              168 arch/arm64/kernel/debug-monitors.c static void unregister_debug_hook(struct list_head *node)
node              171 arch/arm64/kernel/debug-monitors.c 	list_del_rcu(node);
node              178 arch/arm64/kernel/debug-monitors.c 	register_debug_hook(&hook->node, &user_step_hook);
node              183 arch/arm64/kernel/debug-monitors.c 	unregister_debug_hook(&hook->node);
node              188 arch/arm64/kernel/debug-monitors.c 	register_debug_hook(&hook->node, &kernel_step_hook);
node              193 arch/arm64/kernel/debug-monitors.c 	unregister_debug_hook(&hook->node);
node              214 arch/arm64/kernel/debug-monitors.c 	list_for_each_entry_rcu(hook, list, node)	{
node              282 arch/arm64/kernel/debug-monitors.c 	register_debug_hook(&hook->node, &user_break_hook);
node              287 arch/arm64/kernel/debug-monitors.c 	unregister_debug_hook(&hook->node);
node              292 arch/arm64/kernel/debug-monitors.c 	register_debug_hook(&hook->node, &kernel_break_hook);
node              297 arch/arm64/kernel/debug-monitors.c 	unregister_debug_hook(&hook->node);
node              312 arch/arm64/kernel/debug-monitors.c 	list_for_each_entry_rcu(hook, list, node) {
node               27 arch/arm64/kernel/kaslr.c 	int node, len;
node               31 arch/arm64/kernel/kaslr.c 	node = fdt_path_offset(fdt, "/chosen");
node               32 arch/arm64/kernel/kaslr.c 	if (node < 0)
node               35 arch/arm64/kernel/kaslr.c 	prop = fdt_getprop_w(fdt, node, "kaslr-seed", &len);
node               49 arch/arm64/kernel/kaslr.c 		int node;
node               52 arch/arm64/kernel/kaslr.c 		node = fdt_path_offset(fdt, "/chosen");
node               53 arch/arm64/kernel/kaslr.c 		if (node < 0)
node               56 arch/arm64/kernel/kaslr.c 		prop = fdt_getprop(fdt, node, "bootargs", NULL);
node              204 arch/arm64/kernel/pci.c 	list_for_each_entry(child, &bus->children, node)
node              291 arch/arm64/kernel/traps.c 	list_add(&hook->node, &undef_hook);
node              300 arch/arm64/kernel/traps.c 	list_del(&hook->node);
node              340 arch/arm64/kernel/traps.c 	list_for_each_entry(hook, &undef_hook, node)
node              123 arch/arm64/mm/init.c static int __init early_init_dt_scan_elfcorehdr(unsigned long node,
node              132 arch/arm64/mm/init.c 	reg = of_get_flat_dt_prop(node, "linux,elfcorehdr", &len);
node              276 arch/arm64/mm/init.c static int __init early_init_dt_scan_usablemem(unsigned long node,
node              286 arch/arm64/mm/init.c 	reg = of_get_flat_dt_prop(node, "linux,usable-memory-range", &len);
node               34 arch/arm64/mm/kasan_init.c static phys_addr_t __init kasan_alloc_zeroed_page(int node)
node               38 arch/arm64/mm/kasan_init.c 					      MEMBLOCK_ALLOC_KASAN, node);
node               41 arch/arm64/mm/kasan_init.c 		      __func__, PAGE_SIZE, PAGE_SIZE, node,
node               47 arch/arm64/mm/kasan_init.c static phys_addr_t __init kasan_alloc_raw_page(int node)
node               51 arch/arm64/mm/kasan_init.c 						MEMBLOCK_ALLOC_KASAN, node);
node               54 arch/arm64/mm/kasan_init.c 		      __func__, PAGE_SIZE, PAGE_SIZE, node,
node               60 arch/arm64/mm/kasan_init.c static pte_t *__init kasan_pte_offset(pmd_t *pmdp, unsigned long addr, int node,
node               66 arch/arm64/mm/kasan_init.c 					: kasan_alloc_zeroed_page(node);
node               74 arch/arm64/mm/kasan_init.c static pmd_t *__init kasan_pmd_offset(pud_t *pudp, unsigned long addr, int node,
node               80 arch/arm64/mm/kasan_init.c 					: kasan_alloc_zeroed_page(node);
node               87 arch/arm64/mm/kasan_init.c static pud_t *__init kasan_pud_offset(pgd_t *pgdp, unsigned long addr, int node,
node               93 arch/arm64/mm/kasan_init.c 					: kasan_alloc_zeroed_page(node);
node              101 arch/arm64/mm/kasan_init.c 				      unsigned long end, int node, bool early)
node              104 arch/arm64/mm/kasan_init.c 	pte_t *ptep = kasan_pte_offset(pmdp, addr, node, early);
node              109 arch/arm64/mm/kasan_init.c 					: kasan_alloc_raw_page(node);
node              118 arch/arm64/mm/kasan_init.c 				      unsigned long end, int node, bool early)
node              121 arch/arm64/mm/kasan_init.c 	pmd_t *pmdp = kasan_pmd_offset(pudp, addr, node, early);
node              125 arch/arm64/mm/kasan_init.c 		kasan_pte_populate(pmdp, addr, next, node, early);
node              130 arch/arm64/mm/kasan_init.c 				      unsigned long end, int node, bool early)
node              133 arch/arm64/mm/kasan_init.c 	pud_t *pudp = kasan_pud_offset(pgdp, addr, node, early);
node              137 arch/arm64/mm/kasan_init.c 		kasan_pmd_populate(pudp, addr, next, node, early);
node              142 arch/arm64/mm/kasan_init.c 				      int node, bool early)
node              150 arch/arm64/mm/kasan_init.c 		kasan_pud_populate(pgdp, addr, next, node, early);
node              168 arch/arm64/mm/kasan_init.c 				      int node)
node              170 arch/arm64/mm/kasan_init.c 	kasan_pgd_populate(start & PAGE_MASK, PAGE_ALIGN(end), node, false);
node              730 arch/arm64/mm/mmu.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              733 arch/arm64/mm/mmu.c 	return vmemmap_populate_basepages(start, end, node);
node              736 arch/arm64/mm/mmu.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              748 arch/arm64/mm/mmu.c 		pgdp = vmemmap_pgd_populate(addr, node);
node              752 arch/arm64/mm/mmu.c 		pudp = vmemmap_pud_populate(pgdp, addr, node);
node              760 arch/arm64/mm/mmu.c 			p = vmemmap_alloc_block_buf(PMD_SIZE, node);
node              766 arch/arm64/mm/mmu.c 			vmemmap_verify((pte_t *)pmdp, node, addr, next);
node               47 arch/arm64/mm/numa.c const struct cpumask *cpumask_of_node(int node)
node               49 arch/arm64/mm/numa.c 	if (WARN_ON(node >= nr_node_ids))
node               52 arch/arm64/mm/numa.c 	if (WARN_ON(node_to_cpumask_map[node] == NULL))
node               55 arch/arm64/mm/numa.c 	return node_to_cpumask_map[node];
node               99 arch/arm64/mm/numa.c 	int node;
node              106 arch/arm64/mm/numa.c 	for (node = 0; node < nr_node_ids; node++) {
node              107 arch/arm64/mm/numa.c 		alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]);
node              108 arch/arm64/mm/numa.c 		cpumask_clear(node_to_cpumask_map[node]);
node               80 arch/c6x/include/asm/clock.h 	struct list_head	node;
node               95 arch/c6x/kernel/setup.c 	struct device_node *node;
node              189 arch/c6x/kernel/setup.c 	for_each_of_cpu_node(node)
node              192 arch/c6x/kernel/setup.c 	node = of_find_node_by_name(NULL, "soc");
node              193 arch/c6x/kernel/setup.c 	if (node) {
node              194 arch/c6x/kernel/setup.c 		if (of_property_read_string(node, "model", &c6x_soc_name))
node              196 arch/c6x/kernel/setup.c 		of_node_put(node);
node              426 arch/c6x/platforms/cache.c 	struct device_node *node;
node              428 arch/c6x/platforms/cache.c 	node = of_find_compatible_node(NULL, NULL, "ti,c64x+cache");
node              429 arch/c6x/platforms/cache.c 	if (!node)
node              432 arch/c6x/platforms/cache.c 	cache_base = of_iomap(node, 0);
node              434 arch/c6x/platforms/cache.c 	of_node_put(node);
node               99 arch/c6x/platforms/dscr.c 	void (*init)(struct device_node *node);
node              281 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstat(struct device_node *node,
node              287 arch/c6x/platforms/dscr.c 	err = of_property_read_u32_array(node, "ti,dscr-devstat", &val, 1);
node              293 arch/c6x/platforms/dscr.c static void __init dscr_parse_silicon_rev(struct device_node *node,
node              299 arch/c6x/platforms/dscr.c 	err = of_property_read_u32_array(node, "ti,dscr-silicon-rev", vals, 3);
node              324 arch/c6x/platforms/dscr.c static void __init dscr_parse_mac_fuse(struct device_node *node,
node              330 arch/c6x/platforms/dscr.c 	err = of_property_read_u32_array(node, "ti,dscr-mac-fuse-regs",
node              343 arch/c6x/platforms/dscr.c static void __init dscr_parse_rmii_resets(struct device_node *node,
node              350 arch/c6x/platforms/dscr.c 	p = of_get_property(node, "ti,dscr-rmii-resets", &size);
node              365 arch/c6x/platforms/dscr.c static void __init dscr_parse_privperm(struct device_node *node,
node              371 arch/c6x/platforms/dscr.c 	err = of_property_read_u32_array(node, "ti,dscr-privperm", vals, 2);
node              393 arch/c6x/platforms/dscr.c static void __init dscr_parse_locked_regs(struct device_node *node,
node              400 arch/c6x/platforms/dscr.c 	p = of_get_property(node, "ti,dscr-locked-regs", &size);
node              428 arch/c6x/platforms/dscr.c static void __init dscr_parse_kick_regs(struct device_node *node,
node              434 arch/c6x/platforms/dscr.c 	err = of_property_read_u32_array(node, "ti,dscr-kick-regs", vals, 4);
node              466 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstate_ctl_regs(struct device_node *node,
node              473 arch/c6x/platforms/dscr.c 	p = of_get_property(node, "ti,dscr-devstate-ctl-regs", &size);
node              522 arch/c6x/platforms/dscr.c static void __init dscr_parse_devstate_stat_regs(struct device_node *node,
node              529 arch/c6x/platforms/dscr.c 	p = of_get_property(node, "ti,dscr-devstate-stat-regs", &size);
node              569 arch/c6x/platforms/dscr.c 	struct device_node *node;
node              574 arch/c6x/platforms/dscr.c 	node = of_find_matching_node(NULL, dscr_ids);
node              575 arch/c6x/platforms/dscr.c 	if (!node)
node              578 arch/c6x/platforms/dscr.c 	base = of_iomap(node, 0);
node              580 arch/c6x/platforms/dscr.c 		of_node_put(node);
node              586 arch/c6x/platforms/dscr.c 	dscr_parse_devstat(node, base);
node              587 arch/c6x/platforms/dscr.c 	dscr_parse_silicon_rev(node, base);
node              588 arch/c6x/platforms/dscr.c 	dscr_parse_mac_fuse(node, base);
node              589 arch/c6x/platforms/dscr.c 	dscr_parse_rmii_resets(node, base);
node              590 arch/c6x/platforms/dscr.c 	dscr_parse_locked_regs(node, base);
node              591 arch/c6x/platforms/dscr.c 	dscr_parse_kick_regs(node, base);
node              592 arch/c6x/platforms/dscr.c 	dscr_parse_devstate_ctl_regs(node, base);
node              593 arch/c6x/platforms/dscr.c 	dscr_parse_devstate_stat_regs(node, base);
node              594 arch/c6x/platforms/dscr.c 	dscr_parse_privperm(node, base);
node               44 arch/c6x/platforms/emif.c 	struct device_node *node;
node               49 arch/c6x/platforms/emif.c 	node = of_find_matching_node(NULL, emifa_match);
node               50 arch/c6x/platforms/emif.c 	if (!node)
node               53 arch/c6x/platforms/emif.c 	regs = of_iomap(node, 0);
node               58 arch/c6x/platforms/emif.c 	err = of_property_read_u32_array(node, "ti,dscr-dev-enable", &val, 1);
node               63 arch/c6x/platforms/emif.c 	p = of_get_property(node, "ti,emifa-ce-config", &len);
node               72 arch/c6x/platforms/emif.c 	err = of_property_read_u32_array(node, "ti,emifa-burst-priority", &val, 1);
node               76 arch/c6x/platforms/emif.c 	err = of_property_read_u32_array(node, "ti,emifa-async-wait-control", &val, 1);
node               81 arch/c6x/platforms/emif.c 	of_node_put(node);
node              166 arch/c6x/platforms/pll.c 	list_add_tail(&clk->node, &clocks);
node              193 arch/c6x/platforms/pll.c 	list_del(&clk->node);
node              324 arch/c6x/platforms/pll.c 	INIT_LIST_HEAD(&clk->node);
node              412 arch/c6x/platforms/pll.c 	list_for_each_entry(clk, &clocks, node)
node               29 arch/c6x/platforms/plldata.c 	.node = LIST_HEAD_INIT(clkin1.node),
node              168 arch/c6x/platforms/plldata.c static void __init c6455_setup_clocks(struct device_node *node)
node              206 arch/c6x/platforms/plldata.c static void __init c6457_setup_clocks(struct device_node *node)
node              256 arch/c6x/platforms/plldata.c static void __init c6472_setup_clocks(struct device_node *node)
node              305 arch/c6x/platforms/plldata.c static void __init c6474_setup_clocks(struct device_node *node)
node              354 arch/c6x/platforms/plldata.c static void __init c6678_setup_clocks(struct device_node *node)
node              422 arch/c6x/platforms/plldata.c 	struct device_node *node;
node              427 arch/c6x/platforms/plldata.c 	node = of_find_matching_node(NULL, c6x_clkc_match);
node              428 arch/c6x/platforms/plldata.c 	if (!node)
node              431 arch/c6x/platforms/plldata.c 	pll->base = of_iomap(node, 0);
node              435 arch/c6x/platforms/plldata.c 	err = of_property_read_u32(node, "clock-frequency", &val);
node              438 arch/c6x/platforms/plldata.c 		       node, (int)val / 1000000);
node              443 arch/c6x/platforms/plldata.c 	err = of_property_read_u32(node, "ti,c64x+pll-bypass-delay", &val);
node              448 arch/c6x/platforms/plldata.c 	err = of_property_read_u32(node, "ti,c64x+pll-reset-delay", &val);
node              453 arch/c6x/platforms/plldata.c 	err = of_property_read_u32(node, "ti,c64x+pll-lock-delay", &val);
node              459 arch/c6x/platforms/plldata.c 	id = of_match_node(c6x_clkc_match, node);
node              462 arch/c6x/platforms/plldata.c 		__setup_clocks(node);
node              466 arch/c6x/platforms/plldata.c 	of_node_put(node);
node             1296 arch/csky/kernel/perf_event.c 	struct device_node *node = pdev->dev.of_node;
node             1305 arch/csky/kernel/perf_event.c 	if (of_property_read_u32(node, "count-width",
node              138 arch/csky/kernel/smp.c 	struct device_node *node = NULL;
node              141 arch/csky/kernel/smp.c 	for_each_of_cpu_node(node) {
node              142 arch/csky/kernel/smp.c 		if (!of_device_is_available(node))
node              145 arch/csky/kernel/smp.c 		if (of_property_read_u32(node, "reg", &cpu))
node              212 arch/ia64/hp/common/sba_iommu.c 	unsigned int	node;		/* node where this IOC lives */
node             1116 arch/ia64/hp/common/sba_iommu.c 	int node = -1;
node             1122 arch/ia64/hp/common/sba_iommu.c 	node = ioc->node;
node             1125 arch/ia64/hp/common/sba_iommu.c 	page = alloc_pages_node(node, flags, get_order(size));
node             1878 arch/ia64/hp/common/sba_iommu.c 	if (ioc->node != NUMA_NO_NODE)
node             1879 arch/ia64/hp/common/sba_iommu.c 		seq_printf(s, "NUMA node       : %d\n", ioc->node);
node             1970 arch/ia64/hp/common/sba_iommu.c 	unsigned int node;
node             1972 arch/ia64/hp/common/sba_iommu.c 	node = acpi_get_node(handle);
node             1973 arch/ia64/hp/common/sba_iommu.c 	if (node != NUMA_NO_NODE && !node_online(node))
node             1974 arch/ia64/hp/common/sba_iommu.c 		node = NUMA_NO_NODE;
node             1976 arch/ia64/hp/common/sba_iommu.c 	ioc->node = node;
node               69 arch/ia64/include/asm/meminit.h static inline int vmemmap_find_next_valid_pfn(int node, int i)
node               28 arch/ia64/include/asm/nodedata.h 	short			node;
node               50 arch/ia64/include/asm/pci.h 	int node;		/* nearest node with memory or NUMA_NO_NODE for global allocation */
node              455 arch/ia64/include/asm/sal.h 		    node            : 1,
node              474 arch/ia64/include/asm/sal.h 	u16 node;
node               58 arch/ia64/include/asm/thread_info.h #define alloc_thread_stack_node(tsk, node)	\
node               63 arch/ia64/include/asm/thread_info.h #define alloc_thread_stack_node(tsk, node)	((unsigned long *) 0)
node               83 arch/ia64/include/asm/thread_info.h #define alloc_task_struct_node(node)						\
node               85 arch/ia64/include/asm/thread_info.h 	struct page *page = alloc_pages_node(node, GFP_KERNEL | __GFP_COMP,	\
node               28 arch/ia64/include/asm/topology.h #define cpumask_of_node(node) ((node) == -1 ?				\
node               30 arch/ia64/include/asm/topology.h 			       &node_to_cpu_mask[node])
node               35 arch/ia64/include/asm/topology.h #define pcibus_to_node(bus) PCI_CONTROLLER(bus)->node
node              834 arch/ia64/kernel/acpi.c 	int node;
node              862 arch/ia64/kernel/acpi.c 	node = acpi_get_node(handle);
node              863 arch/ia64/kernel/acpi.c 	if (node == NUMA_NO_NODE || !node_online(node) ||
node              864 arch/ia64/kernel/acpi.c 	    cpumask_empty(cpumask_of_node(node)))
node              868 arch/ia64/kernel/acpi.c 	map_iosapic_to_node(gsi_base, node);
node              122 arch/ia64/kernel/iosapic.c 	unsigned short	node;		/* numa node association via pxm */
node              655 arch/ia64/kernel/iosapic.c 		    iosapic_lists[iosapic_index].node == MAX_NUMNODES)
node              658 arch/ia64/kernel/iosapic.c 		cpu_mask = cpumask_of_node(iosapic_lists[iosapic_index].node);
node             1066 arch/ia64/kernel/iosapic.c 	iosapic_lists[index].node = MAX_NUMNODES;
node             1123 arch/ia64/kernel/iosapic.c void map_iosapic_to_node(unsigned int gsi_base, int node)
node             1133 arch/ia64/kernel/iosapic.c 	iosapic_lists[index].node = node;
node              175 arch/ia64/kernel/msi_ia64.c int dmar_alloc_hwirq(int id, int node, void *arg)
node               59 arch/ia64/kernel/numa.c 	int cpu, i, node;
node               61 arch/ia64/kernel/numa.c 	for(node=0; node < MAX_NUMNODES; node++)
node               62 arch/ia64/kernel/numa.c 		cpumask_clear(&node_to_cpu_mask[node]);
node               65 arch/ia64/kernel/numa.c 		node = NUMA_NO_NODE;
node               68 arch/ia64/kernel/numa.c 				node = node_cpuid[i].nid;
node               71 arch/ia64/kernel/numa.c 		map_cpu_to_node(cpu, node);
node               55 arch/ia64/mm/discontig.c #define NODEDATA_ALIGN(addr, node)						\
node               57 arch/ia64/mm/discontig.c 	     (((node)*PERCPU_PAGE_SIZE) & (MAX_NODE_ALIGN_OFFSET - 1)))
node               72 arch/ia64/mm/discontig.c 				  int node)
node               79 arch/ia64/mm/discontig.c 	if (!mem_data[node].min_pfn) {
node               80 arch/ia64/mm/discontig.c 		mem_data[node].min_pfn = spfn;
node               81 arch/ia64/mm/discontig.c 		mem_data[node].max_pfn = epfn;
node               83 arch/ia64/mm/discontig.c 		mem_data[node].min_pfn = min(spfn, mem_data[node].min_pfn);
node               84 arch/ia64/mm/discontig.c 		mem_data[node].max_pfn = max(epfn, mem_data[node].max_pfn);
node               98 arch/ia64/mm/discontig.c static int __meminit early_nr_cpus_node(int node)
node              103 arch/ia64/mm/discontig.c 		if (node == node_cpuid[cpu].nid)
node              113 arch/ia64/mm/discontig.c static unsigned long __meminit compute_pernodesize(int node)
node              117 arch/ia64/mm/discontig.c 	cpus = early_nr_cpus_node(node);
node              119 arch/ia64/mm/discontig.c 	pernodesize += node * L1_CACHE_BYTES;
node              136 arch/ia64/mm/discontig.c static void *per_cpu_node_setup(void *cpu_data, int node)
node              144 arch/ia64/mm/discontig.c 		if (node != node_cpuid[cpu].nid)
node              189 arch/ia64/mm/discontig.c 	int node, prev_node, unit, nr_units;
node              205 arch/ia64/mm/discontig.c 	for_each_node(node)
node              207 arch/ia64/mm/discontig.c 			if (node == node_cpuid[cpu].nid)
node              234 arch/ia64/mm/discontig.c 		node = node_cpuid[cpu].nid;
node              236 arch/ia64/mm/discontig.c 		if (node == prev_node) {
node              240 arch/ia64/mm/discontig.c 		prev_node = node;
node              259 arch/ia64/mm/discontig.c static void __init fill_pernode(int node, unsigned long pernode,
node              263 arch/ia64/mm/discontig.c 	int cpus = early_nr_cpus_node(node);
node              265 arch/ia64/mm/discontig.c 	mem_data[node].pernode_addr = pernode;
node              266 arch/ia64/mm/discontig.c 	mem_data[node].pernode_size = pernodesize;
node              271 arch/ia64/mm/discontig.c 	pernode += node * L1_CACHE_BYTES;
node              273 arch/ia64/mm/discontig.c 	pgdat_list[node] = __va(pernode);
node              276 arch/ia64/mm/discontig.c 	mem_data[node].node_data = __va(pernode);
node              280 arch/ia64/mm/discontig.c 	cpu_data = per_cpu_node_setup(cpu_data, node);
node              314 arch/ia64/mm/discontig.c 				     int node)
node              326 arch/ia64/mm/discontig.c 	if (spfn < mem_data[node].min_pfn || epfn > mem_data[node].max_pfn)
node              330 arch/ia64/mm/discontig.c 	if (mem_data[node].pernode_addr)
node              337 arch/ia64/mm/discontig.c 	pernodesize = compute_pernodesize(node);
node              338 arch/ia64/mm/discontig.c 	pernode = NODEDATA_ALIGN(start, node);
node              342 arch/ia64/mm/discontig.c 		fill_pernode(node, pernode, pernodesize);
node              357 arch/ia64/mm/discontig.c 	int node;
node              359 arch/ia64/mm/discontig.c 	for_each_online_node(node) {
node              360 arch/ia64/mm/discontig.c 		if (node_isset(node, memory_less_mask))
node              364 arch/ia64/mm/discontig.c 		size = mem_data[node].pernode_size;
node              365 arch/ia64/mm/discontig.c 		base = __pa(mem_data[node].pernode_addr);
node              373 arch/ia64/mm/discontig.c 	int node;
node              383 arch/ia64/mm/discontig.c 	for_each_node(node) {
node              384 arch/ia64/mm/discontig.c 		if (pgdat_list[node]) {
node              385 arch/ia64/mm/discontig.c 			dst = LOCAL_DATA_ADDR(pgdat_list[node])->pg_data_ptrs;
node              400 arch/ia64/mm/discontig.c 	int cpu, node;
node              407 arch/ia64/mm/discontig.c 		node = node_cpuid[cpu].nid;
node              409 arch/ia64/mm/discontig.c 			mem_data[node].node_data;
node              415 arch/ia64/mm/discontig.c 		node = node_cpuid[cpu].nid;
node              418 arch/ia64/mm/discontig.c 		cpu0_cpu_info->node_data = mem_data[node].node_data;
node              434 arch/ia64/mm/discontig.c 	int bestnode = NUMA_NO_NODE, node, anynode = 0;
node              436 arch/ia64/mm/discontig.c 	for_each_online_node(node) {
node              437 arch/ia64/mm/discontig.c 		if (node_isset(node, memory_less_mask))
node              439 arch/ia64/mm/discontig.c 		else if (node_distance(nid, node) < best) {
node              440 arch/ia64/mm/discontig.c 			best = node_distance(nid, node);
node              441 arch/ia64/mm/discontig.c 			bestnode = node;
node              443 arch/ia64/mm/discontig.c 		anynode = node;
node              469 arch/ia64/mm/discontig.c 	int node;
node              471 arch/ia64/mm/discontig.c 	for_each_node_mask(node, memory_less_mask) {
node              472 arch/ia64/mm/discontig.c 		pernodesize = compute_pernodesize(node);
node              473 arch/ia64/mm/discontig.c 		pernode = memory_less_node_alloc(node, pernodesize);
node              474 arch/ia64/mm/discontig.c 		fill_pernode(node, __pa(pernode), pernodesize);
node              488 arch/ia64/mm/discontig.c 	int node;
node              507 arch/ia64/mm/discontig.c 	for_each_online_node(node)
node              508 arch/ia64/mm/discontig.c 		if (mem_data[node].min_pfn)
node              509 arch/ia64/mm/discontig.c 			node_clear(node, memory_less_mask);
node              599 arch/ia64/mm/discontig.c 	int node;
node              615 arch/ia64/mm/discontig.c 	for_each_online_node(node) {
node              616 arch/ia64/mm/discontig.c 		pfn_offset = mem_data[node].min_pfn;
node              619 arch/ia64/mm/discontig.c 		NODE_DATA(node)->node_mem_map = vmem_map + pfn_offset;
node              621 arch/ia64/mm/discontig.c 		if (mem_data[node].max_pfn > max_pfn)
node              622 arch/ia64/mm/discontig.c 			max_pfn = mem_data[node].max_pfn;
node              656 arch/ia64/mm/discontig.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              659 arch/ia64/mm/discontig.c 	return vmemmap_populate_basepages(start, end, node);
node              373 arch/ia64/mm/init.c int vmemmap_find_next_valid_pfn(int node, int i)
node              377 arch/ia64/mm/init.c 	pg_data_t *pgdat = NODE_DATA(node);
node              431 arch/ia64/mm/init.c 	int node;
node              442 arch/ia64/mm/init.c 	node = paddr_to_nid(__pa(start));
node              447 arch/ia64/mm/init.c 			pud = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node);
node              455 arch/ia64/mm/init.c 			pmd = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node);
node              463 arch/ia64/mm/init.c 			pte = memblock_alloc_node(PAGE_SIZE, PAGE_SIZE, node);
node              472 arch/ia64/mm/init.c 							 node);
node              483 arch/ia64/mm/init.c 	      __func__, PAGE_SIZE, PAGE_SIZE, node);
node              305 arch/ia64/pci/pci.c 	info->controller.node = acpi_get_node(device->handle);
node              116 arch/m68k/include/asm/dvma.h 	int node;                /* Prom node for this DMA device */
node              278 arch/m68k/include/asm/openprom.h 	int (*no_nextnode)(int node);
node              279 arch/m68k/include/asm/openprom.h 	int (*no_child)(int node);
node              280 arch/m68k/include/asm/openprom.h 	int (*no_proplen)(int node, char *name);
node              281 arch/m68k/include/asm/openprom.h 	int (*no_getprop)(int node, char *name, char *val);
node              282 arch/m68k/include/asm/openprom.h 	int (*no_setprop)(int node, char *name, char *val, int len);
node              283 arch/m68k/include/asm/openprom.h 	char * (*no_nextprop)(int node, char *name);
node              219 arch/m68k/include/asm/oplib.h extern int prom_getsibling(int node);
node              233 arch/m68k/include/asm/oplib.h extern int prom_getint(int node, char *property);
node              236 arch/m68k/include/asm/oplib.h extern int prom_getintdefault(int node, char *property, int defval);
node              239 arch/m68k/include/asm/oplib.h extern int prom_getbool(int node, char *prop);
node              242 arch/m68k/include/asm/oplib.h extern void prom_getstring(int node, char *prop, char *buf, int bufsize);
node              255 arch/m68k/include/asm/oplib.h extern char *prom_firstprop(int node);
node              260 arch/m68k/include/asm/oplib.h extern char *prom_nextprop(int node, char *prev_property);
node              263 arch/m68k/include/asm/oplib.h extern int prom_node_has_property(int node, char *property);
node              268 arch/m68k/include/asm/oplib.h extern int prom_setprop(int node, char *prop_name, char *prop_value,
node              288 arch/m68k/include/asm/oplib.h extern void prom_apply_generic_ranges(int node, int parent,
node               19 arch/m68k/include/asm/pgalloc.h extern void m68k_setup_node(int node);
node               60 arch/m68k/mm/init.c void __init m68k_setup_node(int node)
node               63 arch/m68k/mm/init.c 	struct m68k_mem_info *info = m68k_memory + node;
node               70 arch/m68k/mm/init.c 			pr_warn("overlap at %u for chunk %u\n", i, node);
node               71 arch/m68k/mm/init.c 		pg_data_table[i] = pg_data_map + node;
node               74 arch/m68k/mm/init.c 	node_set_online(node);
node              115 arch/m68k/mm/motorola.c static void __init map_node(int node)
node              124 arch/m68k/mm/motorola.c 	size = m68k_memory[node].size;
node              125 arch/m68k/mm/motorola.c 	physaddr = m68k_memory[node].addr;
node              133 arch/m68k/sun3x/prom.c int prom_getintdefault(int node, char *property, int deflt)
node              138 arch/m68k/sun3x/prom.c int prom_getbool (int node, char *prop)
node              130 arch/microblaze/include/asm/pci-bridge.h 			struct device_node *node);
node              132 arch/microblaze/pci/pci-common.c struct pci_controller *pci_find_hose_for_OF_device(struct device_node *node)
node              134 arch/microblaze/pci/pci-common.c 	while (node) {
node              137 arch/microblaze/pci/pci-common.c 			if (hose->dn == node)
node              139 arch/microblaze/pci/pci-common.c 		node = node->parent;
node              741 arch/microblaze/pci/pci-common.c 	list_for_each_entry(b, &bus->children, node)
node              883 arch/microblaze/pci/pci-common.c 	list_for_each_entry(b, &pci_root_buses, node)
node              893 arch/microblaze/pci/pci-common.c 	list_for_each_entry(b, &pci_root_buses, node)
node              967 arch/microblaze/pci/pci-common.c 	struct device_node *node = hose->dn;
node              969 arch/microblaze/pci/pci-common.c 	pr_debug("PCI: Scanning PHB %pOF\n", node);
node               45 arch/mips/cavium-octeon/octeon-irq.c 	int			node;
node             1147 arch/mips/cavium-octeon/octeon-irq.c 				struct device_node *node,
node             1157 arch/mips/cavium-octeon/octeon-irq.c 	if (irq_domain_get_of_node(d) != node)
node             1184 arch/mips/cavium-octeon/octeon-irq.c 		       node,
node             1196 arch/mips/cavium-octeon/octeon-irq.c 			       struct device_node *node,
node             1891 arch/mips/cavium-octeon/octeon-irq.c 				struct device_node *node,
node             2161 arch/mips/cavium-octeon/octeon-irq.c 				   struct device_node *node,
node             2334 arch/mips/cavium-octeon/octeon-irq.c 			 struct device_node *node,
node             2547 arch/mips/cavium-octeon/octeon-irq.c 						       ciu3_info->node);
node             2553 arch/mips/cavium-octeon/octeon-irq.c 	cd->ciu_node = ciu3_info->node;
node             2850 arch/mips/cavium-octeon/octeon-irq.c 	int node;
node             2857 arch/mips/cavium-octeon/octeon-irq.c 	node = 0; /* of_node_to_nid(ciu_node); */
node             2858 arch/mips/cavium-octeon/octeon-irq.c 	ciu3_info = kzalloc_node(sizeof(*ciu3_info), GFP_KERNEL, node);
node             2871 arch/mips/cavium-octeon/octeon-irq.c 	ciu3_info->node = node;
node             2881 arch/mips/cavium-octeon/octeon-irq.c 	if (node == cvmx_get_node_num()) {
node             2886 arch/mips/cavium-octeon/octeon-irq.c 		i = irq_alloc_descs_from(OCTEON_IRQ_MBOX0, 8, node);
node             2903 arch/mips/cavium-octeon/octeon-irq.c 	octeon_ciu3_info_per_node[node] = ciu3_info;
node             2905 arch/mips/cavium-octeon/octeon-irq.c 	if (node == cvmx_get_node_num()) {
node             2908 arch/mips/cavium-octeon/octeon-irq.c 		if (node == 0)
node             2972 arch/mips/cavium-octeon/octeon-irq.c struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block)
node             2976 arch/mips/cavium-octeon/octeon-irq.c 	ciu3_info = octeon_ciu3_info_per_node[node & CVMX_NODE_MASK];
node              591 arch/mips/cavium-octeon/octeon-platform.c static void __init octeon_fdt_rm_ethernet(int node)
node              595 arch/mips/cavium-octeon/octeon-platform.c 	phy_handle = fdt_getprop(initial_boot_params, node, "phy-handle", NULL);
node              603 arch/mips/cavium-octeon/octeon-platform.c 	fdt_nop_node(initial_boot_params, node);
node              235 arch/mips/cavium-octeon/octeon-usb.c 	struct device_node *node = dev->of_node;
node              238 arch/mips/cavium-octeon/octeon-usb.c 	if (of_find_property(node, "power", &len) != NULL) {
node              240 arch/mips/cavium-octeon/octeon-usb.c 			of_property_read_u32_array(node, "power", gpio_pwr, 3);
node              244 arch/mips/cavium-octeon/octeon-usb.c 			of_property_read_u32_array(node, "power", gpio_pwr, 2);
node              500 arch/mips/cavium-octeon/octeon-usb.c 	struct device_node *node;
node              508 arch/mips/cavium-octeon/octeon-usb.c 	node = NULL;
node              510 arch/mips/cavium-octeon/octeon-usb.c 		node = of_find_node_by_name(node, "uctl");
node              511 arch/mips/cavium-octeon/octeon-usb.c 		if (!node)
node              514 arch/mips/cavium-octeon/octeon-usb.c 		if (of_device_is_compatible(node, compat_node_name)) {
node              515 arch/mips/cavium-octeon/octeon-usb.c 			pdev = of_find_device_by_node(node);
node              545 arch/mips/cavium-octeon/octeon-usb.c 	} while (node != NULL);
node               22 arch/mips/include/asm/clock.h 	struct list_head node;
node               46 arch/mips/include/asm/mach-ip27/kernel-entry-init.h 	li	t0, 0x1c000		# Offset of text into node memory
node               18 arch/mips/include/asm/mach-ip27/topology.h #define cpumask_of_node(node)	((node) == -1 ?				\
node               20 arch/mips/include/asm/mach-ip27/topology.h 				 &hub_data(node)->h_cpus)
node                8 arch/mips/include/asm/mach-loongson64/topology.h #define cpumask_of_node(node)	(&__node_data[(node)]->cpumask)
node               72 arch/mips/include/asm/mach-netlogic/multi-node.h void nlm_node_init(int node);
node               94 arch/mips/include/asm/netlogic/common.h uint64_t nlm_pci_irqmask(int node);
node               95 arch/mips/include/asm/netlogic/common.h void nlm_setup_pic_irq(int node, int picirq, int irq, int irt);
node               96 arch/mips/include/asm/netlogic/common.h void nlm_set_pic_extra_ack(int node, int irq,  void (*xack)(struct irq_data *));
node               99 arch/mips/include/asm/netlogic/common.h void nlm_dispatch_msi(int node, int lirq);
node              100 arch/mips/include/asm/netlogic/common.h void nlm_dispatch_msix(int node, int msixirq);
node              106 arch/mips/include/asm/netlogic/common.h static inline int nlm_irq_to_xirq(int node, int irq)
node              108 arch/mips/include/asm/netlogic/common.h 	return node * NR_IRQS / NLM_NR_NODES + irq;
node              180 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_get_bridge_pcibase(node)	nlm_pcicfg_base(cpu_is_xlp9xx() ? \
node              181 arch/mips/include/asm/netlogic/xlp-hal/bridge.h 		XLP9XX_IO_BRIDGE_OFFSET(node) : XLP_IO_BRIDGE_OFFSET(node))
node              182 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_get_bridge_regbase(node)	\
node              183 arch/mips/include/asm/netlogic/xlp-hal/bridge.h 			(nlm_get_bridge_pcibase(node) + XLP_IO_PCI_HDRSZ)
node               50 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_DEV(node, dev)		((dev) + (node) * 8)
node               53 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_HDR_OFFSET(node, bus, dev, fn) \
node               54 arch/mips/include/asm/netlogic/xlp-hal/iomap.h 		XLP_IO_PCI_OFFSET(bus, XLP_IO_DEV(node, dev), fn)
node               56 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_BRIDGE_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 0, 0)
node               58 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 0, 1)
node               59 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 0, 2)
node               60 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CIC2_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 0, 3)
node               61 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PIC_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 0, 4)
node               63 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE_OFFSET(node, i)	XLP_HDR_OFFSET(node, 0, 1, i)
node               64 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 1, 0)
node               65 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 1, 1)
node               66 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE2_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 1, 2)
node               67 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCIE3_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 1, 3)
node               69 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OFFSET(node, i)	XLP_HDR_OFFSET(node, 0, 2, i)
node               70 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_EHCI0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 0)
node               71 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 1)
node               72 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 2)
node               73 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_EHCI1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 3)
node               74 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI2_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 4)
node               75 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_USB_OHCI3_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 2, 5)
node               77 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SATA_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 3, 2)
node               80 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_OFFSET(node, i)	XLP_HDR_OFFSET(node, 0, 4, i)
node               81 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 4, 1)
node               82 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 4, 2)
node               83 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_USB_XHCI2_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 4, 3)
node               85 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NAE_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 3, 0)
node               86 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_POE_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 3, 1)
node               88 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CMS_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 4, 0)
node               90 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_DMA_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 5, 1)
node               91 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SEC_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 5, 2)
node               92 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_CMP_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 5, 3)
node               94 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART_OFFSET(node, i)	XLP_HDR_OFFSET(node, 0, 6, i)
node               95 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 0)
node               96 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_UART1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 1)
node               97 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C_OFFSET(node, i)	XLP_HDR_OFFSET(node, 0, 6, 2 + i)
node               98 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C0_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 2)
node               99 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_I2C1_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 3)
node              100 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_GPIO_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 4)
node              102 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP2XX_IO_I2C_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 7)
node              105 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SYS_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 6, 5)
node              106 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_JTAG_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 6, 6)
node              109 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NOR_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 7, 0)
node              110 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_NAND_OFFSET(node)	XLP_HDR_OFFSET(node, 0, 7, 1)
node              111 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_SPI_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 7, 2)
node              112 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_MMC_OFFSET(node)		XLP_HDR_OFFSET(node, 0, 7, 3)
node              118 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_BRIDGE_OFFSET(node)	XLP_IO_PCI_OFFSET(0, 0, node)
node              119 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PIC_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 2, 0)
node              120 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_UART_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 2, 2)
node              121 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SYS_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 6, 0)
node              122 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_FUSE_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 6, 1)
node              123 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_CLOCK_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 6, 2)
node              124 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_POWER_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 6, 3)
node              125 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_JTAG_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 6, 4)
node              127 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE_OFFSET(node, i)	XLP9XX_HDR_OFFSET(node, 1, i)
node              128 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE0_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 1, 0)
node              129 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE2_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 1, 2)
node              130 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_PCIE3_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 1, 3)
node              133 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_OFFSET(node, i)		XLP9XX_HDR_OFFSET(node, 4, i)
node              134 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_XHCI0_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 4, 1)
node              135 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_USB_XHCI1_OFFSET(node)	XLP9XX_HDR_OFFSET(node, 4, 2)
node              138 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SATA_OFFSET(node)		XLP9XX_HDR_OFFSET(node, 3, 2)
node              141 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_NOR_OFFSET(node)		XLP9XX_HDR_OFFSET(node, 7, 0)
node              142 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_NAND_OFFSET(node)		XLP9XX_HDR_OFFSET(node, 7, 1)
node              143 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_SPI_OFFSET(node)		XLP9XX_HDR_OFFSET(node, 7, 2)
node              144 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP9XX_IO_MMC_OFFSET(node)		XLP9XX_HDR_OFFSET(node, 7, 3)
node              203 arch/mips/include/asm/netlogic/xlp-hal/iomap.h static inline int xlp9xx_get_socbus(int node)
node              207 arch/mips/include/asm/netlogic/xlp-hal/iomap.h 	if (node == 0)
node              209 arch/mips/include/asm/netlogic/xlp-hal/iomap.h 	socbridge = nlm_pcicfg_base(XLP9XX_IO_BRIDGE_OFFSET(node));
node              101 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h #define nlm_get_pcie_base(node, inst)	nlm_pcicfg_base(cpu_is_xlp9xx() ? \
node              102 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h 	XLP9XX_IO_PCIE_OFFSET(node, inst) : XLP_IO_PCIE_OFFSET(node, inst))
node              105 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h void xlp_init_node_msi_irqs(int node, int link);
node              107 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h static inline void xlp_init_node_msi_irqs(int node, int link) {}
node              219 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_get_pic_pcibase(node)	nlm_pcicfg_base(cpu_is_xlp9xx() ? \
node              220 arch/mips/include/asm/netlogic/xlp-hal/pic.h 		XLP9XX_IO_PIC_OFFSET(node) : XLP_IO_PIC_OFFSET(node))
node              221 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_get_pic_regbase(node) (nlm_get_pic_pcibase(node) + XLP_IO_PCI_HDRSZ)
node              196 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_sys_pcibase(node)	nlm_pcicfg_base(cpu_is_xlp9xx() ? \
node              197 arch/mips/include/asm/netlogic/xlp-hal/sys.h 		XLP9XX_IO_SYS_OFFSET(node) : XLP_IO_SYS_OFFSET(node))
node              198 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_sys_regbase(node) (nlm_get_sys_pcibase(node) + XLP_IO_PCI_HDRSZ)
node              201 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_fuse_pcibase(node)	\
node              202 arch/mips/include/asm/netlogic/xlp-hal/sys.h 			nlm_pcicfg_base(XLP9XX_IO_FUSE_OFFSET(node))
node              203 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_fuse_regbase(node)	\
node              204 arch/mips/include/asm/netlogic/xlp-hal/sys.h 			(nlm_get_fuse_pcibase(node) + XLP_IO_PCI_HDRSZ)
node              206 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_clock_pcibase(node)	\
node              207 arch/mips/include/asm/netlogic/xlp-hal/sys.h 			nlm_pcicfg_base(XLP9XX_IO_CLOCK_OFFSET(node))
node              208 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_get_clock_regbase(node)	\
node              209 arch/mips/include/asm/netlogic/xlp-hal/sys.h 			(nlm_get_clock_pcibase(node) + XLP_IO_PCI_HDRSZ)
node              211 arch/mips/include/asm/netlogic/xlp-hal/sys.h unsigned int nlm_get_pic_frequency(int node);
node               96 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_get_uart_pcibase(node, inst)	\
node               97 arch/mips/include/asm/netlogic/xlp-hal/uart.h 	nlm_pcicfg_base(cpu_is_xlp9xx() ?  XLP9XX_IO_UART_OFFSET(node) : \
node               98 arch/mips/include/asm/netlogic/xlp-hal/uart.h 						XLP_IO_UART_OFFSET(node, inst))
node               99 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_get_uart_regbase(node, inst)	\
node              100 arch/mips/include/asm/netlogic/xlp-hal/uart.h 			(nlm_get_uart_pcibase(node, inst) + XLP_IO_PCI_HDRSZ)
node               93 arch/mips/include/asm/netlogic/xlp-hal/xlp.h int nlm_get_dram_map(int node, uint64_t *dram_map, int nentries);
node              367 arch/mips/include/asm/octeon/cvmx.h static inline void cvmx_write_csr_node(uint64_t node, uint64_t csr_addr,
node              372 arch/mips/include/asm/octeon/cvmx.h 	node_addr = (node & CVMX_NODE_MASK) << CVMX_NODE_IO_SHIFT;
node              380 arch/mips/include/asm/octeon/cvmx.h static inline uint64_t cvmx_read_csr_node(uint64_t node, uint64_t csr_addr)
node              385 arch/mips/include/asm/octeon/cvmx.h 		    (node & CVMX_NODE_MASK) << CVMX_NODE_IO_SHIFT;
node              313 arch/mips/include/asm/octeon/octeon.h 			 struct device_node *node,
node              364 arch/mips/include/asm/octeon/octeon.h struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block);
node               74 arch/mips/include/asm/pci.h 			       struct device_node *node);
node               77 arch/mips/include/asm/pci.h 				      struct device_node *node) {}
node              680 arch/mips/include/asm/r4kcache.h static inline void blast_##pfx##cache##lsize##_node(long node)		\
node              682 arch/mips/include/asm/r4kcache.h 	unsigned long start = CAC_BASE | nid_to_addrbase(node);		\
node              130 arch/mips/include/asm/sn/sn0/hubio.h #define IIO_IGFX_INIT(widget, node, cpu, valid)				(\
node              132 arch/mips/include/asm/sn/sn0/hubio.h 	(((node)   & IIO_IGFX_N_NUM_MASK) << IIO_IGFX_N_NUM_SHIFT) |	 \
node              915 arch/mips/include/asm/sn/sn0/hubio.h 		    node  : 9,
node              216 arch/mips/kernel/relocate.c 		int node, len;
node              219 arch/mips/kernel/relocate.c 		node = fdt_path_offset(initial_boot_params, "/chosen");
node              220 arch/mips/kernel/relocate.c 		if (node >= 0) {
node              221 arch/mips/kernel/relocate.c 			prop = fdt_getprop_w(initial_boot_params, node,
node              238 arch/mips/kernel/smp.c 	struct device_node *node;
node              240 arch/mips/kernel/smp.c 	node = of_irq_find_parent(of_root);
node              241 arch/mips/kernel/smp.c 	ipidomain = irq_find_matching_host(node, DOMAIN_BUS_IPI);
node              248 arch/mips/kernel/smp.c 	if (node && !ipidomain)
node              295 arch/mips/kernel/smp.c 	struct device_node *node;
node              297 arch/mips/kernel/smp.c 	node = of_irq_find_parent(of_root);
node              298 arch/mips/kernel/smp.c 	ipidomain = irq_find_matching_host(node, DOMAIN_BUS_IPI);
node              305 arch/mips/kernel/smp.c 	if (node && !ipidomain)
node              337 arch/mips/lantiq/irq.c int __init icu_of_init(struct device_node *node, struct device_node *parent)
node              345 arch/mips/lantiq/irq.c 		if (of_address_to_resource(node, vpe, &res))
node              379 arch/mips/lantiq/irq.c 	ltq_domain = irq_domain_add_linear(node,
node               60 arch/mips/loongson64/lemote-2f/clock.c 	list_for_each_entry(clkp, &clock_list, node) {
node              119 arch/mips/loongson64/loongson-3/numa.c static void __init szmem(unsigned int node)
node              128 arch/mips/loongson64/loongson-3/numa.c 		if (node_id != node)
node              146 arch/mips/loongson64/loongson-3/numa.c 				PFN_PHYS(end_pfn - start_pfn), node);
node              158 arch/mips/loongson64/loongson-3/numa.c 				PFN_PHYS(end_pfn - start_pfn), node);
node              170 arch/mips/loongson64/loongson-3/numa.c static void __init node_mem_init(unsigned int node)
node              175 arch/mips/loongson64/loongson-3/numa.c 	node_addrspace_offset = nid_to_addroffset(node);
node              177 arch/mips/loongson64/loongson-3/numa.c 			node, node_addrspace_offset);
node              179 arch/mips/loongson64/loongson-3/numa.c 	get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
node              181 arch/mips/loongson64/loongson-3/numa.c 		node, start_pfn, end_pfn);
node              183 arch/mips/loongson64/loongson-3/numa.c 	__node_data[node] = prealloc__node_data + node;
node              185 arch/mips/loongson64/loongson-3/numa.c 	NODE_DATA(node)->node_start_pfn = start_pfn;
node              186 arch/mips/loongson64/loongson-3/numa.c 	NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
node              188 arch/mips/loongson64/loongson-3/numa.c 	if (node == 0) {
node              208 arch/mips/loongson64/loongson-3/numa.c 	unsigned int node, cpu, active_cpu = 0;
node              213 arch/mips/loongson64/loongson-3/numa.c 	for (node = 0; node < loongson_sysconf.nr_nodes; node++) {
node              214 arch/mips/loongson64/loongson-3/numa.c 		if (node_online(node)) {
node              215 arch/mips/loongson64/loongson-3/numa.c 			szmem(node);
node              216 arch/mips/loongson64/loongson-3/numa.c 			node_mem_init(node);
node              217 arch/mips/loongson64/loongson-3/numa.c 			cpumask_clear(&__node_data[(node)]->cpumask);
node              224 arch/mips/loongson64/loongson-3/numa.c 		node = cpu / loongson_sysconf.cores_per_node;
node              225 arch/mips/loongson64/loongson-3/numa.c 		if (node >= num_online_nodes())
node              226 arch/mips/loongson64/loongson-3/numa.c 			node = 0;
node              231 arch/mips/loongson64/loongson-3/numa.c 		cpumask_set_cpu(active_cpu, &__node_data[(node)]->cpumask);
node              232 arch/mips/loongson64/loongson-3/numa.c 		pr_info("NUMA: set cpumask cpu %d on node %d\n", active_cpu, node);
node              456 arch/mips/loongson64/loongson-3/smp.c 	register long cpuid, core, node, count;
node              508 arch/mips/loongson64/loongson-3/smp.c 		: [core] "=&r" (core), [node] "=&r" (node),
node              518 arch/mips/loongson64/loongson-3/smp.c 	register long cpuid, core, node, count;
node              591 arch/mips/loongson64/loongson-3/smp.c 		: [core] "=&r" (core), [node] "=&r" (node),
node              601 arch/mips/loongson64/loongson-3/smp.c 	register long cpuid, core, node, count;
node              655 arch/mips/loongson64/loongson-3/smp.c 		: [core] "=&r" (core), [node] "=&r" (node),
node              462 arch/mips/mm/c-r4k.c static void (*r4k_blast_scache_node)(long node);
node              210 arch/mips/mti-malta/malta-time.c 	struct device_node *node;
node              214 arch/mips/mti-malta/malta-time.c 	node = of_find_compatible_node(NULL, NULL, "mti,gic-timer");
node              215 arch/mips/mti-malta/malta-time.c 	if (!node) {
node              220 arch/mips/mti-malta/malta-time.c 	if (of_update_property(node, &gic_frequency_prop) < 0)
node               81 arch/mips/netlogic/common/irq.c 	struct	nlm_soc_info *node;
node               93 arch/mips/netlogic/common/irq.c 	spin_lock_irqsave(&pd->node->piclock, flags);
node               94 arch/mips/netlogic/common/irq.c 	nlm_pic_enable_irt(pd->node->picbase, pd->irt);
node               95 arch/mips/netlogic/common/irq.c 	spin_unlock_irqrestore(&pd->node->piclock, flags);
node              104 arch/mips/netlogic/common/irq.c 	spin_lock_irqsave(&pd->node->piclock, flags);
node              105 arch/mips/netlogic/common/irq.c 	nlm_pic_disable_irt(pd->node->picbase, pd->irt);
node              106 arch/mips/netlogic/common/irq.c 	spin_unlock_irqrestore(&pd->node->piclock, flags);
node              130 arch/mips/netlogic/common/irq.c 	nlm_pic_ack(pd->node->picbase, pd->irt);
node              184 arch/mips/netlogic/common/irq.c void nlm_setup_pic_irq(int node, int picirq, int irq, int irt)
node              189 arch/mips/netlogic/common/irq.c 	xirq = nlm_irq_to_xirq(node, irq);
node              194 arch/mips/netlogic/common/irq.c 	pic_data->node = nlm_get_node(node);
node              199 arch/mips/netlogic/common/irq.c void nlm_set_pic_extra_ack(int node, int irq, void (*xack)(struct irq_data *))
node              204 arch/mips/netlogic/common/irq.c 	xirq = nlm_irq_to_xirq(node, irq);
node              211 arch/mips/netlogic/common/irq.c static void nlm_init_node_irqs(int node)
node              216 arch/mips/netlogic/common/irq.c 	pr_info("Init IRQ for node %d\n", node);
node              217 arch/mips/netlogic/common/irq.c 	nodep = nlm_get_node(node);
node              228 arch/mips/netlogic/common/irq.c 				node * nlm_threads_per_node(), 0);
node              229 arch/mips/netlogic/common/irq.c 		nlm_setup_pic_irq(node, i, i, irt);
node              235 arch/mips/netlogic/common/irq.c 	int cpu, node;
node              238 arch/mips/netlogic/common/irq.c 	node = hwtid / nlm_threads_per_node();
node              240 arch/mips/netlogic/common/irq.c 	if (cpu == 0 && node != 0)
node              241 arch/mips/netlogic/common/irq.c 		nlm_init_node_irqs(node);
node              242 arch/mips/netlogic/common/irq.c 	write_c0_eimr(nlm_get_node(node)->irqmask);
node              248 arch/mips/netlogic/common/irq.c 	int i, node;
node              250 arch/mips/netlogic/common/irq.c 	node = nlm_nodeid();
node              265 arch/mips/netlogic/common/irq.c 		nlm_dispatch_msi(node, i);
node              269 arch/mips/netlogic/common/irq.c 		nlm_dispatch_msix(node, i);
node              275 arch/mips/netlogic/common/irq.c 	do_IRQ(nlm_irq_to_xirq(node, i));
node              283 arch/mips/netlogic/common/irq.c static int __init xlp_of_pic_init(struct device_node *node,
node              292 arch/mips/netlogic/common/irq.c 	ret = of_address_to_resource(node, 0, &res);
node              294 arch/mips/netlogic/common/irq.c 		pr_err("PIC %pOFn: reg property not found!\n", node);
node              308 arch/mips/netlogic/common/irq.c 					node, bus);
node              315 arch/mips/netlogic/common/irq.c 							node, socid);
node              321 arch/mips/netlogic/common/irq.c 		pr_err("PIC %pOFn: node %d does not exist!\n", node, socid);
node              325 arch/mips/netlogic/common/irq.c 	xlp_pic_domain = irq_domain_add_legacy(node, n_picirqs,
node              329 arch/mips/netlogic/common/irq.c 		pr_err("PIC %pOFn: Creating legacy domain failed!\n", node);
node              171 arch/mips/netlogic/common/smp.c 	int num_cpus, i, ncore, node;
node              193 arch/mips/netlogic/common/smp.c 			node = nlm_hwtid_to_node(i);
node              194 arch/mips/netlogic/common/smp.c 			cpumask_set_cpu(num_cpus, &nlm_get_node(node)->cpumask);
node              145 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_get_sata_pcibase(node)	\
node              146 arch/mips/netlogic/xlp/ahci-init-xlp2.c 		nlm_pcicfg_base(XLP9XX_IO_SATA_OFFSET(node))
node              147 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_get_sata_regbase(node)	\
node              148 arch/mips/netlogic/xlp/ahci-init-xlp2.c 		(nlm_get_sata_pcibase(node) + 0x100)
node              253 arch/mips/netlogic/xlp/ahci-init-xlp2.c static void nlm_sata_firmware_init(int node)
node              260 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	regbase = nlm_get_sata_regbase(node);
node              335 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	int node;
node              339 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	for (node = 0; node < NLM_NR_NODES; node++)
node              340 arch/mips/netlogic/xlp/ahci-init-xlp2.c 		if (nlm_node_present(node))
node              341 arch/mips/netlogic/xlp/ahci-init-xlp2.c 			nlm_sata_firmware_init(node);
node              349 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	int node;
node              351 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	node = data->irq / NLM_IRQS_PER_NODE;
node              352 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	regbase = nlm_get_sata_regbase(node);
node              367 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	int node;
node              370 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	node = xlp_socdev_to_node(dev);
node              371 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	regbase = nlm_get_sata_regbase(node);
node              380 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	dev->irq = nlm_irq_to_xirq(node, PIC_SATA_IRQ);
node              381 arch/mips/netlogic/xlp/ahci-init-xlp2.c 	nlm_set_pic_extra_ack(node, PIC_SATA_IRQ, nlm_sata_intr_ack);
node               89 arch/mips/netlogic/xlp/ahci-init.c #define nlm_get_sata_pcibase(node)	\
node               90 arch/mips/netlogic/xlp/ahci-init.c 		nlm_pcicfg_base(XLP_IO_SATA_OFFSET(node))
node               92 arch/mips/netlogic/xlp/ahci-init.c #define nlm_get_sata_regbase(node)	\
node               93 arch/mips/netlogic/xlp/ahci-init.c 		(nlm_get_sata_pcibase(node) + 0x900)
node              111 arch/mips/netlogic/xlp/ahci-init.c static void nlm_sata_firmware_init(int node)
node              118 arch/mips/netlogic/xlp/ahci-init.c 	regbase = nlm_get_sata_regbase(node);
node              153 arch/mips/netlogic/xlp/ahci-init.c 	int node = 0;
node              157 arch/mips/netlogic/xlp/ahci-init.c 		nlm_sata_firmware_init(node);
node              185 arch/mips/netlogic/xlp/ahci-init.c 	int node = 0; /* XLP3XX does not support multi-node */
node              187 arch/mips/netlogic/xlp/ahci-init.c 	regbase = nlm_get_sata_regbase(node);
node              201 arch/mips/netlogic/xlp/ahci-init.c 	nlm_set_pic_extra_ack(node, PIC_SATA_IRQ, nlm_sata_intr_ack);
node               52 arch/mips/netlogic/xlp/nlm_hal.c void nlm_node_init(int node)
node               56 arch/mips/netlogic/xlp/nlm_hal.c 	nodep = nlm_get_node(node);
node               57 arch/mips/netlogic/xlp/nlm_hal.c 	if (node == 0)
node               59 arch/mips/netlogic/xlp/nlm_hal.c 	nodep->sysbase = nlm_get_sys_regbase(node);
node               60 arch/mips/netlogic/xlp/nlm_hal.c 	nodep->picbase = nlm_get_pic_regbase(node);
node               63 arch/mips/netlogic/xlp/nlm_hal.c 		nodep->socbus = xlp9xx_get_socbus(node);
node              228 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp2_get_core_frequency(int node, int core)
node              234 arch/mips/netlogic/xlp/nlm_hal.c 		clockbase = nlm_get_clock_regbase(node);
node              240 arch/mips/netlogic/xlp/nlm_hal.c 		sysbase = nlm_get_node(node)->sysbase;
node              274 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp_get_core_frequency(int node, int core)
node              280 arch/mips/netlogic/xlp/nlm_hal.c 	sysbase = nlm_get_node(node)->sysbase;
node              295 arch/mips/netlogic/xlp/nlm_hal.c unsigned int nlm_get_core_frequency(int node, int core)
node              298 arch/mips/netlogic/xlp/nlm_hal.c 		return nlm_xlp2_get_core_frequency(node, core);
node              300 arch/mips/netlogic/xlp/nlm_hal.c 		return nlm_xlp_get_core_frequency(node, core);
node              308 arch/mips/netlogic/xlp/nlm_hal.c static unsigned int nlm_xlp2_get_pic_frequency(int node)
node              314 arch/mips/netlogic/xlp/nlm_hal.c 	sysbase = nlm_get_node(node)->sysbase;
node              315 arch/mips/netlogic/xlp/nlm_hal.c 	clockbase = nlm_get_clock_regbase(node);
node              449 arch/mips/netlogic/xlp/nlm_hal.c unsigned int nlm_get_pic_frequency(int node)
node              452 arch/mips/netlogic/xlp/nlm_hal.c 		return nlm_xlp2_get_pic_frequency(node);
node              466 arch/mips/netlogic/xlp/nlm_hal.c int nlm_get_dram_map(int node, uint64_t *dram_map, int nentries)
node              488 arch/mips/netlogic/xlp/nlm_hal.c 		if (node >= 0) {
node              492 arch/mips/netlogic/xlp/nlm_hal.c 			if (n != node)
node               87 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_xlpii_get_usb_pcibase(node, inst)			\
node               89 arch/mips/netlogic/xlp/usb-init-xlp2.c 			XLP9XX_IO_USB_OFFSET(node, inst) :	\
node               90 arch/mips/netlogic/xlp/usb-init-xlp2.c 			XLP2XX_IO_USB_OFFSET(node, inst))
node               91 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_xlpii_get_usb_regbase(node, inst)		\
node               92 arch/mips/netlogic/xlp/usb-init-xlp2.c 	(nlm_xlpii_get_usb_pcibase(node, inst) + XLP_IO_PCI_HDRSZ)
node              118 arch/mips/netlogic/xlp/usb-init-xlp2.c 	int node, irq;
node              122 arch/mips/netlogic/xlp/usb-init-xlp2.c 	node = data->irq / NLM_IRQS_PER_NODE;
node              126 arch/mips/netlogic/xlp/usb-init-xlp2.c 		port_addr = nlm_xlpii_get_usb_regbase(node, 1);
node              129 arch/mips/netlogic/xlp/usb-init-xlp2.c 		port_addr = nlm_xlpii_get_usb_regbase(node, 2);
node              132 arch/mips/netlogic/xlp/usb-init-xlp2.c 		port_addr = nlm_xlpii_get_usb_regbase(node, 3);
node              135 arch/mips/netlogic/xlp/usb-init-xlp2.c 		pr_err("No matching USB irq %d node  %d!\n", irq, node);
node              141 arch/mips/netlogic/xlp/usb-init-xlp2.c static void nlm_xlpii_usb_hw_reset(int node, int port)
node              147 arch/mips/netlogic/xlp/usb-init-xlp2.c 	port_addr = nlm_xlpii_get_usb_regbase(node, port);
node              180 arch/mips/netlogic/xlp/usb-init-xlp2.c 	pci_base = nlm_xlpii_get_usb_pcibase(node, port);
node              210 arch/mips/netlogic/xlp/usb-init-xlp2.c 	int node;
node              229 arch/mips/netlogic/xlp/usb-init-xlp2.c 	for (node = 0; node < NLM_NR_NODES; node++) {
node              230 arch/mips/netlogic/xlp/usb-init-xlp2.c 		if (!nlm_node_present(node))
node              232 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_xlpii_usb_hw_reset(node, 1);
node              233 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_xlpii_usb_hw_reset(node, 2);
node              234 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_xlpii_usb_hw_reset(node, 3);
node              235 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_0_IRQ, xlp9xx_usb_ack);
node              236 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_1_IRQ, xlp9xx_usb_ack);
node              237 arch/mips/netlogic/xlp/usb-init-xlp2.c 		nlm_set_pic_extra_ack(node, PIC_9XX_XHCI_2_IRQ, xlp9xx_usb_ack);
node              249 arch/mips/netlogic/xlp/usb-init-xlp2.c 	int node;
node              251 arch/mips/netlogic/xlp/usb-init-xlp2.c 	node = xlp_socdev_to_node(dev);
node              256 arch/mips/netlogic/xlp/usb-init-xlp2.c 		dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_0_IRQ);
node              259 arch/mips/netlogic/xlp/usb-init-xlp2.c 		dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_1_IRQ);
node              262 arch/mips/netlogic/xlp/usb-init-xlp2.c 		dev->irq = nlm_irq_to_xirq(node, PIC_9XX_XHCI_2_IRQ);
node               65 arch/mips/netlogic/xlp/usb-init.c #define nlm_get_usb_pcibase(node, inst)		\
node               66 arch/mips/netlogic/xlp/usb-init.c 	nlm_pcicfg_base(XLP_IO_USB_OFFSET(node, inst))
node               67 arch/mips/netlogic/xlp/usb-init.c #define nlm_get_usb_regbase(node, inst)		\
node               68 arch/mips/netlogic/xlp/usb-init.c 	(nlm_get_usb_pcibase(node, inst) + XLP_IO_PCI_HDRSZ)
node               70 arch/mips/netlogic/xlp/usb-init.c static void nlm_usb_intr_en(int node, int port)
node               75 arch/mips/netlogic/xlp/usb-init.c 	port_addr = nlm_get_usb_regbase(node, port);
node               82 arch/mips/netlogic/xlp/usb-init.c static void nlm_usb_hw_reset(int node, int port)
node               88 arch/mips/netlogic/xlp/usb-init.c 	port_addr = nlm_get_usb_regbase(node, port);
node               53 arch/mips/netlogic/xlp/wakeup.c static int xlp_wakeup_core(uint64_t sysbase, int node, int core)
node              113 arch/mips/pci/msi-xlp.c 	struct nlm_soc_info *node;
node              227 arch/mips/pci/msi-xlp.c 		nlm_pic_ack(md->node->picbase,
node              293 arch/mips/pci/msi-xlp.c static int xlp_setup_msi(uint64_t lnkbase, int node, int link,
node              304 arch/mips/pci/msi-xlp.c 	xirq = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0));
node              306 arch/mips/pci/msi-xlp.c 	msiaddr = MSI_LINK_ADDR(node, link);
node              316 arch/mips/pci/msi-xlp.c 		nlm_setup_pic_irq(node, lirq, lirq, irt);
node              317 arch/mips/pci/msi-xlp.c 		nlm_pic_init_irt(nlm_get_node(node)->picbase, irt, lirq,
node              318 arch/mips/pci/msi-xlp.c 				 node * nlm_threads_per_node(), 1 /*en */);
node              400 arch/mips/pci/msi-xlp.c static int xlp_setup_msix(uint64_t lnkbase, int node, int link,
node              411 arch/mips/pci/msi-xlp.c 	xirq = nlm_irq_to_xirq(node, nlm_link_msixirq(link, 0));
node              413 arch/mips/pci/msi-xlp.c 	msixaddr = MSIX_LINK_ADDR(node, link);
node              448 arch/mips/pci/msi-xlp.c 	int node, link, slot;
node              457 arch/mips/pci/msi-xlp.c 	node = slot / 8;
node              458 arch/mips/pci/msi-xlp.c 	lnkbase = nlm_get_pcie_base(node, link);
node              461 arch/mips/pci/msi-xlp.c 		return xlp_setup_msix(lnkbase, node, link, desc);
node              463 arch/mips/pci/msi-xlp.c 		return xlp_setup_msi(lnkbase, node, link, desc);
node              466 arch/mips/pci/msi-xlp.c void __init xlp_init_node_msi_irqs(int node, int link)
node              472 arch/mips/pci/msi-xlp.c 	pr_info("[%d %d] Init node PCI IRT\n", node, link);
node              473 arch/mips/pci/msi-xlp.c 	nodep = nlm_get_node(node);
node              481 arch/mips/pci/msi-xlp.c 	md->node = nodep;
node              482 arch/mips/pci/msi-xlp.c 	md->lnkbase = nlm_get_pcie_base(node, link);
node              485 arch/mips/pci/msi-xlp.c 	irq = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0));
node              493 arch/mips/pci/msi-xlp.c 			val = ((node * nlm_threads_per_node()) << 7 |
node              505 arch/mips/pci/msi-xlp.c 					node * nlm_threads_per_node(), 1);
node              509 arch/mips/pci/msi-xlp.c 		irq = nlm_irq_to_xirq(node, nlm_link_msixirq(link, i));
node              515 arch/mips/pci/msi-xlp.c void nlm_dispatch_msi(int node, int lirq)
node              522 arch/mips/pci/msi-xlp.c 	irqbase = nlm_irq_to_xirq(node, nlm_link_msiirq(link, 0));
node              539 arch/mips/pci/msi-xlp.c 		nlm_pic_ack(md->node->picbase,
node              542 arch/mips/pci/msi-xlp.c 		nlm_pic_ack(md->node->picbase, PIC_IRT_PCIE_LINK_INDEX(link));
node              545 arch/mips/pci/msi-xlp.c void nlm_dispatch_msix(int node, int lirq)
node              552 arch/mips/pci/msi-xlp.c 	irqbase = nlm_irq_to_xirq(node, nlm_link_msixirq(link, 0));
node               95 arch/mips/pci/pci-lantiq.c 	struct device_node *node = pdev->dev.of_node;
node              114 arch/mips/pci/pci-lantiq.c 	bus_clk = of_get_property(node, "lantiq,bus-clock", NULL);
node              120 arch/mips/pci/pci-lantiq.c 	if (of_find_property(node, "lantiq,external-clock", NULL))
node              126 arch/mips/pci/pci-lantiq.c 	reset_gpio = of_get_named_gpio(node, "gpio-reset", 0);
node              150 arch/mips/pci/pci-lantiq.c 	req_mask = of_get_property(node, "req-mask", NULL);
node              131 arch/mips/pci/pci-legacy.c 		list_for_each_entry(child, &bus->children, node)
node              138 arch/mips/pci/pci-legacy.c void pci_load_of_ranges(struct pci_controller *hose, struct device_node *node)
node              143 arch/mips/pci/pci-legacy.c 	pr_info("PCI host bridge %pOF ranges:\n", node);
node              144 arch/mips/pci/pci-legacy.c 	hose->of_node = node;
node              146 arch/mips/pci/pci-legacy.c 	if (of_pci_range_parser_init(&parser, node))
node              170 arch/mips/pci/pci-legacy.c 			of_pci_range_to_resource(&range, node, res);
node              211 arch/mips/pci/pci-xlp.c 	int lnkfunc, node;
node              222 arch/mips/pci/pci-xlp.c 	node = xlp_socdev_to_node(lnkdev);
node              224 arch/mips/pci/pci-xlp.c 	return nlm_irq_to_xirq(node, PIC_PCIE_LINK_LEGACY_IRQ(lnkfunc));
node              239 arch/mips/pci/pci-xlp.c static void xlp_config_pci_bswap(int node, int link)
node              244 arch/mips/pci/pci-xlp.c 	nbubase = nlm_get_bridge_regbase(node);
node              245 arch/mips/pci/pci-xlp.c 	lnkbase = nlm_get_pcie_base(node, link);
node              286 arch/mips/pci/pci-xlp.c static inline void xlp_config_pci_bswap(int node, int link) {}
node               26 arch/mips/pic32/pic32mzda/time.c 	static struct device_node *node;
node               29 arch/mips/pic32/pic32mzda/time.c 	node = of_find_matching_node(NULL, pic32_infra_match);
node               31 arch/mips/pic32/pic32mzda/time.c 	if (WARN_ON(!node))
node               34 arch/mips/pic32/pic32mzda/time.c 	irq = irq_of_parse_and_map(node, 0);
node               32 arch/mips/ralink/common.h __iomem void *plat_of_remap_node(const char *node);
node              146 arch/mips/ralink/irq.c static int __init intc_of_init(struct device_node *node,
node              153 arch/mips/ralink/irq.c 	if (!of_property_read_u32_array(node, "ralink,intc-registers",
node              157 arch/mips/ralink/irq.c 	irq = irq_of_parse_and_map(node, 0);
node              161 arch/mips/ralink/irq.c 	if (of_address_to_resource(node, 0, &res))
node              179 arch/mips/ralink/irq.c 	domain = irq_domain_add_legacy(node, RALINK_INTC_IRQ_COUNT,
node               29 arch/mips/ralink/of.c __iomem void *plat_of_remap_node(const char *node)
node               34 arch/mips/ralink/of.c 	np = of_find_compatible_node(NULL, NULL, node);
node               36 arch/mips/ralink/of.c 		panic("Failed to find %s node", node);
node               39 arch/mips/ralink/of.c 		panic("Failed to get resource for %s", node);
node               44 arch/mips/ralink/of.c 		panic("Failed to request resources for %s", node);
node               56 arch/mips/ralink/of.c static int __init early_init_dt_find_memory(unsigned long node,
node              151 arch/mips/sgi-ip27/ip27-irq.c 	desc->irq_common_data.node = info->nasid;
node              264 arch/mips/sgi-ip27/ip27-memory.c static unsigned long __init slot_psize_compute(cnodeid_t node, int slot)
node              271 arch/mips/sgi-ip27/ip27-memory.c 	nasid = COMPACT_TO_NASID_NODEID(node);
node              353 arch/mips/sgi-ip27/ip27-memory.c 	cnodeid_t node;
node              355 arch/mips/sgi-ip27/ip27-memory.c 	for_each_online_node(node) {
node              358 arch/mips/sgi-ip27/ip27-memory.c 			slot_psize = slot_psize_compute(node, slot);
node              373 arch/mips/sgi-ip27/ip27-memory.c 								slot, node);
node              377 arch/mips/sgi-ip27/ip27-memory.c 			memblock_add_node(PFN_PHYS(slot_getbasepfn(node, slot)),
node              378 arch/mips/sgi-ip27/ip27-memory.c 					  PFN_PHYS(slot_psize), node);
node              383 arch/mips/sgi-ip27/ip27-memory.c static void __init node_mem_init(cnodeid_t node)
node              385 arch/mips/sgi-ip27/ip27-memory.c 	unsigned long slot_firstpfn = slot_getbasepfn(node, 0);
node              386 arch/mips/sgi-ip27/ip27-memory.c 	unsigned long slot_freepfn = node_getfirstfree(node);
node              389 arch/mips/sgi-ip27/ip27-memory.c 	get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
node              394 arch/mips/sgi-ip27/ip27-memory.c 	__node_data[node] = __va(slot_freepfn << PAGE_SHIFT);
node              395 arch/mips/sgi-ip27/ip27-memory.c 	memset(__node_data[node], 0, PAGE_SIZE);
node              397 arch/mips/sgi-ip27/ip27-memory.c 	NODE_DATA(node)->node_start_pfn = start_pfn;
node              398 arch/mips/sgi-ip27/ip27-memory.c 	NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
node              400 arch/mips/sgi-ip27/ip27-memory.c 	cpumask_clear(&hub_data(node)->h_cpus);
node              405 arch/mips/sgi-ip27/ip27-memory.c 	free_bootmem_with_active_regions(node, end_pfn);
node              410 arch/mips/sgi-ip27/ip27-memory.c 	sparse_memory_present_with_active_regions(node);
node              430 arch/mips/sgi-ip27/ip27-memory.c 	cnodeid_t node;
node              436 arch/mips/sgi-ip27/ip27-memory.c 	for (node = 0; node < MAX_COMPACT_NODES; node++) {
node              437 arch/mips/sgi-ip27/ip27-memory.c 		if (node_online(node)) {
node              438 arch/mips/sgi-ip27/ip27-memory.c 			node_mem_init(node);
node              441 arch/mips/sgi-ip27/ip27-memory.c 		__node_data[node] = &null_node;
node              214 arch/mips/sgi-ip27/ip27-nmi.c 		for_each_online_node(node)
node              215 arch/mips/sgi-ip27/ip27-nmi.c 			if (NODEPDA(node)->dump_count == 0)
node              217 arch/mips/sgi-ip27/ip27-nmi.c 		if (node == MAX_NUMNODES)
node              220 arch/mips/sgi-ip27/ip27-nmi.c 			for_each_online_node(node)
node              221 arch/mips/sgi-ip27/ip27-nmi.c 				if (NODEPDA(node)->dump_count == 0) {
node              222 arch/mips/sgi-ip27/ip27-nmi.c 					cpu = cpumask_first(cpumask_of_node(node));
node              223 arch/mips/sgi-ip27/ip27-nmi.c 					for (n=0; n < CNODE_NUM_CPUS(node); cpu++, n++) {
node               36 arch/mips/sgi-ip27/ip27-smp.c 	cnodeid_t node = get_cpu_cnode(cpu);
node               37 arch/mips/sgi-ip27/ip27-smp.c 	nasid_t nasid = COMPACT_TO_NASID_NODEID(node);
node               40 arch/mips/sgi-ip27/ip27-smp.c 	sn_cpu_info[cpunum].p_nodeid = node;
node             1132 arch/nds32/kernel/perf_event_cpu.c 	struct device_node *node = pdev->dev.of_node;
node             1146 arch/nds32/kernel/perf_event_cpu.c 	if (node && of_id) {
node               65 arch/nios2/kernel/irq.c 	struct device_node *node;
node               67 arch/nios2/kernel/irq.c 	node = of_find_compatible_node(NULL, NULL, "altr,nios2-1.0");
node               68 arch/nios2/kernel/irq.c 	if (!node)
node               69 arch/nios2/kernel/irq.c 		node = of_find_compatible_node(NULL, NULL, "altr,nios2-1.1");
node               71 arch/nios2/kernel/irq.c 	BUG_ON(!node);
node               73 arch/nios2/kernel/irq.c 	domain = irq_domain_add_linear(node, NIOS2_CPU_NR_IRQS, &irq_ops, NULL);
node               77 arch/nios2/kernel/irq.c 	of_node_put(node);
node               22 arch/powerpc/boot/cuboot-8xx.c 	void *node;
node               28 arch/powerpc/boot/cuboot-8xx.c 	node = finddevice("/soc/cpm");
node               29 arch/powerpc/boot/cuboot-8xx.c 	if (node)
node               30 arch/powerpc/boot/cuboot-8xx.c 		setprop(node, "clock-frequency", &bd.bi_busfreq, 4);
node               32 arch/powerpc/boot/cuboot-8xx.c 	node = finddevice("/soc/cpm/brg");
node               33 arch/powerpc/boot/cuboot-8xx.c 	if (node)
node               34 arch/powerpc/boot/cuboot-8xx.c 		setprop(node, "clock-frequency",  &bd.bi_busfreq, 4);
node              127 arch/powerpc/boot/cuboot-pq2.c 	void *node, *parent_node;
node              130 arch/powerpc/boot/cuboot-pq2.c 	node = finddevice("/pci");
node              131 arch/powerpc/boot/cuboot-pq2.c 	if (!node || !dt_is_compatible(node, "fsl,pq2-pci"))
node              135 arch/powerpc/boot/cuboot-pq2.c 		if (!dt_xlate_reg(node, i,
node              143 arch/powerpc/boot/cuboot-pq2.c 	dt_get_reg_format(node, &naddr, &nsize);
node              147 arch/powerpc/boot/cuboot-pq2.c 	parent_node = get_parent(node);
node              155 arch/powerpc/boot/cuboot-pq2.c 	len = getprop(node, "ranges", pci_ranges_buf,
node              244 arch/powerpc/boot/cuboot-pq2.c 	void *node;
node              250 arch/powerpc/boot/cuboot-pq2.c 	node = finddevice("/soc/cpm");
node              251 arch/powerpc/boot/cuboot-pq2.c 	if (node)
node              252 arch/powerpc/boot/cuboot-pq2.c 		setprop(node, "clock-frequency", &bd.bi_cpmfreq, 4);
node              254 arch/powerpc/boot/cuboot-pq2.c 	node = finddevice("/soc/cpm/brg");
node              255 arch/powerpc/boot/cuboot-pq2.c 	if (node)
node              256 arch/powerpc/boot/cuboot-pq2.c 		setprop(node, "clock-frequency",  &bd.bi_brgfreq, 4);
node              132 arch/powerpc/boot/devtree.c void dt_get_reg_format(void *node, u32 *naddr, u32 *nsize)
node              134 arch/powerpc/boot/devtree.c 	if (getprop(node, "#address-cells", naddr, 4) != 4)
node              136 arch/powerpc/boot/devtree.c 	if (getprop(node, "#size-cells", nsize, 4) != 4)
node              228 arch/powerpc/boot/devtree.c static int dt_xlate(void *node, int res, int reglen, unsigned long *addr,
node              238 arch/powerpc/boot/devtree.c 	parent = get_parent(node);
node              264 arch/powerpc/boot/devtree.c 		node = parent;
node              266 arch/powerpc/boot/devtree.c 		parent = get_parent(node);
node              272 arch/powerpc/boot/devtree.c 		buflen = getprop(node, "ranges", prop_buf,
node              313 arch/powerpc/boot/devtree.c int dt_xlate_reg(void *node, int res, unsigned long *addr, unsigned long *size)
node              317 arch/powerpc/boot/devtree.c 	reglen = getprop(node, "reg", prop_buf, sizeof(prop_buf)) / 4;
node              318 arch/powerpc/boot/devtree.c 	return dt_xlate(node, res, reglen, addr, size);
node              321 arch/powerpc/boot/devtree.c int dt_xlate_addr(void *node, u32 *buf, int buflen, unsigned long *xlated_addr)
node              328 arch/powerpc/boot/devtree.c 	return dt_xlate(node, 0, buflen / 4, xlated_addr, NULL);
node              331 arch/powerpc/boot/devtree.c int dt_is_compatible(void *node, const char *compat)
node              336 arch/powerpc/boot/devtree.c 	len = getprop(node, "compatible", buf, MAX_PROP_LEN);
node              350 arch/powerpc/boot/devtree.c int dt_get_virtual_reg(void *node, void **addr, int nres)
node              355 arch/powerpc/boot/devtree.c 	n = getprop(node, "virtual-reg", addr, nres * 4);
node              360 arch/powerpc/boot/devtree.c 		if (!dt_xlate_reg(node, n, &xaddr, NULL))
node               57 arch/powerpc/boot/mpc8xx.c 	void *node;
node               61 arch/powerpc/boot/mpc8xx.c 	node = finddevice("/soc/cpm");
node               62 arch/powerpc/boot/mpc8xx.c 	if (node)
node               63 arch/powerpc/boot/mpc8xx.c 		setprop(node, "clock-frequency", &sysclk, 4);
node               65 arch/powerpc/boot/mpc8xx.c 	node = finddevice("/soc/cpm/brg");
node               66 arch/powerpc/boot/mpc8xx.c 	if (node)
node               67 arch/powerpc/boot/mpc8xx.c 		setprop(node, "clock-frequency", &sysclk, 4);
node               96 arch/powerpc/boot/ops.h int dt_xlate_reg(void *node, int res, unsigned long *addr, unsigned long *size);
node               97 arch/powerpc/boot/ops.h int dt_xlate_addr(void *node, u32 *buf, int buflen, unsigned long *xlated_addr);
node               98 arch/powerpc/boot/ops.h int dt_is_compatible(void *node, const char *compat);
node               99 arch/powerpc/boot/ops.h void dt_get_reg_format(void *node, u32 *naddr, u32 *nsize);
node              100 arch/powerpc/boot/ops.h int dt_get_virtual_reg(void *node, void **addr, int nres);
node              108 arch/powerpc/boot/planetcore.c 	void *node, *chosen;
node              114 arch/powerpc/boot/planetcore.c 	node = find_node_by_prop_value_str(NULL, "linux,planetcore-label",
node              116 arch/powerpc/boot/planetcore.c 	if (!node)
node              119 arch/powerpc/boot/planetcore.c 	path = get_path(node, prop_buf, MAX_PROP_LEN);
node               77 arch/powerpc/boot/pq2.c 	void *node;
node               81 arch/powerpc/boot/pq2.c 	node = finddevice("/soc/cpm");
node               82 arch/powerpc/boot/pq2.c 	if (node)
node               83 arch/powerpc/boot/pq2.c 		setprop(node, "clock-frequency", &sysfreq, 4);
node               85 arch/powerpc/boot/pq2.c 	node = finddevice("/soc/cpm/brg");
node               86 arch/powerpc/boot/pq2.c 	if (node)
node               87 arch/powerpc/boot/pq2.c 		setprop(node, "clock-frequency", &brgfreq, 4);
node               24 arch/powerpc/boot/redboot-83xx.c 	void *node;
node               30 arch/powerpc/boot/redboot-83xx.c 	node = finddevice("/soc/cpm/brg");
node               31 arch/powerpc/boot/redboot-83xx.c 	if (node) {
node               34 arch/powerpc/boot/redboot-83xx.c 		setprop(node, "clock-frequency",  &bd.bi_busfreq, 4);
node               23 arch/powerpc/boot/redboot-8xx.c 	void *node;
node               29 arch/powerpc/boot/redboot-8xx.c 	node = finddevice("/soc/cpm/brg");
node               30 arch/powerpc/boot/redboot-8xx.c 	if (node) {
node               33 arch/powerpc/boot/redboot-8xx.c 		setprop(node, "clock-frequency",  &bd.bi_busfreq, 4);
node               30 arch/powerpc/boot/simpleboot.c 	int node, size, i;
node               37 arch/powerpc/boot/simpleboot.c 	node = fdt_path_offset(_dtb_start, "/");
node               38 arch/powerpc/boot/simpleboot.c 	if (node < 0)
node               40 arch/powerpc/boot/simpleboot.c 	na = fdt_getprop(_dtb_start, node, "#address-cells", &size);
node               43 arch/powerpc/boot/simpleboot.c 	ns = fdt_getprop(_dtb_start, node, "#size-cells", &size);
node               48 arch/powerpc/boot/simpleboot.c 	node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type",
node               50 arch/powerpc/boot/simpleboot.c 	if (node < 0)
node               52 arch/powerpc/boot/simpleboot.c 	reg = fdt_getprop(_dtb_start, node, "reg", &size);
node               69 arch/powerpc/boot/simpleboot.c 	node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type",
node               71 arch/powerpc/boot/simpleboot.c 	if (!node)
node               73 arch/powerpc/boot/simpleboot.c 	timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size);
node               97 arch/powerpc/boot/treeboot-akebono.c 	int node, size;
node              147 arch/powerpc/boot/treeboot-akebono.c 	node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type",
node              149 arch/powerpc/boot/treeboot-akebono.c 	if (!node)
node              151 arch/powerpc/boot/treeboot-akebono.c 	timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size);
node               84 arch/powerpc/boot/treeboot-currituck.c 	int node, size;
node              103 arch/powerpc/boot/treeboot-currituck.c 	node = fdt_node_offset_by_prop_value(_dtb_start, -1, "device_type",
node              105 arch/powerpc/boot/treeboot-currituck.c 	if (!node)
node              107 arch/powerpc/boot/treeboot-currituck.c 	timebase = fdt_getprop(_dtb_start, node, "timebase-frequency", &size);
node               87 arch/powerpc/include/asm/cell-pmu.h extern void cbe_sync_irq(int node);
node              321 arch/powerpc/include/asm/cell-regs.h extern u32 cbe_node_to_cpu(int node);
node               97 arch/powerpc/include/asm/drmem.h void __init walk_drmem_lmbs_early(unsigned long node,
node              156 arch/powerpc/include/asm/fadump-internal.h extern void rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node);
node              159 arch/powerpc/include/asm/fadump-internal.h rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) { }
node              163 arch/powerpc/include/asm/fadump-internal.h extern void opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node);
node              166 arch/powerpc/include/asm/fadump-internal.h opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node) { }
node               31 arch/powerpc/include/asm/fadump.h extern int early_init_dt_scan_fw_dump(unsigned long node, const char *uname,
node               32 arch/powerpc/include/asm/hvcserver.h 	struct list_head node;
node                8 arch/powerpc/include/asm/i8259.h extern void i8259_init(struct device_node *node, unsigned long intack_addr);
node              265 arch/powerpc/include/asm/iommu.h 				  unsigned long mask, gfp_t flag, int node);
node               72 arch/powerpc/include/asm/ipic.h extern struct ipic * ipic_init(struct device_node *node, unsigned int flags);
node              100 arch/powerpc/include/asm/kvm_book3s_64.h #define for_each_nest_rmap_safe(pos, node, rmapp)			       \
node              101 arch/powerpc/include/asm/kvm_book3s_64.h 	for ((pos) = llist_entry((node), typeof(*(pos)), list);		       \
node              102 arch/powerpc/include/asm/kvm_book3s_64.h 	     (node) &&							       \
node              103 arch/powerpc/include/asm/kvm_book3s_64.h 	     (*(rmapp) = ((RMAP_NESTED_IS_SINGLE_ENTRY & ((u64) (node))) ?     \
node              104 arch/powerpc/include/asm/kvm_book3s_64.h 			  ((u64) (node)) : ((pos)->rmap))) &&		       \
node              105 arch/powerpc/include/asm/kvm_book3s_64.h 	     (((node) = ((RMAP_NESTED_IS_SINGLE_ENTRY & ((u64) (node))) ?      \
node              108 arch/powerpc/include/asm/kvm_book3s_64.h 	     (pos) = llist_entry((node), typeof(*(pos)), list))
node              277 arch/powerpc/include/asm/mpc52xx.h extern unsigned int mpc52xx_get_xtal_freq(struct device_node *node);
node              335 arch/powerpc/include/asm/mpc52xx.h extern int __init mpc52xx_add_bridge(struct device_node *node);
node               14 arch/powerpc/include/asm/mpc5xxx.h extern unsigned long mpc5xxx_get_bus_frequency(struct device_node *node);
node              259 arch/powerpc/include/asm/mpic.h 	struct device_node *node;
node              427 arch/powerpc/include/asm/mpic.h extern struct mpic *mpic_alloc(struct device_node *node,
node              309 arch/powerpc/include/asm/opal.h extern int early_init_dt_scan_opal(unsigned long node, const char *uname,
node              311 arch/powerpc/include/asm/opal.h extern int early_init_dt_scan_recoverable_ranges(unsigned long node,
node               56 arch/powerpc/include/asm/pci-bridge.h 	int node;
node              171 arch/powerpc/include/asm/pci-bridge.h extern int pci_device_from_OF_node(struct device_node *node,
node              264 arch/powerpc/include/asm/pci-bridge.h #define PHB_SET_NODE(PHB, NODE)		((PHB)->node = (NODE))
node              266 arch/powerpc/include/asm/pci-bridge.h #define PHB_SET_NODE(PHB, NODE)		((PHB)->node = NUMA_NO_NODE)
node              273 arch/powerpc/include/asm/pci-bridge.h 			struct device_node* node);
node               99 arch/powerpc/include/asm/pci.h extern struct pci_dev *of_create_pci_dev(struct device_node *node,
node              106 arch/powerpc/include/asm/pci.h extern void of_scan_bus(struct device_node *node, struct pci_bus *bus);
node              107 arch/powerpc/include/asm/pci.h extern void of_rescan_bus(struct device_node *node, struct pci_bus *bus);
node              146 arch/powerpc/include/asm/pmac_feature.h static inline long pmac_call_feature(int selector, struct device_node* node,
node              151 arch/powerpc/include/asm/pmac_feature.h 	return ppc_md.feature_call(selector, node, param, value);
node               55 arch/powerpc/include/asm/pmac_low_i2c.h extern struct pmac_i2c_bus *pmac_i2c_find_bus(struct device_node *node);
node              129 arch/powerpc/include/asm/pmac_pfunc.h 	struct device_node	*node;
node               42 arch/powerpc/include/asm/pmi.h 	struct list_head node;
node              513 arch/powerpc/include/asm/ps3.h void ps3_sync_irq(int node);
node              383 arch/powerpc/include/asm/rtas.h extern int early_init_dt_scan_rtas(unsigned long node,
node              120 arch/powerpc/include/asm/spu.h 	u32 node;
node               20 arch/powerpc/include/asm/topology.h #define cpumask_of_node(node) ((node) == -1 ?				\
node               22 arch/powerpc/include/asm/topology.h 			       node_to_cpumask_map[node])
node               48 arch/powerpc/include/asm/topology.h static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node)
node               50 arch/powerpc/include/asm/topology.h 	numa_cpu_lookup_table[cpu] = node;
node               86 arch/powerpc/include/asm/topology.h static inline void update_numa_cpu_lookup_table(unsigned int cpu, int node) {}
node               26 arch/powerpc/include/asm/tsi108_pci.h extern void tsi108_pci_int_init(struct device_node *node);
node               15 arch/powerpc/include/asm/ultravisor.h int early_init_dt_scan_ultravisor(unsigned long node, const char *uname,
node               91 arch/powerpc/include/asm/xics.h 	int (*host_match)(struct ics *ics, struct device_node *node);
node              317 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_lookup_by_node(const struct device_node *node)
node              323 arch/powerpc/kernel/cacheinfo.c 		if (iter->ofnode != node)
node              353 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode_unified(struct device_node *node, int level)
node              355 arch/powerpc/kernel/cacheinfo.c 	pr_debug("creating L%d ucache for %pOF\n", level, node);
node              357 arch/powerpc/kernel/cacheinfo.c 	return new_cache(cache_is_unified_d(node), level, node);
node              360 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode_split(struct device_node *node,
node              366 arch/powerpc/kernel/cacheinfo.c 		 node);
node              368 arch/powerpc/kernel/cacheinfo.c 	dcache = new_cache(CACHE_TYPE_DATA, level, node);
node              369 arch/powerpc/kernel/cacheinfo.c 	icache = new_cache(CACHE_TYPE_INSTRUCTION, level, node);
node              383 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_do_one_devnode(struct device_node *node, int level)
node              387 arch/powerpc/kernel/cacheinfo.c 	if (cache_node_is_unified(node))
node              388 arch/powerpc/kernel/cacheinfo.c 		cache = cache_do_one_devnode_unified(node, level);
node              390 arch/powerpc/kernel/cacheinfo.c 		cache = cache_do_one_devnode_split(node, level);
node              395 arch/powerpc/kernel/cacheinfo.c static struct cache *cache_lookup_or_instantiate(struct device_node *node,
node              400 arch/powerpc/kernel/cacheinfo.c 	cache = cache_lookup_by_node(node);
node              407 arch/powerpc/kernel/cacheinfo.c 		cache = cache_do_one_devnode(node, level);
node               51 arch/powerpc/kernel/dt_cpu_ftrs.c 	unsigned long node;
node              795 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init fdt_find_cpu_features(unsigned long node, const char *uname,
node              798 arch/powerpc/kernel/dt_cpu_ftrs.c 	if (of_flat_dt_is_compatible(node, "ibm,powerpc-cpu-features")
node              799 arch/powerpc/kernel/dt_cpu_ftrs.c 	    && of_get_flat_dt_prop(node, "isa", NULL))
node              833 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init process_cpufeatures_node(unsigned long node,
node              842 arch/powerpc/kernel/dt_cpu_ftrs.c 	f->node = node;
node              846 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "isa", &len);
node              853 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "usable-privilege", &len);
node              860 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "hv-support", &len);
node              866 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "os-support", &len);
node              872 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "hfscr-bit-nr", &len);
node              877 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "fscr-bit-nr", &len);
node              882 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "hwcap-bit-nr", &len);
node              939 arch/powerpc/kernel/dt_cpu_ftrs.c 	if (!of_get_flat_dt_prop(node, "dependencies", &len)) {
node              959 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(f->node, "dependencies", &len);
node              974 arch/powerpc/kernel/dt_cpu_ftrs.c 			if (of_get_flat_dt_phandle(d->node) == phandle) {
node              990 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init scan_cpufeatures_subnodes(unsigned long node,
node              996 arch/powerpc/kernel/dt_cpu_ftrs.c 	process_cpufeatures_node(node, uname, *count);
node             1003 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init count_cpufeatures_subnodes(unsigned long node,
node             1014 arch/powerpc/kernel/dt_cpu_ftrs.c static int __init dt_cpu_ftrs_scan_callback(unsigned long node, const char
node             1022 arch/powerpc/kernel/dt_cpu_ftrs.c 	if (!of_flat_dt_is_compatible(node, "ibm,powerpc-cpu-features"))
node             1025 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "isa", NULL);
node             1033 arch/powerpc/kernel/dt_cpu_ftrs.c 	of_scan_flat_dt_subnodes(node, count_cpufeatures_subnodes,
node             1046 arch/powerpc/kernel/dt_cpu_ftrs.c 	of_scan_flat_dt_subnodes(node, scan_cpufeatures_subnodes, &count);
node             1055 arch/powerpc/kernel/dt_cpu_ftrs.c 	prop = of_get_flat_dt_prop(node, "display-name", NULL);
node               23 arch/powerpc/kernel/epapr_paravirt.c static int __init early_init_dt_scan_epapr(unsigned long node,
node               31 arch/powerpc/kernel/epapr_paravirt.c 	insts = of_get_flat_dt_prop(node, "hcall-instructions", &len);
node               47 arch/powerpc/kernel/epapr_paravirt.c 	if (of_get_flat_dt_prop(node, "has-idle", NULL))
node              108 arch/powerpc/kernel/fadump.c int __init early_init_dt_scan_fw_dump(unsigned long node, const char *uname,
node              115 arch/powerpc/kernel/fadump.c 		rtas_fadump_dt_scan(&fw_dump, node);
node              120 arch/powerpc/kernel/fadump.c 		opal_fadump_dt_scan(&fw_dump, node);
node             1503 arch/powerpc/kernel/fadump.c int __init early_init_dt_scan_fw_dump(unsigned long node, const char *uname,
node             1509 arch/powerpc/kernel/fadump.c 	opal_fadump_dt_scan(&fw_dump, node);
node              854 arch/powerpc/kernel/iommu.c 			   unsigned long mask, gfp_t flag, int node)
node              880 arch/powerpc/kernel/iommu.c 	page = alloc_pages_node(node, flag, order);
node              235 arch/powerpc/kernel/machine_kexec.c static void __init export_crashk_values(struct device_node *node)
node              239 arch/powerpc/kernel/machine_kexec.c 	of_remove_property(node, of_find_property(node,
node              241 arch/powerpc/kernel/machine_kexec.c 	of_remove_property(node, of_find_property(node,
node              246 arch/powerpc/kernel/machine_kexec.c 		of_add_property(node, &crashk_base_prop);
node              248 arch/powerpc/kernel/machine_kexec.c 		of_add_property(node, &crashk_size_prop);
node              256 arch/powerpc/kernel/machine_kexec.c 	of_update_property(node, &memory_limit_prop);
node              261 arch/powerpc/kernel/machine_kexec.c 	struct device_node *node;
node              263 arch/powerpc/kernel/machine_kexec.c 	node = of_find_node_by_path("/chosen");
node              264 arch/powerpc/kernel/machine_kexec.c 	if (!node)
node              268 arch/powerpc/kernel/machine_kexec.c 	of_remove_property(node, of_find_property(node, kernel_end_prop.name, NULL));
node              272 arch/powerpc/kernel/machine_kexec.c 	of_add_property(node, &kernel_end_prop);
node              274 arch/powerpc/kernel/machine_kexec.c 	export_crashk_values(node);
node              276 arch/powerpc/kernel/machine_kexec.c 	of_node_put(node);
node               40 arch/powerpc/kernel/machine_kexec_64.c 	struct device_node *node;
node               54 arch/powerpc/kernel/machine_kexec_64.c 	for_each_node_by_type(node, "pci") {
node               55 arch/powerpc/kernel/machine_kexec_64.c 		basep = of_get_property(node, "linux,tce-base", NULL);
node               56 arch/powerpc/kernel/machine_kexec_64.c 		sizep = of_get_property(node, "linux,tce-size", NULL);
node              394 arch/powerpc/kernel/machine_kexec_64.c 	struct device_node *node;
node              400 arch/powerpc/kernel/machine_kexec_64.c 	node = of_find_node_by_path("/chosen");
node              401 arch/powerpc/kernel/machine_kexec_64.c 	if (!node)
node              405 arch/powerpc/kernel/machine_kexec_64.c 	of_remove_property(node, of_find_property(node, htab_base_prop.name, NULL));
node              406 arch/powerpc/kernel/machine_kexec_64.c 	of_remove_property(node, of_find_property(node, htab_size_prop.name, NULL));
node              409 arch/powerpc/kernel/machine_kexec_64.c 	of_add_property(node, &htab_base_prop);
node              411 arch/powerpc/kernel/machine_kexec_64.c 	of_add_property(node, &htab_size_prop);
node              413 arch/powerpc/kernel/machine_kexec_64.c 	of_node_put(node);
node              339 arch/powerpc/kernel/pci-common.c struct pci_controller* pci_find_hose_for_OF_device(struct device_node* node)
node              341 arch/powerpc/kernel/pci-common.c 	while(node) {
node              344 arch/powerpc/kernel/pci-common.c 			if (hose->dn == node)
node              346 arch/powerpc/kernel/pci-common.c 		node = node->parent;
node             1226 arch/powerpc/kernel/pci-common.c 	list_for_each_entry(b, &bus->children, node)
node             1359 arch/powerpc/kernel/pci-common.c 	list_for_each_entry(b, &pci_root_buses, node)
node             1371 arch/powerpc/kernel/pci-common.c 		list_for_each_entry(b, &pci_root_buses, node)
node             1413 arch/powerpc/kernel/pci-common.c 	list_for_each_entry(child_bus, &bus->children, node)
node             1596 arch/powerpc/kernel/pci-common.c 	struct device_node *node = hose->dn;
node             1599 arch/powerpc/kernel/pci-common.c 	pr_debug("PCI: Scanning PHB %pOF\n", node);
node             1625 arch/powerpc/kernel/pci-common.c 	if (node && hose->controller_ops.probe_mode)
node             1629 arch/powerpc/kernel/pci-common.c 		of_scan_bus(node, bus);
node             1646 arch/powerpc/kernel/pci-common.c 		list_for_each_entry(child, &bus->children, node)
node               29 arch/powerpc/kernel/pci-hotplug.c 	list_for_each_entry(tmp, &bus->children, node) {
node               85 arch/powerpc/kernel/pci-hotplug.c 	list_for_each_entry(child_bus, &bus->children, node)
node               75 arch/powerpc/kernel/pci_32.c make_one_node_map(struct device_node* node, u8 pci_bus)
node               82 arch/powerpc/kernel/pci_32.c 	bus_range = of_get_property(node, "bus-range", &len);
node               85 arch/powerpc/kernel/pci_32.c 		       "assuming it starts at 0\n", node);
node               90 arch/powerpc/kernel/pci_32.c 	for_each_child_of_node(node, node) {
node               94 arch/powerpc/kernel/pci_32.c 		class_code = of_get_property(node, "class-code", NULL);
node               98 arch/powerpc/kernel/pci_32.c 		reg = of_get_property(node, "reg", NULL);
node              107 arch/powerpc/kernel/pci_32.c 		make_one_node_map(node, dev->subordinate->number);
node              134 arch/powerpc/kernel/pci_32.c 		struct device_node* node = hose->dn;
node              136 arch/powerpc/kernel/pci_32.c 		if (!node)
node              138 arch/powerpc/kernel/pci_32.c 		make_one_node_map(node, hose->first_busno);
node              161 arch/powerpc/kernel/pci_32.c int pci_device_from_OF_node(struct device_node *node, u8 *bus, u8 *devfn)
node              168 arch/powerpc/kernel/pci_32.c 	if (!pci_find_hose_for_OF_device(node))
node              171 arch/powerpc/kernel/pci_32.c 	reg = of_get_property(node, "reg", &size);
node              233 arch/powerpc/kernel/pci_64.c 	list_for_each_entry(tmp_bus, &pci_root_buses, node) {
node              266 arch/powerpc/kernel/pci_64.c 	return phb->node;
node              117 arch/powerpc/kernel/pci_of_scan.c static void of_pci_parse_addrs(struct device_node *node, struct pci_dev *dev)
node              128 arch/powerpc/kernel/pci_of_scan.c 	addrs = of_get_property(node, "assigned-addresses", &proplen);
node              130 arch/powerpc/kernel/pci_of_scan.c 		addrs = of_get_property(node, "reg", &proplen);
node              175 arch/powerpc/kernel/pci_of_scan.c struct pci_dev *of_create_pci_dev(struct device_node *node,
node              185 arch/powerpc/kernel/pci_of_scan.c 		 of_node_get_device_type(node));
node              187 arch/powerpc/kernel/pci_of_scan.c 	dev->dev.of_node = of_node_get(node);
node              196 arch/powerpc/kernel/pci_of_scan.c 	dev->vendor = get_int_prop(node, "vendor-id", 0xffff);
node              197 arch/powerpc/kernel/pci_of_scan.c 	dev->device = get_int_prop(node, "device-id", 0xffff);
node              198 arch/powerpc/kernel/pci_of_scan.c 	dev->subsystem_vendor = get_int_prop(node, "subsystem-vendor-id", 0);
node              199 arch/powerpc/kernel/pci_of_scan.c 	dev->subsystem_device = get_int_prop(node, "subsystem-id", 0);
node              205 arch/powerpc/kernel/pci_of_scan.c 	dev->class = get_int_prop(node, "class-code", 0);
node              206 arch/powerpc/kernel/pci_of_scan.c 	dev->revision = get_int_prop(node, "revision-id", 0);
node              218 arch/powerpc/kernel/pci_of_scan.c 	if (of_node_is_type(node, "pci") || of_node_is_type(node, "pciex")) {
node              223 arch/powerpc/kernel/pci_of_scan.c 	} else if (of_node_is_type(node, "cardbus")) {
node              232 arch/powerpc/kernel/pci_of_scan.c 	of_pci_parse_addrs(node, dev);
node              252 arch/powerpc/kernel/pci_of_scan.c 	struct device_node *node = dev->dev.of_node;
node              262 arch/powerpc/kernel/pci_of_scan.c 	pr_debug("of_scan_pci_bridge(%pOF)\n", node);
node              265 arch/powerpc/kernel/pci_of_scan.c 	busrange = of_get_property(node, "bus-range", &len);
node              268 arch/powerpc/kernel/pci_of_scan.c 		       node);
node              271 arch/powerpc/kernel/pci_of_scan.c 	ranges = of_get_property(node, "ranges", &len);
node              274 arch/powerpc/kernel/pci_of_scan.c 		       node);
node              285 arch/powerpc/kernel/pci_of_scan.c 			       node);
node              313 arch/powerpc/kernel/pci_of_scan.c 				       " for bridge %pOF\n", node);
node              319 arch/powerpc/kernel/pci_of_scan.c 				       " for bridge %pOF\n", node);
node              342 arch/powerpc/kernel/pci_of_scan.c 		of_scan_bus(node, bus);
node              395 arch/powerpc/kernel/pci_of_scan.c static void __of_scan_bus(struct device_node *node, struct pci_bus *bus,
node              402 arch/powerpc/kernel/pci_of_scan.c 		 node, bus->number);
node              405 arch/powerpc/kernel/pci_of_scan.c 	for_each_child_of_node(node, child) {
node              429 arch/powerpc/kernel/pci_of_scan.c void of_scan_bus(struct device_node *node, struct pci_bus *bus)
node              431 arch/powerpc/kernel/pci_of_scan.c 	__of_scan_bus(node, bus, 0);
node              443 arch/powerpc/kernel/pci_of_scan.c void of_rescan_bus(struct device_node *node, struct pci_bus *bus)
node              445 arch/powerpc/kernel/pci_of_scan.c 	__of_scan_bus(node, bus, 1);
node              180 arch/powerpc/kernel/prom.c static void __init scan_features(unsigned long node, const unsigned char *ftrs,
node              219 arch/powerpc/kernel/prom.c static void __init check_cpu_pa_features(unsigned long node)
node              224 arch/powerpc/kernel/prom.c 	pa_ftrs = of_get_flat_dt_prop(node, "ibm,pa-features", &tablelen);
node              228 arch/powerpc/kernel/prom.c 	scan_features(node, pa_ftrs, tablelen,
node              233 arch/powerpc/kernel/prom.c static void __init init_mmu_slb_size(unsigned long node)
node              237 arch/powerpc/kernel/prom.c 	slb_size_ptr = of_get_flat_dt_prop(node, "slb-size", NULL) ? :
node              238 arch/powerpc/kernel/prom.c 			of_get_flat_dt_prop(node, "ibm,slb-size", NULL);
node              244 arch/powerpc/kernel/prom.c #define init_mmu_slb_size(node) do { } while(0)
node              269 arch/powerpc/kernel/prom.c static inline void identical_pvr_fixup(unsigned long node)
node              272 arch/powerpc/kernel/prom.c 	const char *model = of_get_flat_dt_prop(node, "model", NULL);
node              288 arch/powerpc/kernel/prom.c #define identical_pvr_fixup(node) do { } while(0)
node              291 arch/powerpc/kernel/prom.c static void __init check_cpu_feature_properties(unsigned long node)
node              298 arch/powerpc/kernel/prom.c 		prop = of_get_flat_dt_prop(node, fp->name, NULL);
node              306 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_cpus(unsigned long node,
node              310 arch/powerpc/kernel/prom.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              323 arch/powerpc/kernel/prom.c 	intserv = of_get_flat_dt_prop(node, "ibm,ppc-interrupt-server#s", &len);
node              325 arch/powerpc/kernel/prom.c 		intserv = of_get_flat_dt_prop(node, "reg", &len);
node              373 arch/powerpc/kernel/prom.c 		prop = of_get_flat_dt_prop(node, "cpu-version", NULL);
node              377 arch/powerpc/kernel/prom.c 		check_cpu_feature_properties(node);
node              378 arch/powerpc/kernel/prom.c 		check_cpu_pa_features(node);
node              381 arch/powerpc/kernel/prom.c 	identical_pvr_fixup(node);
node              382 arch/powerpc/kernel/prom.c 	init_mmu_slb_size(node);
node              396 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_chosen_ppc(unsigned long node,
node              403 arch/powerpc/kernel/prom.c 	if (early_init_dt_scan_chosen(node, uname, depth, data) == 0)
node              408 arch/powerpc/kernel/prom.c 	if (of_get_flat_dt_prop(node, "linux,iommu-off", NULL) != NULL)
node              410 arch/powerpc/kernel/prom.c 	if (of_get_flat_dt_prop(node, "linux,iommu-force-on", NULL) != NULL)
node              415 arch/powerpc/kernel/prom.c 	lprop = of_get_flat_dt_prop(node, "linux,memory-limit", NULL);
node              420 arch/powerpc/kernel/prom.c 	lprop = of_get_flat_dt_prop(node, "linux,tce-alloc-start", NULL);
node              423 arch/powerpc/kernel/prom.c 	lprop = of_get_flat_dt_prop(node, "linux,tce-alloc-end", NULL);
node              429 arch/powerpc/kernel/prom.c 	lprop = of_get_flat_dt_prop(node, "linux,crashkernel-base", NULL);
node              433 arch/powerpc/kernel/prom.c 	lprop = of_get_flat_dt_prop(node, "linux,crashkernel-size", NULL);
node              525 arch/powerpc/kernel/prom.c static int __init early_init_dt_scan_memory_ppc(unsigned long node,
node              532 arch/powerpc/kernel/prom.c 		walk_drmem_lmbs_early(node, early_init_drmem_lmb);
node              537 arch/powerpc/kernel/prom.c 	return early_init_dt_scan_memory(node, uname, depth, data);
node              618 arch/powerpc/kernel/prom_init.c 	phandle node;
node              620 arch/powerpc/kernel/prom_init.c 	if ((node = *nodep) != 0
node              621 arch/powerpc/kernel/prom_init.c 	    && (*nodep = call_prom("child", 1, 1, node)) != 0)
node              623 arch/powerpc/kernel/prom_init.c 	if ((*nodep = call_prom("peer", 1, 1, node)) != 0)
node              626 arch/powerpc/kernel/prom_init.c 		if ((node = call_prom("parent", 1, 1, node)) == 0)
node              628 arch/powerpc/kernel/prom_init.c 		if ((*nodep = call_prom("peer", 1, 1, node)) != 0)
node              633 arch/powerpc/kernel/prom_init.c static inline int __init prom_getprop(phandle node, const char *pname,
node              636 arch/powerpc/kernel/prom_init.c 	return call_prom("getprop", 4, 1, node, ADDR(pname),
node              640 arch/powerpc/kernel/prom_init.c static inline int __init prom_getproplen(phandle node, const char *pname)
node              642 arch/powerpc/kernel/prom_init.c 	return call_prom("getproplen", 2, 1, node, ADDR(pname));
node              671 arch/powerpc/kernel/prom_init.c static int __init prom_setprop(phandle node, const char *nodename,
node              677 arch/powerpc/kernel/prom_init.c 		return call_prom("setprop", 4, 1, node, ADDR(pname),
node             1165 arch/powerpc/kernel/prom_init.c 	phandle node;
node             1170 arch/powerpc/kernel/prom_init.c 	for (node = 0; prom_next_node(&node); ) {
node             1172 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "device_type", type, sizeof(type));
node             1181 arch/powerpc/kernel/prom_init.c 		plen = prom_getproplen(node, "ibm,ppc-interrupt-server#s");
node             1594 arch/powerpc/kernel/prom_init.c 	phandle node;
node             1617 arch/powerpc/kernel/prom_init.c 	for (node = 0; prom_next_node(&node); ) {
node             1619 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "device_type", type, sizeof(type));
node             1626 arch/powerpc/kernel/prom_init.c 			prom_getprop(node, "name", type, sizeof(type));
node             1631 arch/powerpc/kernel/prom_init.c 		plen = prom_getprop(node, "reg", regbuf, sizeof(regbuf));
node             1641 arch/powerpc/kernel/prom_init.c 		call_prom("package-to-path", 3, 1, node, prom_scratch,
node             1921 arch/powerpc/kernel/prom_init.c 	phandle node;
node             1941 arch/powerpc/kernel/prom_init.c 	for (node = 0; prom_next_node(&node); ) {
node             1945 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "compatible",
node             1947 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "device_type", type, sizeof(type));
node             1948 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "model", model, sizeof(model));
node             1966 arch/powerpc/kernel/prom_init.c 		if (prom_getprop(node, "tce-table-minalign", &minalign,
node             1969 arch/powerpc/kernel/prom_init.c 		if (prom_getprop(node, "tce-table-minsize", &minsize,
node             1992 arch/powerpc/kernel/prom_init.c 		if (call_prom("package-to-path", 3, 1, node,
node             1998 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, path, "linux,tce-base", &base, sizeof(base));
node             1999 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, path, "linux,tce-size", &minsize, sizeof(minsize));
node             2002 arch/powerpc/kernel/prom_init.c 		prom_debug("\tnode = 0x%x\n", node);
node             2069 arch/powerpc/kernel/prom_init.c 	phandle node;
node             2105 arch/powerpc/kernel/prom_init.c 	for (node = 0; prom_next_node(&node); ) {
node             2110 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "device_type", type, sizeof(type));
node             2115 arch/powerpc/kernel/prom_init.c 		if (prom_getprop(node, "status", type, sizeof(type)) > 0)
node             2120 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "reg", &reg, sizeof(reg));
node             2134 arch/powerpc/kernel/prom_init.c 			call_prom("start-cpu", 3, 0, node,
node             2321 arch/powerpc/kernel/prom_init.c 	phandle node;
node             2346 arch/powerpc/kernel/prom_init.c 	for (node = 0; prom_next_node(&node); ) {
node             2348 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "device_type", type, sizeof(type));
node             2360 arch/powerpc/kernel/prom_init.c 		if (call_prom("package-to-path", 3, 1, node, path,
node             2373 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, path, "linux,opened", NULL, 0);
node             2392 arch/powerpc/kernel/prom_init.c 		if (prom_getprop(node, "linux,boot-display", NULL, 0) !=
node             2397 arch/powerpc/kernel/prom_init.c 			prom_getprop(node, "width", &width, 4);
node             2398 arch/powerpc/kernel/prom_init.c 			prom_getprop(node, "height", &height, 4);
node             2399 arch/powerpc/kernel/prom_init.c 			prom_getprop(node, "linebytes", &pitch, 4);
node             2400 arch/powerpc/kernel/prom_init.c 			prom_getprop(node, "address", &addr, 4);
node             2467 arch/powerpc/kernel/prom_init.c static void __init scan_dt_build_strings(phandle node,
node             2482 arch/powerpc/kernel/prom_init.c 		if (call_prom("nextprop", 3, 1, node, prev_name, namep) != 1) {
node             2508 arch/powerpc/kernel/prom_init.c 	child = call_prom("child", 1, 1, node);
node             2515 arch/powerpc/kernel/prom_init.c static void __init scan_dt_build_struct(phandle node, unsigned long *mem_start,
node             2532 arch/powerpc/kernel/prom_init.c 	l = call_prom("package-to-path", 3, 1, node, namep, room);
node             2538 arch/powerpc/kernel/prom_init.c 			call_prom("package-to-path", 3, 1, node, namep, l);
node             2559 arch/powerpc/kernel/prom_init.c 	call_prom("package-to-path", 3, 1, node, path, sizeof(prom_scratch) - 1);
node             2565 arch/powerpc/kernel/prom_init.c 		if (call_prom("nextprop", 3, 1, node, prev_name,
node             2585 arch/powerpc/kernel/prom_init.c 		l = call_prom("getproplen", 2, 1, node, pname);
node             2598 arch/powerpc/kernel/prom_init.c 		call_prom("getprop", 4, 1, node, pname, valp, l);
node             2615 arch/powerpc/kernel/prom_init.c 			*(__be32 *)valp = cpu_to_be32(node);
node             2620 arch/powerpc/kernel/prom_init.c 	child = call_prom("child", 1, 1, node);
node             2915 arch/powerpc/kernel/prom_init.c 	u32 node;
node             2920 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/builtin/ethernet"));
node             2921 arch/powerpc/kernel/prom_init.c 	if (!PHANDLE_VALID(node))
node             2925 arch/powerpc/kernel/prom_init.c 	rv = prom_getprop(node, "phy-handle", prop, sizeof(prop));
node             2935 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/builtin/mdio"));
node             2936 arch/powerpc/kernel/prom_init.c 	if (!PHANDLE_VALID(node)) {
node             2956 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1,
node             2958 arch/powerpc/kernel/prom_init.c 	if (!PHANDLE_VALID(node)) {
node             2982 arch/powerpc/kernel/prom_init.c 	u32 node;
node             2987 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/"));
node             2988 arch/powerpc/kernel/prom_init.c 	if (!PHANDLE_VALID(node))
node             2991 arch/powerpc/kernel/prom_init.c 	rv = prom_getprop(node, "model", prop, sizeof(prop));
node             3000 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/"));
node             3001 arch/powerpc/kernel/prom_init.c 	rv = prom_getprop(node, "device_type", prop, sizeof(prop));
node             3003 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, "/", "device_type", "efika", sizeof("efika"));
node             3007 arch/powerpc/kernel/prom_init.c 	rv = prom_getprop(node, "CODEGEN,description", prop, sizeof(prop));
node             3009 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, "/", "CODEGEN,description",
node             3014 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/builtin/bestcomm"));
node             3015 arch/powerpc/kernel/prom_init.c 	if (PHANDLE_VALID(node)) {
node             3016 arch/powerpc/kernel/prom_init.c 		len = prom_getproplen(node, "interrupts");
node             3019 arch/powerpc/kernel/prom_init.c 			prom_setprop(node, "/builtin/bestcom", "interrupts",
node             3025 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR("/builtin/sound"));
node             3026 arch/powerpc/kernel/prom_init.c 	if (PHANDLE_VALID(node)) {
node             3027 arch/powerpc/kernel/prom_init.c 		rv = prom_getprop(node, "interrupts", prop, sizeof(prop));
node             3030 arch/powerpc/kernel/prom_init.c 			prom_setprop(node, "/builtin/sound", "interrupts",
node             3053 arch/powerpc/kernel/prom_init.c 	phandle iob, node;
node             3070 arch/powerpc/kernel/prom_init.c 	node = call_prom("finddevice", 1, 1, ADDR(pci_name));
node             3073 arch/powerpc/kernel/prom_init.c 	for( ; prom_next_node(&node); ) {
node             3075 arch/powerpc/kernel/prom_init.c 		if (!PHANDLE_VALID(node))
node             3078 arch/powerpc/kernel/prom_init.c 		rval = prom_getproplen(node, "interrupts");
node             3082 arch/powerpc/kernel/prom_init.c 		prom_getprop(node, "interrupts", &interrupts, sizeof(interrupts));
node             3096 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, pci_name, "interrupts", interrupts,
node             3098 arch/powerpc/kernel/prom_init.c 		prom_setprop(node, pci_name, "interrupt-parent", &parent,
node             1200 arch/powerpc/kernel/rtas.c int __init early_init_dt_scan_rtas(unsigned long node,
node             1208 arch/powerpc/kernel/rtas.c 	basep  = of_get_flat_dt_prop(node, "linux,rtas-base", NULL);
node             1209 arch/powerpc/kernel/rtas.c 	entryp = of_get_flat_dt_prop(node, "linux,rtas-entry", NULL);
node             1210 arch/powerpc/kernel/rtas.c 	sizep  = of_get_flat_dt_prop(node, "rtas-size", NULL);
node             1219 arch/powerpc/kernel/rtas.c 	basep = of_get_flat_dt_prop(node, "put-term-char", NULL);
node             1223 arch/powerpc/kernel/rtas.c 	basep = of_get_flat_dt_prop(node, "get-term-char", NULL);
node             1016 arch/powerpc/kernel/sysfs.c 	struct node *node = node_devices[nid];
node             1017 arch/powerpc/kernel/sysfs.c 	return sysfs_create_link(&node->dev.kobj, &dev->kobj,
node             1024 arch/powerpc/kernel/sysfs.c 	struct node *node = node_devices[nid];
node             1025 arch/powerpc/kernel/sysfs.c 	sysfs_remove_link(&node->dev.kobj, kobject_name(&dev->kobj));
node              673 arch/powerpc/kernel/vdso.c 	unsigned long cpu, node, val;
node              682 arch/powerpc/kernel/vdso.c 	node = cpu_to_node(cpu);
node              683 arch/powerpc/kernel/vdso.c 	WARN_ON_ONCE(node > 0xffff);
node              685 arch/powerpc/kernel/vdso.c 	val = (cpu & 0xfff) | ((node & 0xffff) << 16);
node             5452 arch/powerpc/kvm/book3s_hv.c 		int node = cpu_to_node(first_cpu);
node             5460 arch/powerpc/kvm/book3s_hv.c 							GFP_KERNEL, node);
node              354 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_seg_sizes(unsigned long node,
node              358 arch/powerpc/mm/book3s64/hash_utils.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              366 arch/powerpc/mm/book3s64/hash_utils.c 	prop = of_get_flat_dt_prop(node, "ibm,processor-segment-sizes", &size);
node              410 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_page_sizes(unsigned long node,
node              414 arch/powerpc/mm/book3s64/hash_utils.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              422 arch/powerpc/mm/book3s64/hash_utils.c 	prop = of_get_flat_dt_prop(node, "ibm,segment-page-sizes", &size);
node              493 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_hugepage_blocks(unsigned long node,
node              496 arch/powerpc/mm/book3s64/hash_utils.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              511 arch/powerpc/mm/book3s64/hash_utils.c 	page_count_prop = of_get_flat_dt_prop(node, "ibm,expected#pages", NULL);
node              515 arch/powerpc/mm/book3s64/hash_utils.c 	addr_prop = of_get_flat_dt_prop(node, "reg", NULL);
node              712 arch/powerpc/mm/book3s64/hash_utils.c static int __init htab_dt_scan_pftsize(unsigned long node,
node              716 arch/powerpc/mm/book3s64/hash_utils.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              723 arch/powerpc/mm/book3s64/hash_utils.c 	prop = of_get_flat_dt_prop(node, "ibm,pft-size", NULL);
node              423 arch/powerpc/mm/book3s64/radix_pgtable.c static int __init radix_dt_scan_page_sizes(unsigned long node,
node              431 arch/powerpc/mm/book3s64/radix_pgtable.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node              438 arch/powerpc/mm/book3s64/radix_pgtable.c 	prop = of_get_flat_dt_prop(node, "ibm,mmu-pid-bits", &size);
node              443 arch/powerpc/mm/book3s64/radix_pgtable.c 	prop = of_get_flat_dt_prop(node, "ibm,processor-radix-AP-encodings", &size);
node              268 arch/powerpc/mm/drmem.c void __init walk_drmem_lmbs_early(unsigned long node,
node              274 arch/powerpc/mm/drmem.c 	prop = of_get_flat_dt_prop(node, "ibm,lmb-size", &len);
node              280 arch/powerpc/mm/drmem.c 	usm = of_get_flat_dt_prop(node, "linux,drconf-usable-memory", &len);
node              282 arch/powerpc/mm/drmem.c 	prop = of_get_flat_dt_prop(node, "ibm,dynamic-memory", &len);
node              286 arch/powerpc/mm/drmem.c 		prop = of_get_flat_dt_prop(node, "ibm,dynamic-memory-v2",
node              129 arch/powerpc/mm/init_64.c static __meminit struct vmemmap_backing * vmemmap_list_alloc(int node)
node              143 arch/powerpc/mm/init_64.c 		next = vmemmap_alloc_block(PAGE_SIZE, node);
node              158 arch/powerpc/mm/init_64.c 					    int node)
node              162 arch/powerpc/mm/init_64.c 	vmem_back = vmemmap_list_alloc(node);
node              190 arch/powerpc/mm/init_64.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              198 arch/powerpc/mm/init_64.c 	pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node);
node              218 arch/powerpc/mm/init_64.c 			p = vmemmap_alloc_block_buf(page_size, node);
node              222 arch/powerpc/mm/init_64.c 		vmemmap_list_populate(__pa(p), start, node);
node               71 arch/powerpc/mm/numa.c 	unsigned int node;
node               78 arch/powerpc/mm/numa.c 	for_each_node(node)
node               79 arch/powerpc/mm/numa.c 		alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]);
node              140 arch/powerpc/mm/numa.c static void map_cpu_to_node(int cpu, int node)
node              142 arch/powerpc/mm/numa.c 	update_numa_cpu_lookup_table(cpu, node);
node              144 arch/powerpc/mm/numa.c 	dbg("adding cpu %d to node %d\n", cpu, node);
node              146 arch/powerpc/mm/numa.c 	if (!(cpumask_test_cpu(cpu, node_to_cpumask_map[node])))
node              147 arch/powerpc/mm/numa.c 		cpumask_set_cpu(cpu, node_to_cpumask_map[node]);
node              153 arch/powerpc/mm/numa.c 	int node = numa_cpu_lookup_table[cpu];
node              155 arch/powerpc/mm/numa.c 	dbg("removing cpu %lu from node %d\n", cpu, node);
node              157 arch/powerpc/mm/numa.c 	if (cpumask_test_cpu(cpu, node_to_cpumask_map[node])) {
node              158 arch/powerpc/mm/numa.c 		cpumask_clear_cpu(cpu, node_to_cpumask_map[node]);
node              161 arch/powerpc/mm/numa.c 		       cpu, node);
node              505 arch/powerpc/mm/numa.c static void verify_cpu_node_mapping(int cpu, int node)
node              518 arch/powerpc/mm/numa.c 		if (cpu_to_node(sibling) != node) {
node              767 arch/powerpc/mm/numa.c 	unsigned int node;
node              773 arch/powerpc/mm/numa.c 	for_each_online_node(node) {
node              774 arch/powerpc/mm/numa.c 		pr_info("Node %d CPUs:", node);
node              783 arch/powerpc/mm/numa.c 					node_to_cpumask_map[node])) {
node              136 arch/powerpc/oprofile/cell/spu_profiler.c 	int cpu, node, k, num_samples, spu_num;
node              145 arch/powerpc/oprofile/cell/spu_profiler.c 		node = cbe_cpu_to_node(cpu);
node              166 arch/powerpc/oprofile/cell/spu_profiler.c 			spu_num = k + (node * SPUS_PER_NODE);
node              200 arch/powerpc/oprofile/op_model_cell.c static void pm_rtas_reset_signals(u32 node)
node              215 arch/powerpc/oprofile/op_model_cell.c 	pm_signal_local.cpu = node;
node              235 arch/powerpc/oprofile/op_model_cell.c static int pm_rtas_activate_signals(u32 node, u32 count)
node              254 arch/powerpc/oprofile/op_model_cell.c 			pm_signal_local[i].cpu = node;
node              588 arch/powerpc/oprofile/op_model_cell.c 	int node;
node              620 arch/powerpc/oprofile/op_model_cell.c 		node = cbe_cpu_to_node(cpu);
node              621 arch/powerpc/oprofile/op_model_cell.c 		cur_phys_spu = (node * NUM_SPUS_PER_NODE)
node              623 arch/powerpc/oprofile/op_model_cell.c 		nxt_phys_spu = (node * NUM_SPUS_PER_NODE)
node             1076 arch/powerpc/oprofile/op_model_cell.c static int pm_rtas_activate_spu_profiling(u32 node)
node             1086 arch/powerpc/oprofile/op_model_cell.c 		pm_signal_local[i].cpu = node;
node              528 arch/powerpc/perf/hv-24x7.c 	struct rb_node node;
node              570 arch/powerpc/perf/hv-24x7.c 		it = rb_entry(*new, struct event_uniq, node);
node              599 arch/powerpc/perf/hv-24x7.c 	rb_link_node(&data->node, parent, new);
node              600 arch/powerpc/perf/hv-24x7.c 	rb_insert_color(&data->node, root);
node              614 arch/powerpc/perf/hv-24x7.c 	rbtree_postorder_for_each_entry_safe(pos, n, root, node)
node              209 arch/powerpc/perf/imc-pmu.c static int update_events_in_group(struct device_node *node, struct imc_pmu *pmu)
node              219 arch/powerpc/perf/imc-pmu.c 	if (!of_property_read_u32(node, "events", &handle))
node              232 arch/powerpc/perf/imc-pmu.c 	if (of_property_read_string(node, "events-prefix", &prefix))
node              236 arch/powerpc/perf/imc-pmu.c 	if (of_property_read_string(node, "scale", &g_scale))
node              239 arch/powerpc/perf/imc-pmu.c 	if (of_property_read_string(node, "unit", &g_unit))
node              243 arch/powerpc/perf/imc-pmu.c 	of_property_read_u32(node, "reg", &base_reg);
node              638 arch/powerpc/platforms/4xx/pci.c 	struct device_node	*node;
node             1244 arch/powerpc/platforms/4xx/pci.c 			port->node);
node             1525 arch/powerpc/platforms/4xx/pci.c 		if (of_device_is_compatible(port->node,
node             1745 arch/powerpc/platforms/4xx/pci.c 		if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx"))
node             1750 arch/powerpc/platforms/4xx/pci.c 				port->node, "ibm,plb-pciex-476fpe") ||
node             1752 arch/powerpc/platforms/4xx/pci.c 				port->node, "ibm,plb-pciex-476gtr"))
node             1802 arch/powerpc/platforms/4xx/pci.c 			       port->node);
node             1881 arch/powerpc/platforms/4xx/pci.c 		if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx") ||
node             1883 arch/powerpc/platforms/4xx/pci.c 			    port->node, "ibm,plb-pciex-476fpe") ||
node             1885 arch/powerpc/platforms/4xx/pci.c 			    port->node, "ibm,plb-pciex-476gtr"))
node             1925 arch/powerpc/platforms/4xx/pci.c 	if (of_get_property(port->node, "primary", NULL))
node             1929 arch/powerpc/platforms/4xx/pci.c 	bus_range = of_get_property(port->node, "bus-range", NULL);
node             1932 arch/powerpc/platforms/4xx/pci.c 	hose = pcibios_alloc_controller(port->node);
node             1965 arch/powerpc/platforms/4xx/pci.c 			       port->node);
node             1977 arch/powerpc/platforms/4xx/pci.c 		       port->node);
node             1982 arch/powerpc/platforms/4xx/pci.c 	pr_debug("PCIE %pOF, bus %d..%d\n", port->node,
node             2007 arch/powerpc/platforms/4xx/pci.c 	pci_process_bridge_OF_ranges(hose, port->node, primary);
node             2027 arch/powerpc/platforms/4xx/pci.c 	pval = of_get_property(port->node, "vendor-id", NULL);
node             2038 arch/powerpc/platforms/4xx/pci.c 	pval = of_get_property(port->node, "device-id", NULL);
node             2050 arch/powerpc/platforms/4xx/pci.c 	if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx"))
node             2113 arch/powerpc/platforms/4xx/pci.c 	port->node = of_node_get(np);
node             2128 arch/powerpc/platforms/4xx/pci.c 	if (of_node_is_type(port->node, "pci-endpoint")) {
node             2130 arch/powerpc/platforms/4xx/pci.c 	} else if (of_node_is_type(port->node, "pci")) {
node              228 arch/powerpc/platforms/4xx/uic.c static struct uic * __init uic_init_one(struct device_node *node)
node              234 arch/powerpc/platforms/4xx/uic.c 	BUG_ON(! of_device_is_compatible(node, "ibm,uic"));
node              241 arch/powerpc/platforms/4xx/uic.c 	indexp = of_get_property(node, "cell-index", &len);
node              244 arch/powerpc/platforms/4xx/uic.c 		       "cell-index property\n", node);
node              249 arch/powerpc/platforms/4xx/uic.c 	dcrreg = of_get_property(node, "dcr-reg", &len);
node              252 arch/powerpc/platforms/4xx/uic.c 		       "dcr-reg property\n", node);
node              257 arch/powerpc/platforms/4xx/uic.c 	uic->irqhost = irq_domain_add_linear(node, NR_UIC_INTS, &uic_host_ops,
node              123 arch/powerpc/platforms/512x/mpc5121_ads_cpld.c cpld_pic_host_match(struct irq_domain *h, struct device_node *node,
node              126 arch/powerpc/platforms/512x/mpc5121_ads_cpld.c 	return cpld_pic_node == node;
node              214 arch/powerpc/platforms/52xx/mpc52xx_common.c unsigned int mpc52xx_get_xtal_freq(struct device_node *node)
node              222 arch/powerpc/platforms/52xx/mpc52xx_common.c 	freq = mpc5xxx_get_bus_frequency(node);
node              241 arch/powerpc/platforms/52xx/mpc52xx_gpt.c mpc52xx_gpt_irq_setup(struct mpc52xx_gpt_priv *gpt, struct device_node *node)
node              247 arch/powerpc/platforms/52xx/mpc52xx_gpt.c 	cascade_virq = irq_of_parse_and_map(node, 0);
node              251 arch/powerpc/platforms/52xx/mpc52xx_gpt.c 	gpt->irqhost = irq_domain_add_linear(node, 1, &mpc52xx_gpt_irq_ops, gpt);
node              321 arch/powerpc/platforms/52xx/mpc52xx_gpt.c mpc52xx_gpt_gpio_setup(struct mpc52xx_gpt_priv *gpt, struct device_node *node)
node              327 arch/powerpc/platforms/52xx/mpc52xx_gpt.c 	if (!of_find_property(node, "gpio-controller", NULL))
node              330 arch/powerpc/platforms/52xx/mpc52xx_gpt.c 	gpt->gc.label = kasprintf(GFP_KERNEL, "%pOF", node);
node              342 arch/powerpc/platforms/52xx/mpc52xx_gpt.c 	gpt->gc.of_node = node;
node              364 arch/powerpc/platforms/52xx/mpc52xx_pci.c mpc52xx_add_bridge(struct device_node *node)
node              372 arch/powerpc/platforms/52xx/mpc52xx_pci.c 	pr_debug("Adding MPC52xx PCI host bridge %pOF\n", node);
node              376 arch/powerpc/platforms/52xx/mpc52xx_pci.c 	if (of_address_to_resource(node, 0, &rsrc) != 0) {
node              377 arch/powerpc/platforms/52xx/mpc52xx_pci.c 		printk(KERN_ERR "Can't get %pOF resources\n", node);
node              381 arch/powerpc/platforms/52xx/mpc52xx_pci.c 	bus_range = of_get_property(node, "bus-range", &len);
node              384 arch/powerpc/platforms/52xx/mpc52xx_pci.c 		       node);
node              396 arch/powerpc/platforms/52xx/mpc52xx_pci.c 	hose = pcibios_alloc_controller(node);
node              409 arch/powerpc/platforms/52xx/mpc52xx_pci.c 	pci_process_bridge_OF_ranges(hose, node, 1);
node              114 arch/powerpc/platforms/82xx/ep8248e.c 	struct device_node *node;
node              117 arch/powerpc/platforms/82xx/ep8248e.c 	node = of_get_parent(ofdev->dev.of_node);
node              118 arch/powerpc/platforms/82xx/ep8248e.c 	of_node_put(node);
node              119 arch/powerpc/platforms/82xx/ep8248e.c 	if (node != ep8248e_bcsr_node)
node               27 arch/powerpc/platforms/82xx/pq2ads-pci-pic.c 	struct device_node *node;
node              128 arch/powerpc/platforms/85xx/mpc85xx_ds.c 	struct device_node *node;
node              132 arch/powerpc/platforms/85xx/mpc85xx_ds.c 	node = of_find_node_by_name(NULL, "uli1575");
node              133 arch/powerpc/platforms/85xx/mpc85xx_ds.c 	while ((pci_with_uli = of_get_parent(node))) {
node              134 arch/powerpc/platforms/85xx/mpc85xx_ds.c 		of_node_put(node);
node              135 arch/powerpc/platforms/85xx/mpc85xx_ds.c 		node = pci_with_uli;
node               73 arch/powerpc/platforms/85xx/sgy_cts1000.c 	struct device_node *node = pdev->dev.of_node;
node               77 arch/powerpc/platforms/85xx/sgy_cts1000.c 	if (!node)
node               81 arch/powerpc/platforms/85xx/sgy_cts1000.c 	halt_node = of_find_matching_node(node, child_match);
node              161 arch/powerpc/platforms/cell/cbe_regs.c u32 cbe_node_to_cpu(int node)
node              163 arch/powerpc/platforms/cell/cbe_regs.c 	return cpumask_first(&cbe_local_mask[node]);
node               41 arch/powerpc/platforms/cell/interrupt.c 	struct device_node *node;
node               52 arch/powerpc/platforms/cell/interrupt.c 	unsigned char node = bits.source >> 4;
node               59 arch/powerpc/platforms/cell/interrupt.c 		return (node << IIC_IRQ_NODE_SHIFT) | (class << 4) | unit;
node              207 arch/powerpc/platforms/cell/interrupt.c static int iic_host_match(struct irq_domain *h, struct device_node *node,
node              210 arch/powerpc/platforms/cell/interrupt.c 	return of_device_is_compatible(node,
node              236 arch/powerpc/platforms/cell/interrupt.c 	unsigned int node, ext, unit, class;
node              248 arch/powerpc/platforms/cell/interrupt.c 	node = intspec[0] >> 24;
node              254 arch/powerpc/platforms/cell/interrupt.c 	if (node > 1)
node              258 arch/powerpc/platforms/cell/interrupt.c 	*out_hwirq = (node << IIC_IRQ_NODE_SHIFT);
node              278 arch/powerpc/platforms/cell/interrupt.c 				struct device_node *node)
node              290 arch/powerpc/platforms/cell/interrupt.c 	iic->node = of_node_get(node);
node              294 arch/powerpc/platforms/cell/interrupt.c 	       hw_cpu, iic->target_id, node);
node              301 arch/powerpc/platforms/cell/interrupt.c 	unsigned int node, cascade, found = 0;
node              330 arch/powerpc/platforms/cell/interrupt.c 		node = np[0] >> 1;
node              332 arch/powerpc/platforms/cell/interrupt.c 		cascade = node << IIC_IRQ_NODE_SHIFT;
node              346 arch/powerpc/platforms/cell/interrupt.c 			 (node << 4)		/* dest node */ |
node              383 arch/powerpc/platforms/cell/interrupt.c 	int node = cpu >> 1;
node              387 arch/powerpc/platforms/cell/interrupt.c 		  CBE_IIC_IR_DEST_NODE(node);
node              683 arch/powerpc/platforms/cell/iommu.c 	int node;
node              688 arch/powerpc/platforms/cell/iommu.c 	for_each_online_node(node) {
node              689 arch/powerpc/platforms/cell/iommu.c 		if (cell_iommu_find_ioc(node, &base))
node              696 arch/powerpc/platforms/cell/iommu.c 		pr_debug("iommu: cleaning up iommu on node %d\n", node);
node              370 arch/powerpc/platforms/cell/pmu.c 	int rc, node;
node              372 arch/powerpc/platforms/cell/pmu.c 	for_each_online_node(node) {
node              374 arch/powerpc/platforms/cell/pmu.c 					       (node << IIC_IRQ_NODE_SHIFT));
node              377 arch/powerpc/platforms/cell/pmu.c 			       node);
node              385 arch/powerpc/platforms/cell/pmu.c 			       node);
node              394 arch/powerpc/platforms/cell/pmu.c void cbe_sync_irq(int node)
node              400 arch/powerpc/platforms/cell/pmu.c 			       | (node << IIC_IRQ_NODE_SHIFT));
node              404 arch/powerpc/platforms/cell/pmu.c 		"for node %d\n", irq, node);
node              158 arch/powerpc/platforms/cell/setup.c 	int node;
node              175 arch/powerpc/platforms/cell/setup.c 	for_each_online_node(node) {
node              176 arch/powerpc/platforms/cell/setup.c 		if (cbe_get_cpu_mic_tm_regs(cbe_node_to_cpu(node)) == NULL)
node              178 arch/powerpc/platforms/cell/setup.c 		platform_device_register_simple("cbe-mic", node, NULL, 0);
node              556 arch/powerpc/platforms/cell/spu_base.c 	sysfs_add_device_to_node(&spu->dev, spu->node);
node              595 arch/powerpc/platforms/cell/spu_base.c 	mutex_lock(&cbe_spu_info[spu->node].list_mutex);
node              596 arch/powerpc/platforms/cell/spu_base.c 	list_add(&spu->cbe_list, &cbe_spu_info[spu->node].spus);
node              597 arch/powerpc/platforms/cell/spu_base.c 	cbe_spu_info[spu->node].n_spus++;
node              598 arch/powerpc/platforms/cell/spu_base.c 	mutex_unlock(&cbe_spu_info[spu->node].list_mutex);
node               83 arch/powerpc/platforms/cell/spu_manage.c 		nid = spu->node;
node              121 arch/powerpc/platforms/cell/spu_manage.c 	struct device_node *node = spu->devnode;
node              126 arch/powerpc/platforms/cell/spu_manage.c 	spu->name = of_get_property(node, "name", NULL);
node              130 arch/powerpc/platforms/cell/spu_manage.c 	prop = of_get_property(node, "local-store", NULL);
node              137 arch/powerpc/platforms/cell/spu_manage.c 		spu_map_prop_old(spu, node, "local-store");
node              141 arch/powerpc/platforms/cell/spu_manage.c 	prop = of_get_property(node, "problem", NULL);
node              146 arch/powerpc/platforms/cell/spu_manage.c 	spu->problem = spu_map_prop_old(spu, node, "problem");
node              150 arch/powerpc/platforms/cell/spu_manage.c 	spu->priv2 = spu_map_prop_old(spu, node, "priv2");
node              155 arch/powerpc/platforms/cell/spu_manage.c 		spu->priv1 = spu_map_prop_old(spu, node, "priv1");
node              266 arch/powerpc/platforms/cell/spu_manage.c 	struct device_node *node;
node              270 arch/powerpc/platforms/cell/spu_manage.c 	for_each_node_by_type(node, "spe") {
node              271 arch/powerpc/platforms/cell/spu_manage.c 		ret = fn(node);
node              274 arch/powerpc/platforms/cell/spu_manage.c 				__func__, node);
node              275 arch/powerpc/platforms/cell/spu_manage.c 			of_node_put(node);
node              292 arch/powerpc/platforms/cell/spu_manage.c 	spu->node = of_node_to_nid(spe);
node              293 arch/powerpc/platforms/cell/spu_manage.c 	if (spu->node >= MAX_NUMNODES) {
node              295 arch/powerpc/platforms/cell/spu_manage.c 		       " node number too big\n", spe, spu->node);
node              364 arch/powerpc/platforms/cell/spu_manage.c static struct spu *spu_lookup_reg(int node, u32 reg)
node              369 arch/powerpc/platforms/cell/spu_manage.c 	list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              379 arch/powerpc/platforms/cell/spu_manage.c 	int node, i;
node              383 arch/powerpc/platforms/cell/spu_manage.c 	for (node = 0; node < MAX_NUMNODES; node++) {
node              387 arch/powerpc/platforms/cell/spu_manage.c 			spu = spu_lookup_reg(node, reg);
node               68 arch/powerpc/platforms/cell/spu_priv1_mmio.c 	if (nr_cpus_node(spu->node)) {
node               69 arch/powerpc/platforms/cell/spu_priv1_mmio.c 		const struct cpumask *spumask = cpumask_of_node(spu->node),
node              306 arch/powerpc/platforms/cell/spufs/inode.c 	int count, node;
node              353 arch/powerpc/platforms/cell/spufs/inode.c 		for (node = 0; node < MAX_NUMNODES; node++) {
node              354 arch/powerpc/platforms/cell/spufs/inode.c 			if ((cbe_spu_info[node].n_spus - atomic_read(
node              355 arch/powerpc/platforms/cell/spufs/inode.c 				&cbe_spu_info[node].reserved_spus)) >= count)
node              359 arch/powerpc/platforms/cell/spufs/inode.c 		if (node == MAX_NUMNODES) {
node              139 arch/powerpc/platforms/cell/spufs/sched.c 	int node;
node              142 arch/powerpc/platforms/cell/spufs/sched.c 		node = ctx->spu->node;
node              147 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node              149 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              155 arch/powerpc/platforms/cell/spufs/sched.c static int __node_allowed(struct spu_context *ctx, int node)
node              157 arch/powerpc/platforms/cell/spufs/sched.c 	if (nr_cpus_node(node)) {
node              158 arch/powerpc/platforms/cell/spufs/sched.c 		const struct cpumask *mask = cpumask_of_node(node);
node              167 arch/powerpc/platforms/cell/spufs/sched.c static int node_allowed(struct spu_context *ctx, int node)
node              172 arch/powerpc/platforms/cell/spufs/sched.c 	rval = __node_allowed(ctx, node);
node              180 arch/powerpc/platforms/cell/spufs/sched.c 	int node;
node              188 arch/powerpc/platforms/cell/spufs/sched.c 	for_each_online_node(node) {
node              191 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node              192 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              201 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              217 arch/powerpc/platforms/cell/spufs/sched.c 		atomic_inc(&cbe_spu_info[spu->node].reserved_spus);
node              253 arch/powerpc/platforms/cell/spufs/sched.c 	BUG_ON(!mutex_is_locked(&cbe_spu_info[spu->node].list_mutex));
node              296 arch/powerpc/platforms/cell/spufs/sched.c 	int node, n;
node              302 arch/powerpc/platforms/cell/spufs/sched.c 	node = cpu_to_node(raw_smp_processor_id());
node              303 arch/powerpc/platforms/cell/spufs/sched.c 	for (n = 0; n < MAX_NUMNODES; n++, node++) {
node              315 arch/powerpc/platforms/cell/spufs/sched.c 		node = (node < MAX_NUMNODES) ? node : 0;
node              316 arch/powerpc/platforms/cell/spufs/sched.c 		if (!node_allowed(ctx, node))
node              320 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node              321 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              328 arch/powerpc/platforms/cell/spufs/sched.c 			mutex_unlock(&cbe_spu_info[node].list_mutex);
node              332 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              335 arch/powerpc/platforms/cell/spufs/sched.c 				mutex_unlock(&cbe_spu_info[node].list_mutex);
node              339 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              368 arch/powerpc/platforms/cell/spufs/sched.c static struct spu *ctx_location(struct spu *ref, int offset, int node)
node              375 arch/powerpc/platforms/cell/spufs/sched.c 			BUG_ON(spu->node != node);
node              383 arch/powerpc/platforms/cell/spufs/sched.c 			BUG_ON(spu->node != node);
node              433 arch/powerpc/platforms/cell/spufs/sched.c 		atomic_dec(&cbe_spu_info[spu->node].reserved_spus);
node              562 arch/powerpc/platforms/cell/spufs/sched.c 	int node, n;
node              572 arch/powerpc/platforms/cell/spufs/sched.c 			node = aff_ref_spu->node;
node              574 arch/powerpc/platforms/cell/spufs/sched.c 			mutex_lock(&cbe_spu_info[node].list_mutex);
node              575 arch/powerpc/platforms/cell/spufs/sched.c 			spu = ctx_location(aff_ref_spu, ctx->aff_offset, node);
node              578 arch/powerpc/platforms/cell/spufs/sched.c 			mutex_unlock(&cbe_spu_info[node].list_mutex);
node              585 arch/powerpc/platforms/cell/spufs/sched.c 	node = cpu_to_node(raw_smp_processor_id());
node              586 arch/powerpc/platforms/cell/spufs/sched.c 	for (n = 0; n < MAX_NUMNODES; n++, node++) {
node              587 arch/powerpc/platforms/cell/spufs/sched.c 		node = (node < MAX_NUMNODES) ? node : 0;
node              588 arch/powerpc/platforms/cell/spufs/sched.c 		if (!node_allowed(ctx, node))
node              591 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node              592 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              596 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              605 arch/powerpc/platforms/cell/spufs/sched.c 	mutex_unlock(&cbe_spu_info[node].list_mutex);
node              621 arch/powerpc/platforms/cell/spufs/sched.c 	int node, n;
node              633 arch/powerpc/platforms/cell/spufs/sched.c 	node = cpu_to_node(raw_smp_processor_id());
node              634 arch/powerpc/platforms/cell/spufs/sched.c 	for (n = 0; n < MAX_NUMNODES; n++, node++) {
node              635 arch/powerpc/platforms/cell/spufs/sched.c 		node = (node < MAX_NUMNODES) ? node : 0;
node              636 arch/powerpc/platforms/cell/spufs/sched.c 		if (!node_allowed(ctx, node))
node              639 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node              640 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list) {
node              651 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              685 arch/powerpc/platforms/cell/spufs/sched.c 			mutex_lock(&cbe_spu_info[node].list_mutex);
node              686 arch/powerpc/platforms/cell/spufs/sched.c 			cbe_spu_info[node].nr_active--;
node              688 arch/powerpc/platforms/cell/spufs/sched.c 			mutex_unlock(&cbe_spu_info[node].list_mutex);
node              707 arch/powerpc/platforms/cell/spufs/sched.c 	int node = spu->node;
node              712 arch/powerpc/platforms/cell/spufs/sched.c 	mutex_lock(&cbe_spu_info[node].list_mutex);
node              715 arch/powerpc/platforms/cell/spufs/sched.c 		cbe_spu_info[node].nr_active++;
node              719 arch/powerpc/platforms/cell/spufs/sched.c 	mutex_unlock(&cbe_spu_info[node].list_mutex);
node              753 arch/powerpc/platforms/cell/spufs/sched.c 	int node = spu->node;
node              755 arch/powerpc/platforms/cell/spufs/sched.c 	mutex_lock(&cbe_spu_info[node].list_mutex);
node              756 arch/powerpc/platforms/cell/spufs/sched.c 	cbe_spu_info[node].nr_active--;
node              762 arch/powerpc/platforms/cell/spufs/sched.c 	mutex_unlock(&cbe_spu_info[node].list_mutex);
node              825 arch/powerpc/platforms/cell/spufs/sched.c static struct spu_context *grab_runnable_context(int prio, int node)
node              837 arch/powerpc/platforms/cell/spufs/sched.c 			if (__node_allowed(ctx, node)) {
node              856 arch/powerpc/platforms/cell/spufs/sched.c 		new = grab_runnable_context(max_prio, spu->node);
node              929 arch/powerpc/platforms/cell/spufs/sched.c 	new = grab_runnable_context(ctx->prio + 1, spu->node);
node              957 arch/powerpc/platforms/cell/spufs/sched.c 	int nr_active = 0, node;
node              959 arch/powerpc/platforms/cell/spufs/sched.c 	for (node = 0; node < MAX_NUMNODES; node++)
node              960 arch/powerpc/platforms/cell/spufs/sched.c 		nr_active += cbe_spu_info[node].nr_active;
node              997 arch/powerpc/platforms/cell/spufs/sched.c 	int node;
node             1002 arch/powerpc/platforms/cell/spufs/sched.c 		for (node = 0; node < MAX_NUMNODES; node++) {
node             1003 arch/powerpc/platforms/cell/spufs/sched.c 			struct mutex *mtx = &cbe_spu_info[node].list_mutex;
node             1006 arch/powerpc/platforms/cell/spufs/sched.c 			list_for_each_entry(spu, &cbe_spu_info[node].spus,
node             1032 arch/powerpc/platforms/cell/spufs/sched.c 	int node;
node             1053 arch/powerpc/platforms/cell/spufs/sched.c 		node = spu->node;
node             1055 arch/powerpc/platforms/cell/spufs/sched.c 			atomic_dec(&cbe_spu_info[node].busy_spus);
node             1057 arch/powerpc/platforms/cell/spufs/sched.c 			atomic_inc(&cbe_spu_info[node].busy_spus);
node             1129 arch/powerpc/platforms/cell/spufs/sched.c 	int node;
node             1137 arch/powerpc/platforms/cell/spufs/sched.c 	for (node = 0; node < MAX_NUMNODES; node++) {
node             1138 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_lock(&cbe_spu_info[node].list_mutex);
node             1139 arch/powerpc/platforms/cell/spufs/sched.c 		list_for_each_entry(spu, &cbe_spu_info[node].spus, cbe_list)
node             1142 arch/powerpc/platforms/cell/spufs/sched.c 		mutex_unlock(&cbe_spu_info[node].list_mutex);
node              259 arch/powerpc/platforms/chrp/setup.c 	struct device_node *node;
node              267 arch/powerpc/platforms/chrp/setup.c 	node = of_find_node_by_path("/");
node              268 arch/powerpc/platforms/chrp/setup.c 	if (!node)
node              270 arch/powerpc/platforms/chrp/setup.c 	property = of_get_property(node, "model", NULL);
node              279 arch/powerpc/platforms/chrp/setup.c 	of_node_put(node);
node              280 arch/powerpc/platforms/chrp/setup.c 	node = of_find_node_by_path(property);
node              281 arch/powerpc/platforms/chrp/setup.c 	if (!node)
node              283 arch/powerpc/platforms/chrp/setup.c 	if (!of_node_is_type(node, "serial"))
node              290 arch/powerpc/platforms/chrp/setup.c 	if (of_node_name_eq(node, "failsafe") || of_node_name_eq(node, "serial"))
node              293 arch/powerpc/platforms/chrp/setup.c 	of_node_put(node);
node               35 arch/powerpc/platforms/maple/pci.c static int __init fixup_one_level_bus_range(struct device_node *node, int higher)
node               37 arch/powerpc/platforms/maple/pci.c 	for (; node != 0;node = node->sibling) {
node               43 arch/powerpc/platforms/maple/pci.c 		class_code = of_get_property(node, "class-code", NULL);
node               47 arch/powerpc/platforms/maple/pci.c 		bus_range = of_get_property(node, "bus-range", &len);
node               52 arch/powerpc/platforms/maple/pci.c 		higher = fixup_one_level_bus_range(node->child, higher);
node               30 arch/powerpc/platforms/pasemi/misc.c static int __init find_i2c_driver(struct device_node *node,
node               36 arch/powerpc/platforms/pasemi/misc.c 		if (!of_device_is_compatible(node, i2c_devices[i].of_device))
node               50 arch/powerpc/platforms/pasemi/misc.c 	struct device_node *node;
node               59 arch/powerpc/platforms/pasemi/misc.c 		node = NULL;
node               60 arch/powerpc/platforms/pasemi/misc.c 		while ((node = of_get_next_child(adap_node, node))) {
node               65 arch/powerpc/platforms/pasemi/misc.c 			addr = of_get_property(node, "reg", &len);
node               72 arch/powerpc/platforms/pasemi/misc.c 			info.irq = irq_of_parse_and_map(node, 0);
node               76 arch/powerpc/platforms/pasemi/misc.c 			if (find_i2c_driver(node, &info) < 0)
node               90 arch/powerpc/platforms/powermac/bootx_init.c 					 unsigned long node,
node               93 arch/powerpc/platforms/powermac/bootx_init.c 	struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node);
node              217 arch/powerpc/platforms/powermac/bootx_init.c 					       unsigned long node,
node              220 arch/powerpc/platforms/powermac/bootx_init.c 	struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node);
node              240 arch/powerpc/platforms/powermac/bootx_init.c 		bootx_node_chosen = node;
node              242 arch/powerpc/platforms/powermac/bootx_init.c 	if (node == bootx_info->dispDeviceRegEntryOffset) {
node              275 arch/powerpc/platforms/powermac/bootx_init.c 					      unsigned long node,
node              278 arch/powerpc/platforms/powermac/bootx_init.c 	struct bootx_dt_node *np = (struct bootx_dt_node *)(base + node);
node              318 arch/powerpc/platforms/powermac/bootx_init.c 		if (node == bootx_node_chosen && !strcmp(namep, "bootargs"))
node              329 arch/powerpc/platforms/powermac/bootx_init.c 	if (node == bootx_node_chosen) {
node              334 arch/powerpc/platforms/powermac/bootx_init.c 	else if (node == bootx_info->dispDeviceRegEntryOffset)
node              115 arch/powerpc/platforms/powermac/feature.c typedef long (*feature_call)(struct device_node *node, long param, long value);
node              136 arch/powerpc/platforms/powermac/feature.c static inline int simple_feature_tweak(struct device_node *node, int type,
node              142 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, type);
node              158 arch/powerpc/platforms/powermac/feature.c static long ohare_htw_scc_enable(struct device_node *node, long param,
node              168 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              171 arch/powerpc/platforms/powermac/feature.c 	if (of_node_name_eq(node, "ch-a"))
node              173 arch/powerpc/platforms/powermac/feature.c 	else if (of_node_name_eq(node, "ch-b"))
node              251 arch/powerpc/platforms/powermac/feature.c static long ohare_floppy_enable(struct device_node *node, long param,
node              254 arch/powerpc/platforms/powermac/feature.c 	return simple_feature_tweak(node, macio_ohare,
node              258 arch/powerpc/platforms/powermac/feature.c static long ohare_mesh_enable(struct device_node *node, long param, long value)
node              260 arch/powerpc/platforms/powermac/feature.c 	return simple_feature_tweak(node, macio_ohare,
node              264 arch/powerpc/platforms/powermac/feature.c static long ohare_ide_enable(struct device_node *node, long param, long value)
node              272 arch/powerpc/platforms/powermac/feature.c 			simple_feature_tweak(node, macio_ohare,
node              274 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_ohare,
node              277 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_ohare,
node              284 arch/powerpc/platforms/powermac/feature.c static long ohare_ide_reset(struct device_node *node, long param, long value)
node              288 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_ohare,
node              291 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_ohare,
node              298 arch/powerpc/platforms/powermac/feature.c static long ohare_sleep_state(struct device_node *node, long param, long value)
node              313 arch/powerpc/platforms/powermac/feature.c static long heathrow_modem_enable(struct device_node *node, long param,
node              320 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, macio_unknown);
node              357 arch/powerpc/platforms/powermac/feature.c static long heathrow_floppy_enable(struct device_node *node, long param,
node              360 arch/powerpc/platforms/powermac/feature.c 	return simple_feature_tweak(node, macio_unknown,
node              366 arch/powerpc/platforms/powermac/feature.c static long heathrow_mesh_enable(struct device_node *node, long param,
node              372 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, macio_unknown);
node              395 arch/powerpc/platforms/powermac/feature.c static long heathrow_ide_enable(struct device_node *node, long param,
node              400 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              403 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              410 arch/powerpc/platforms/powermac/feature.c static long heathrow_ide_reset(struct device_node *node, long param,
node              415 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              418 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              425 arch/powerpc/platforms/powermac/feature.c static long heathrow_bmac_enable(struct device_node *node, long param,
node              431 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              454 arch/powerpc/platforms/powermac/feature.c static long heathrow_sound_enable(struct device_node *node, long param,
node              467 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              581 arch/powerpc/platforms/powermac/feature.c static long heathrow_sleep_state(struct device_node *node, long param,
node              598 arch/powerpc/platforms/powermac/feature.c static long core99_scc_enable(struct device_node *node, long param, long value)
node              605 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              608 arch/powerpc/platforms/powermac/feature.c 	if (of_node_name_eq(node, "ch-a"))
node              610 arch/powerpc/platforms/powermac/feature.c 	else if (of_node_name_eq(node, "ch-b"))
node              699 arch/powerpc/platforms/powermac/feature.c core99_modem_enable(struct device_node *node, long param, long value)
node              706 arch/powerpc/platforms/powermac/feature.c 	if (node == NULL) {
node              709 arch/powerpc/platforms/powermac/feature.c 		node = macio_chips[0].of_node;
node              711 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              751 arch/powerpc/platforms/powermac/feature.c pangea_modem_enable(struct device_node *node, long param, long value)
node              758 arch/powerpc/platforms/powermac/feature.c 	if (node == NULL) {
node              762 arch/powerpc/platforms/powermac/feature.c 		node = macio_chips[0].of_node;
node              764 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              806 arch/powerpc/platforms/powermac/feature.c core99_ata100_enable(struct device_node *node, long value)
node              826 arch/powerpc/platforms/powermac/feature.c 		if (pci_device_from_OF_node(node, &pbus, &pid) == 0)
node              841 arch/powerpc/platforms/powermac/feature.c core99_ide_enable(struct device_node *node, long param, long value)
node              848 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              851 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              854 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              857 arch/powerpc/platforms/powermac/feature.c 		return core99_ata100_enable(node, value);
node              864 arch/powerpc/platforms/powermac/feature.c core99_ide_reset(struct device_node *node, long param, long value)
node              868 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              871 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              874 arch/powerpc/platforms/powermac/feature.c 		return simple_feature_tweak(node, macio_unknown,
node              882 arch/powerpc/platforms/powermac/feature.c core99_gmac_enable(struct device_node *node, long param, long value)
node              899 arch/powerpc/platforms/powermac/feature.c core99_gmac_phy_reset(struct device_node *node, long param, long value)
node              924 arch/powerpc/platforms/powermac/feature.c core99_sound_chip_enable(struct device_node *node, long param, long value)
node              929 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              954 arch/powerpc/platforms/powermac/feature.c core99_airport_enable(struct device_node *node, long param, long value)
node              960 arch/powerpc/platforms/powermac/feature.c 	macio = macio_find(node, 0);
node              967 arch/powerpc/platforms/powermac/feature.c 	if (node != macio->of_node &&
node              968 arch/powerpc/platforms/powermac/feature.c 	    (!node->parent || node->parent != macio->of_node))
node             1041 arch/powerpc/platforms/powermac/feature.c core99_reset_cpu(struct device_node *node, long param, long value)
node             1082 arch/powerpc/platforms/powermac/feature.c core99_usb_enable(struct device_node *node, long param, long value)
node             1095 arch/powerpc/platforms/powermac/feature.c 	prop = of_get_property(node, "AAPL,clock-id", NULL);
node             1235 arch/powerpc/platforms/powermac/feature.c core99_firewire_enable(struct device_node *node, long param, long value)
node             1262 arch/powerpc/platforms/powermac/feature.c core99_firewire_cable_power(struct device_node *node, long param, long value)
node             1293 arch/powerpc/platforms/powermac/feature.c intrepid_aack_delay_enable(struct device_node *node, long param, long value)
node             1314 arch/powerpc/platforms/powermac/feature.c core99_read_gpio(struct device_node *node, long param, long value)
node             1323 arch/powerpc/platforms/powermac/feature.c core99_write_gpio(struct device_node *node, long param, long value)
node             1332 arch/powerpc/platforms/powermac/feature.c static long g5_gmac_enable(struct device_node *node, long param, long value)
node             1337 arch/powerpc/platforms/powermac/feature.c 	if (node == NULL)
node             1346 arch/powerpc/platforms/powermac/feature.c 		k2_skiplist[0] = node;
node             1357 arch/powerpc/platforms/powermac/feature.c static long g5_fw_enable(struct device_node *node, long param, long value)
node             1362 arch/powerpc/platforms/powermac/feature.c 	if (node == NULL)
node             1371 arch/powerpc/platforms/powermac/feature.c 		k2_skiplist[1] = node;
node             1382 arch/powerpc/platforms/powermac/feature.c static long g5_mpic_enable(struct device_node *node, long param, long value)
node             1385 arch/powerpc/platforms/powermac/feature.c 	struct device_node *parent = of_get_parent(node);
node             1402 arch/powerpc/platforms/powermac/feature.c static long g5_eth_phy_reset(struct device_node *node, long param, long value)
node             1412 arch/powerpc/platforms/powermac/feature.c 	phy = of_get_next_child(node, NULL);
node             1432 arch/powerpc/platforms/powermac/feature.c static long g5_i2s_enable(struct device_node *node, long param, long value)
node             1458 arch/powerpc/platforms/powermac/feature.c 	if (strncmp(node->name, "i2s-", 4))
node             1460 arch/powerpc/platforms/powermac/feature.c 	cell = node->name[4] - 'a';
node             1491 arch/powerpc/platforms/powermac/feature.c static long g5_reset_cpu(struct device_node *node, long param, long value)
node             1822 arch/powerpc/platforms/powermac/feature.c core99_sleep_state(struct device_node *node, long param, long value)
node             1855 arch/powerpc/platforms/powermac/feature.c generic_dev_can_wake(struct device_node *node, long param, long value)
node             1866 arch/powerpc/platforms/powermac/feature.c static long generic_get_mb_info(struct device_node *node, long param, long value)
node             2369 arch/powerpc/platforms/powermac/feature.c 	struct device_node *node;
node             2391 arch/powerpc/platforms/powermac/feature.c 	node = (struct device_node*)va_arg(args, void*);
node             2396 arch/powerpc/platforms/powermac/feature.c 	return func(node, param, value);
node             2617 arch/powerpc/platforms/powermac/feature.c 	struct device_node*	node;
node             2624 arch/powerpc/platforms/powermac/feature.c 	for_each_node_by_name(node, name) {
node             2627 arch/powerpc/platforms/powermac/feature.c 		if (of_device_is_compatible(node, compat))
node             2630 arch/powerpc/platforms/powermac/feature.c 	if (!node)
node             2635 arch/powerpc/platforms/powermac/feature.c 		if (macio_chips[i].of_node == node)
node             2641 arch/powerpc/platforms/powermac/feature.c 		printk(KERN_ERR "pmac_feature: %pOF skipped\n", node);
node             2644 arch/powerpc/platforms/powermac/feature.c 	addrp = of_get_pci_address(node, 0, &size, NULL);
node             2647 arch/powerpc/platforms/powermac/feature.c 		       node);
node             2650 arch/powerpc/platforms/powermac/feature.c 	addr = of_translate_address(node, addrp);
node             2653 arch/powerpc/platforms/powermac/feature.c 		       node);
node             2659 arch/powerpc/platforms/powermac/feature.c 		       node);
node             2663 arch/powerpc/platforms/powermac/feature.c 		const u32 *did = of_get_property(node, "device-id", NULL);
node             2671 arch/powerpc/platforms/powermac/feature.c 	macio_chips[i].of_node	= node;
node             2676 arch/powerpc/platforms/powermac/feature.c 	revp = of_get_property(node, "revision-id", NULL);
node              955 arch/powerpc/platforms/powermac/low_i2c.c struct pmac_i2c_bus *pmac_i2c_find_bus(struct device_node *node)
node              957 arch/powerpc/platforms/powermac/low_i2c.c 	struct device_node *p = of_node_get(node);
node             1235 arch/powerpc/platforms/powermac/low_i2c.c 	bus = pmac_i2c_find_bus(func->node);
node             1238 arch/powerpc/platforms/powermac/low_i2c.c 		       func->node);
node             1243 arch/powerpc/platforms/powermac/low_i2c.c 		       func->node);
node             1258 arch/powerpc/platforms/powermac/low_i2c.c 	inst->addr = pmac_i2c_get_dev_addr(func->node);
node               59 arch/powerpc/platforms/powermac/pci.c static int __init fixup_one_level_bus_range(struct device_node *node, int higher)
node               61 arch/powerpc/platforms/powermac/pci.c 	for (; node; node = node->sibling) {
node               67 arch/powerpc/platforms/powermac/pci.c 		class_code = of_get_property(node, "class-code", NULL);
node               71 arch/powerpc/platforms/powermac/pci.c 		bus_range = of_get_property(node, "bus-range", &len);
node               76 arch/powerpc/platforms/powermac/pci.c 		higher = fixup_one_level_bus_range(node->child, higher);
node              952 arch/powerpc/platforms/powermac/pci.c 	struct device_node* node;
node              956 arch/powerpc/platforms/powermac/pci.c 	node = pci_device_to_OF_node(dev);
node              963 arch/powerpc/platforms/powermac/pci.c 	    && !node) {
node              969 arch/powerpc/platforms/powermac/pci.c 	if (!node)
node              972 arch/powerpc/platforms/powermac/pci.c 	uninorth_child = node->parent &&
node              973 arch/powerpc/platforms/powermac/pci.c 		of_device_is_compatible(node->parent, "uni-north");
node              978 arch/powerpc/platforms/powermac/pci.c 	if (uninorth_child && of_node_name_eq(node, "firewire") &&
node              979 arch/powerpc/platforms/powermac/pci.c 	    (of_device_is_compatible(node, "pci106b,18") ||
node              980 arch/powerpc/platforms/powermac/pci.c 	     of_device_is_compatible(node, "pci106b,30") ||
node              981 arch/powerpc/platforms/powermac/pci.c 	     of_device_is_compatible(node, "pci11c1,5811"))) {
node              982 arch/powerpc/platforms/powermac/pci.c 		pmac_call_feature(PMAC_FTR_1394_CABLE_POWER, node, 0, 1);
node              983 arch/powerpc/platforms/powermac/pci.c 		pmac_call_feature(PMAC_FTR_1394_ENABLE, node, 0, 1);
node              986 arch/powerpc/platforms/powermac/pci.c 	if (uninorth_child && of_node_name_eq(node, "ethernet") &&
node              987 arch/powerpc/platforms/powermac/pci.c 	    of_device_is_compatible(node, "gmac")) {
node              988 arch/powerpc/platforms/powermac/pci.c 		pmac_call_feature(PMAC_FTR_GMAC_ENABLE, node, 0, 1);
node             1016 arch/powerpc/platforms/powermac/pci.c 	struct device_node *node = pci_device_to_OF_node(dev);
node             1021 arch/powerpc/platforms/powermac/pci.c 	if (dev->class == PCI_CLASS_SERIAL_USB_OHCI && !node)
node             1235 arch/powerpc/platforms/powermac/pci.c 	struct device_node *node = pci_bus_to_OF_node(bus);
node             1241 arch/powerpc/platforms/powermac/pci.c 	if (bus->self == NULL && (of_device_is_compatible(node, "u3-agp") ||
node             1242 arch/powerpc/platforms/powermac/pci.c 				  of_device_is_compatible(node, "u4-pcie") ||
node             1243 arch/powerpc/platforms/powermac/pci.c 				  of_device_is_compatible(node, "u3-ht")))
node               29 arch/powerpc/platforms/powermac/pfunc_base.c 	unsigned int irq = irq_of_parse_and_map(func->node, 0);
node               32 arch/powerpc/platforms/powermac/pfunc_base.c 	return request_irq(irq, macio_gpio_irq, 0, func->node->name, func);
node               37 arch/powerpc/platforms/powermac/pfunc_base.c 	unsigned int irq = irq_of_parse_and_map(func->node, 0);
node               59 arch/powerpc/platforms/powermac/pfunc_base.c 	    tmp, func->node, addr);
node              541 arch/powerpc/platforms/powermac/pfunc_core.c 	struct device_node	*node;
node              573 arch/powerpc/platforms/powermac/pfunc_core.c 		if (dev->node == np)
node              652 arch/powerpc/platforms/powermac/pfunc_core.c 		func->node = dev->node;
node              688 arch/powerpc/platforms/powermac/pfunc_core.c 	for (pp = dev->node->properties; pp != 0; pp = pp->next) {
node              729 arch/powerpc/platforms/powermac/pfunc_core.c 	dev->node = of_node_get(np);
node              944 arch/powerpc/platforms/powermac/pfunc_core.c 	DBG(" ** pmf_call_one(%pOF/%s) **\n", dev->node, func->name);
node              267 arch/powerpc/platforms/powermac/pic.c static int pmac_pic_host_match(struct irq_domain *h, struct device_node *node,
node              335 arch/powerpc/platforms/powermac/setup.c void note_scsi_host(struct device_node *node, void *host)
node              162 arch/powerpc/platforms/powernv/npu-dma.c 	pnv_pci_link_table_and_group(phb->hose->node, num,
node               31 arch/powerpc/platforms/powernv/opal-fadump.c void __init opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node)
node               39 arch/powerpc/platforms/powernv/opal-fadump.c 	dn = of_get_flat_dt_subnode_by_name(node, "dump");
node              623 arch/powerpc/platforms/powernv/opal-fadump.c void __init opal_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node)
node              635 arch/powerpc/platforms/powernv/opal-fadump.c 	dn = of_get_flat_dt_subnode_by_name(node, "dump");
node               52 arch/powerpc/platforms/powernv/opal-imc.c static void export_imc_mode_and_cmd(struct device_node *node,
node               65 arch/powerpc/platforms/powernv/opal-imc.c 	if (of_property_read_u32(node, "cb_offset", &cb_offset))
node               93 arch/powerpc/platforms/powernv/opal-imc.c static int imc_get_mem_addr_nest(struct device_node *node,
node              101 arch/powerpc/platforms/powernv/opal-imc.c 	nr_chips = of_property_count_u32_elems(node, "chip-id");
node              115 arch/powerpc/platforms/powernv/opal-imc.c 	if (of_property_read_u32_array(node, "chip-id", chipid_arr, nr_chips))
node              118 arch/powerpc/platforms/powernv/opal-imc.c 	if (of_property_read_u64_array(node, "base-addr", base_addr_arr,
node              231 arch/powerpc/platforms/powernv/opal-imc.c 	struct device_node *node;
node              234 arch/powerpc/platforms/powernv/opal-imc.c 	for_each_compatible_node(node, NULL, IMC_DTB_UNIT_COMPAT) {
node              235 arch/powerpc/platforms/powernv/opal-imc.c 		if (of_property_read_u32(node, "type", &type))
node              137 arch/powerpc/platforms/powernv/opal-irqchip.c static int opal_event_match(struct irq_domain *h, struct device_node *node,
node              140 arch/powerpc/platforms/powernv/opal-irqchip.c 	return irq_domain_get_of_node(h) == node;
node              106 arch/powerpc/platforms/powernv/opal-msglog.c struct memcons *memcons_init(struct device_node *node, const char *mc_prop_name)
node              111 arch/powerpc/platforms/powernv/opal-msglog.c 	if (of_property_read_u64(node, mc_prop_name, &mcaddr)) {
node              144 arch/powerpc/platforms/powernv/opal-powercap.c 	struct device_node *powercap, *node;
node              165 arch/powerpc/platforms/powernv/opal-powercap.c 	for_each_child_of_node(powercap, node) {
node              170 arch/powerpc/platforms/powernv/opal-powercap.c 		if (!of_property_read_u32(node, "powercap-min", &min)) {
node              175 arch/powerpc/platforms/powernv/opal-powercap.c 		if (!of_property_read_u32(node, "powercap-max", &max)) {
node              180 arch/powerpc/platforms/powernv/opal-powercap.c 		if (!of_property_read_u32(node, "powercap-current", &cur)) {
node              198 arch/powerpc/platforms/powernv/opal-powercap.c 		pcaps[i].pg.name = kasprintf(GFP_KERNEL, "%pOFn", node);
node               44 arch/powerpc/platforms/powernv/opal-prd.c 	struct device_node *parent, *node;
node               56 arch/powerpc/platforms/powernv/opal-prd.c 	for_each_child_of_node(parent, node) {
node               61 arch/powerpc/platforms/powernv/opal-prd.c 		addrp = of_get_address(node, 0, &range_size, NULL);
node               66 arch/powerpc/platforms/powernv/opal-prd.c 		label = of_get_property(node, "ibm,prd-label", NULL);
node               77 arch/powerpc/platforms/powernv/opal-prd.c 			of_node_put(node);
node              125 arch/powerpc/platforms/powernv/opal-psr.c 	struct device_node *psr, *node;
node              146 arch/powerpc/platforms/powernv/opal-psr.c 	for_each_child_of_node(psr, node) {
node              147 arch/powerpc/platforms/powernv/opal-psr.c 		if (of_property_read_u32(node, "handle",
node              152 arch/powerpc/platforms/powernv/opal-psr.c 		if (of_property_read_string(node, "label",
node              162 arch/powerpc/platforms/powernv/opal-sensor-groups.c 	struct device_node *sg, *node;
node              181 arch/powerpc/platforms/powernv/opal-sensor-groups.c 	for_each_child_of_node(sg, node) {
node              185 arch/powerpc/platforms/powernv/opal-sensor-groups.c 		ops = of_get_property(node, "ops", &len);
node              207 arch/powerpc/platforms/powernv/opal-sensor-groups.c 		if (of_property_read_u32(node, "sensor-group-id", &sgid)) {
node              212 arch/powerpc/platforms/powernv/opal-sensor-groups.c 		if (!of_property_read_u32(node, "ibm,chip-id", &chipid))
node              213 arch/powerpc/platforms/powernv/opal-sensor-groups.c 			sprintf(sgs[i].name, "%pOFn%d", node, chipid);
node              215 arch/powerpc/platforms/powernv/opal-sensor-groups.c 			sprintf(sgs[i].name, "%pOFn", node);
node               98 arch/powerpc/platforms/powernv/opal.c int __init early_init_dt_scan_opal(unsigned long node,
node              107 arch/powerpc/platforms/powernv/opal.c 	basep  = of_get_flat_dt_prop(node, "opal-base-address", &basesz);
node              108 arch/powerpc/platforms/powernv/opal.c 	entryp = of_get_flat_dt_prop(node, "opal-entry-address", &entrysz);
node              109 arch/powerpc/platforms/powernv/opal.c 	sizep = of_get_flat_dt_prop(node, "opal-runtime-size", &runtimesz);
node              125 arch/powerpc/platforms/powernv/opal.c 	if (of_flat_dt_is_compatible(node, "ibm,opal-v3")) {
node              135 arch/powerpc/platforms/powernv/opal.c int __init early_init_dt_scan_recoverable_ranges(unsigned long node,
node              144 arch/powerpc/platforms/powernv/opal.c 	prop = of_get_flat_dt_prop(node, "mcheck-recoverable-ranges", &psize);
node              387 arch/powerpc/platforms/powernv/pci-ioda-tce.c long pnv_pci_link_table_and_group(int node, int num,
node              397 arch/powerpc/platforms/powernv/pci-ioda-tce.c 			node);
node             1807 arch/powerpc/platforms/powernv/pci-ioda.c 	table_pages = alloc_pages_node(pe->phb->hose->node, GFP_KERNEL,
node             2220 arch/powerpc/platforms/powernv/pci-ioda.c 	tbl = pnv_pci_table_alloc(phb->hose->node);
node             2226 arch/powerpc/platforms/powernv/pci-ioda.c 	pnv_pci_link_table_and_group(phb->hose->node, 0, tbl, &pe->table_group);
node             2244 arch/powerpc/platforms/powernv/pci-ioda.c 	tce_mem = alloc_pages_node(phb->hose->node, GFP_KERNEL,
node             2279 arch/powerpc/platforms/powernv/pci-ioda.c 	iommu_init_table(tbl, phb->hose->node, 0, 0);
node             2327 arch/powerpc/platforms/powernv/pci-ioda.c 	pnv_pci_link_table_and_group(phb->hose->node, num,
node             2368 arch/powerpc/platforms/powernv/pci-ioda.c 	int nid = pe->phb->hose->node;
node             2449 arch/powerpc/platforms/powernv/pci-ioda.c 	iommu_init_table(tbl, pe->phb->hose->node, res_start, res_end);
node             3176 arch/powerpc/platforms/powernv/pci-ioda.c 	list_for_each_entry(child, &bus->children, node)
node              241 arch/powerpc/platforms/powernv/pci.h extern long pnv_pci_link_table_and_group(int node, int num,
node               36 arch/powerpc/platforms/powernv/powernv.h struct memcons *memcons_init(struct device_node *node, const char *mc_prop_name);
node               21 arch/powerpc/platforms/powernv/ultravisor.c int __init early_init_dt_scan_ultravisor(unsigned long node, const char *uname,
node               24 arch/powerpc/platforms/powernv/ultravisor.c 	if (!of_flat_dt_is_compatible(node, "ibm,ultravisor"))
node               48 arch/powerpc/platforms/powernv/ultravisor.c 	struct device_node *node;
node               53 arch/powerpc/platforms/powernv/ultravisor.c 	node = of_find_compatible_node(NULL, NULL, "ibm,uv-firmware");
node               54 arch/powerpc/platforms/powernv/ultravisor.c 	if (!node)
node               57 arch/powerpc/platforms/powernv/ultravisor.c 	uv_memcons = memcons_init(node, "memcons");
node               49 arch/powerpc/platforms/powernv/vas.c 	INIT_LIST_HEAD(&vinst->node);
node               82 arch/powerpc/platforms/powernv/vas.c 	list_add(&vinst->node, &vas_instances);
node              112 arch/powerpc/platforms/powernv/vas.c 		vinst = list_entry(ent, struct vas_instance, node);
node              308 arch/powerpc/platforms/powernv/vas.h 	struct list_head node;
node              258 arch/powerpc/platforms/ps3/os-area.c static void os_area_set_property(struct device_node *node,
node              262 arch/powerpc/platforms/ps3/os-area.c 	struct property *tmp = of_find_property(node, prop->name, NULL);
node              266 arch/powerpc/platforms/ps3/os-area.c 		of_remove_property(node, tmp);
node              269 arch/powerpc/platforms/ps3/os-area.c 	result = of_add_property(node, prop);
node              281 arch/powerpc/platforms/ps3/os-area.c static void __init os_area_get_property(struct device_node *node,
node              284 arch/powerpc/platforms/ps3/os-area.c 	const struct property *tmp = of_find_property(node, prop->name, NULL);
node              675 arch/powerpc/platforms/ps3/os-area.c 	struct device_node *node;
node              680 arch/powerpc/platforms/ps3/os-area.c 	node = of_find_node_by_path("/");
node              681 arch/powerpc/platforms/ps3/os-area.c 	if (node) {
node              682 arch/powerpc/platforms/ps3/os-area.c 		os_area_set_property(node, &property_rtc_diff);
node              683 arch/powerpc/platforms/ps3/os-area.c 		of_node_put(node);
node              774 arch/powerpc/platforms/ps3/os-area.c 	struct device_node *node;
node              778 arch/powerpc/platforms/ps3/os-area.c 	node = of_find_node_by_path("/");
node              780 arch/powerpc/platforms/ps3/os-area.c 	if (!saved_params.valid && node) {
node              782 arch/powerpc/platforms/ps3/os-area.c 		os_area_get_property(node, &property_rtc_diff);
node              783 arch/powerpc/platforms/ps3/os-area.c 		os_area_get_property(node, &property_av_multi_out);
node              789 arch/powerpc/platforms/ps3/os-area.c 	if (node) {
node              790 arch/powerpc/platforms/ps3/os-area.c 		os_area_set_property(node, &property_rtc_diff);
node              791 arch/powerpc/platforms/ps3/os-area.c 		os_area_set_property(node, &property_av_multi_out);
node              792 arch/powerpc/platforms/ps3/os-area.c 		of_node_put(node);
node              141 arch/powerpc/platforms/pseries/firmware.c static int __init probe_fw_features(unsigned long node, const char *uname, int
node              153 arch/powerpc/platforms/pseries/firmware.c 		prop = of_get_flat_dt_prop(node, "ibm,hypertas-functions",
node              164 arch/powerpc/platforms/pseries/firmware.c 		prop = of_get_flat_dt_prop(node, "ibm,architecture-vec-5",
node               72 arch/powerpc/platforms/pseries/hvcserver.c 		pi = list_entry(element, struct hvcs_partner_info, node);
node              183 arch/powerpc/platforms/pseries/hvcserver.c 		list_add_tail(&(next_partner_info->node), head);
node               42 arch/powerpc/platforms/pseries/iommu.c static struct iommu_table_group *iommu_pseries_alloc_group(int node)
node               48 arch/powerpc/platforms/pseries/iommu.c 			   node);
node               52 arch/powerpc/platforms/pseries/iommu.c 	tbl = kzalloc_node(sizeof(struct iommu_table), GFP_KERNEL, node);
node              484 arch/powerpc/platforms/pseries/iommu.c 	struct device_node *node;
node              488 arch/powerpc/platforms/pseries/iommu.c 	node = phb->dn;
node              490 arch/powerpc/platforms/pseries/iommu.c 	basep = of_get_property(node, "linux,tce-base", NULL);
node              491 arch/powerpc/platforms/pseries/iommu.c 	sizep = of_get_property(node, "linux,tce-size", NULL);
node              622 arch/powerpc/platforms/pseries/iommu.c 	pci->table_group = iommu_pseries_alloc_group(pci->phb->node);
node              627 arch/powerpc/platforms/pseries/iommu.c 	iommu_init_table(tbl, pci->phb->node, 0, 0);
node              704 arch/powerpc/platforms/pseries/iommu.c 		ppci->table_group = iommu_pseries_alloc_group(ppci->phb->node);
node              709 arch/powerpc/platforms/pseries/iommu.c 		iommu_init_table(tbl, ppci->phb->node, 0, 0);
node              734 arch/powerpc/platforms/pseries/iommu.c 		PCI_DN(dn)->table_group = iommu_pseries_alloc_group(phb->node);
node              738 arch/powerpc/platforms/pseries/iommu.c 		iommu_init_table(tbl, phb->node, 0, 0);
node             1183 arch/powerpc/platforms/pseries/iommu.c 		pci->table_group = iommu_pseries_alloc_group(pci->phb->node);
node             1188 arch/powerpc/platforms/pseries/iommu.c 		iommu_init_table(tbl, pci->phb->node, 0, 0);
node              306 arch/powerpc/platforms/pseries/papr_scm.c static inline int papr_scm_node(int node)
node              311 arch/powerpc/platforms/pseries/papr_scm.c 	if ((node == NUMA_NO_NODE) || node_online(node))
node              312 arch/powerpc/platforms/pseries/papr_scm.c 		return node;
node              316 arch/powerpc/platforms/pseries/papr_scm.c 		dist = node_distance(node, nid);
node              242 arch/powerpc/platforms/pseries/reconfig.c 	struct device_node *node;
node              245 arch/powerpc/platforms/pseries/reconfig.c 	if ((node = of_find_node_by_path(buf)))
node              246 arch/powerpc/platforms/pseries/reconfig.c 		rv = pSeries_reconfig_remove_node(node);
node              248 arch/powerpc/platforms/pseries/reconfig.c 	of_node_put(node);
node              490 arch/powerpc/platforms/pseries/rtas-fadump.c void __init rtas_fadump_dt_scan(struct fw_dump *fadump_conf, u64 node)
node              500 arch/powerpc/platforms/pseries/rtas-fadump.c 	token = of_get_flat_dt_prop(node, "ibm,configure-kernel-dump", NULL);
node              515 arch/powerpc/platforms/pseries/rtas-fadump.c 	fdm_active = of_get_flat_dt_prop(node, "ibm,kernel-dump", NULL);
node              528 arch/powerpc/platforms/pseries/rtas-fadump.c 	sections = of_get_flat_dt_prop(node, "ibm,configure-kernel-dump-sizes",
node              456 arch/powerpc/platforms/pseries/setup.c 	struct device_node *node;
node              460 arch/powerpc/platforms/pseries/setup.c 	for_each_child_of_node(root, node) {
node              461 arch/powerpc/platforms/pseries/setup.c 		if (!of_node_is_type(node, "pci") &&
node              462 arch/powerpc/platforms/pseries/setup.c 		    !of_node_is_type(node, "pciex"))
node              465 arch/powerpc/platforms/pseries/setup.c 		phb = pcibios_alloc_controller(node);
node              469 arch/powerpc/platforms/pseries/setup.c 		pci_process_bridge_OF_ranges(phb, node, 0);
node              230 arch/powerpc/sysdev/cpm2_pic.c void cpm2_pic_init(struct device_node *node)
node              264 arch/powerpc/sysdev/cpm2_pic.c 	cpm2_pic_host = irq_domain_add_linear(node, 64, &cpm2_pic_host_ops, NULL);
node               15 arch/powerpc/sysdev/dcr.c static struct device_node *find_dcr_parent(struct device_node *node)
node               20 arch/powerpc/sysdev/dcr.c 	for (par = of_node_get(node); par;) {
node              180 arch/powerpc/sysdev/ehv_pic.c static int ehv_pic_host_match(struct irq_domain *h, struct device_node *node,
node              185 arch/powerpc/sysdev/ehv_pic.c 	return of_node == NULL || of_node == node;
node              186 arch/powerpc/sysdev/fsl_lbc.c 			     struct device_node *node)
node              198 arch/powerpc/sysdev/fsl_lbc.c 	if (of_device_is_compatible(node, "fsl,elbc"))
node              181 arch/powerpc/sysdev/fsl_pci.c 	struct device_node *node;
node              183 arch/powerpc/sysdev/fsl_pci.c 	node = of_find_node_by_type(NULL, "memory");
node              184 arch/powerpc/sysdev/fsl_pci.c 	if (!node) {
node              189 arch/powerpc/sysdev/fsl_pci.c 	return of_property_read_bool(node, "linux,usable-memory");
node             1310 arch/powerpc/sysdev/fsl_pci.c 	struct device_node *node;
node             1313 arch/powerpc/sysdev/fsl_pci.c 	node = pdev->dev.of_node;
node             1314 arch/powerpc/sysdev/fsl_pci.c 	ret = fsl_add_bridge(pdev, fsl_pci_primary == node);
node              125 arch/powerpc/sysdev/fsl_rio.h 	struct device_node *node);
node              328 arch/powerpc/sysdev/fsl_rmu.c 					&fsl_dbell->mport[i]->dbells, node) {
node             1065 arch/powerpc/sysdev/fsl_rmu.c int fsl_rio_setup_rmu(struct rio_mport *mport, struct device_node *node)
node             1079 arch/powerpc/sysdev/fsl_rmu.c 	if (!node) {
node             1089 arch/powerpc/sysdev/fsl_rmu.c 	aw = of_n_addr_cells(node);
node             1090 arch/powerpc/sysdev/fsl_rmu.c 	msg_addr = of_get_property(node, "reg", &mlen);
node             1093 arch/powerpc/sysdev/fsl_rmu.c 			node);
node             1102 arch/powerpc/sysdev/fsl_rmu.c 	rmu->txirq = irq_of_parse_and_map(node, 0);
node             1103 arch/powerpc/sysdev/fsl_rmu.c 	rmu->rxirq = irq_of_parse_and_map(node, 1);
node             1105 arch/powerpc/sysdev/fsl_rmu.c 		node, rmu->txirq, rmu->rxirq);
node              102 arch/powerpc/sysdev/fsl_soc.c 	struct device_node *node;
node              107 arch/powerpc/sysdev/fsl_soc.c 	node = of_find_compatible_node(NULL, NULL, "fsl,cpm-brg");
node              108 arch/powerpc/sysdev/fsl_soc.c 	if (node) {
node              109 arch/powerpc/sysdev/fsl_soc.c 		of_property_read_u32(node, "clock-frequency", &brgfreq);
node              110 arch/powerpc/sysdev/fsl_soc.c 		of_node_put(node);
node              115 arch/powerpc/sysdev/fsl_soc.c 	node = of_find_node_by_type(NULL, "cpm");
node              116 arch/powerpc/sysdev/fsl_soc.c 	if (!node)
node              117 arch/powerpc/sysdev/fsl_soc.c 		node = of_find_compatible_node(NULL, NULL, "fsl,qe");
node              118 arch/powerpc/sysdev/fsl_soc.c 	if (!node)
node              119 arch/powerpc/sysdev/fsl_soc.c 		node = of_find_node_by_type(NULL, "qe");
node              121 arch/powerpc/sysdev/fsl_soc.c 	if (node) {
node              122 arch/powerpc/sysdev/fsl_soc.c 		of_property_read_u32(node, "brg-frequency", &brgfreq);
node              124 arch/powerpc/sysdev/fsl_soc.c 			if (!of_property_read_u32(node, "bus-frequency",
node              127 arch/powerpc/sysdev/fsl_soc.c 		of_node_put(node);
node              138 arch/powerpc/sysdev/fsl_soc.c 	struct device_node *node;
node              143 arch/powerpc/sysdev/fsl_soc.c 	node = of_find_node_by_type(NULL, "serial");
node              144 arch/powerpc/sysdev/fsl_soc.c 	if (node) {
node              145 arch/powerpc/sysdev/fsl_soc.c 		of_property_read_u32(node, "current-speed", &fs_baudrate);
node              146 arch/powerpc/sysdev/fsl_soc.c 		of_node_put(node);
node              161 arch/powerpc/sysdev/i8259.c static int i8259_host_match(struct irq_domain *h, struct device_node *node,
node              165 arch/powerpc/sysdev/i8259.c 	return of_node == NULL || of_node == node;
node              223 arch/powerpc/sysdev/i8259.c void i8259_init(struct device_node *node, unsigned long intack_addr)
node              263 arch/powerpc/sysdev/i8259.c 	i8259_host = irq_domain_add_legacy_isa(node, &i8259_host_ops, NULL);
node              671 arch/powerpc/sysdev/ipic.c static int ipic_host_match(struct irq_domain *h, struct device_node *node,
node              676 arch/powerpc/sysdev/ipic.c 	return of_node == NULL || of_node == node;
node              699 arch/powerpc/sysdev/ipic.c struct ipic * __init ipic_init(struct device_node *node, unsigned int flags)
node              705 arch/powerpc/sysdev/ipic.c 	ret = of_address_to_resource(node, 0, &res);
node              713 arch/powerpc/sysdev/ipic.c 	ipic->irqhost = irq_domain_add_linear(node, NR_IPIC_INTS,
node               15 arch/powerpc/sysdev/mpc5xxx_clocks.c unsigned long mpc5xxx_get_bus_frequency(struct device_node *node)
node               19 arch/powerpc/sysdev/mpc5xxx_clocks.c 	of_node_get(node);
node               20 arch/powerpc/sysdev/mpc5xxx_clocks.c 	while (node) {
node               21 arch/powerpc/sysdev/mpc5xxx_clocks.c 		p_bus_freq = of_get_property(node, "bus-frequency", NULL);
node               25 arch/powerpc/sysdev/mpc5xxx_clocks.c 		node = of_get_next_parent(node);
node               27 arch/powerpc/sysdev/mpc5xxx_clocks.c 	of_node_put(node);
node              326 arch/powerpc/sysdev/mpic.c 	phys_addr_t phys_addr = dcr_resource_start(mpic->node, 0);
node              327 arch/powerpc/sysdev/mpic.c 	rb->dhost = dcr_map(mpic->node, phys_addr + offset, size);
node              992 arch/powerpc/sysdev/mpic.c static int mpic_host_match(struct irq_domain *h, struct device_node *node,
node              997 arch/powerpc/sysdev/mpic.c 	return of_node == NULL || of_node == node;
node             1213 arch/powerpc/sysdev/mpic.c struct mpic * __init mpic_alloc(struct device_node *node,
node             1239 arch/powerpc/sysdev/mpic.c 	if (node) {
node             1240 arch/powerpc/sysdev/mpic.c 		node = of_node_get(node);
node             1242 arch/powerpc/sysdev/mpic.c 		node = of_find_matching_node(NULL, mpic_device_id);
node             1243 arch/powerpc/sysdev/mpic.c 		if (!node)
node             1250 arch/powerpc/sysdev/mpic.c 		if (of_property_read_bool(node, "dcr-reg")) {
node             1254 arch/powerpc/sysdev/mpic.c 			if (of_address_to_resource(node, 0, &r))
node             1261 arch/powerpc/sysdev/mpic.c 	if (of_get_property(node, "big-endian", NULL))
node             1263 arch/powerpc/sysdev/mpic.c 	if (of_get_property(node, "pic-no-reset", NULL))
node             1265 arch/powerpc/sysdev/mpic.c 	if (of_get_property(node, "single-cpu-affinity", NULL))
node             1267 arch/powerpc/sysdev/mpic.c 	if (of_device_is_compatible(node, "fsl,mpic")) {
node             1278 arch/powerpc/sysdev/mpic.c 	mpic->node = node;
node             1323 arch/powerpc/sysdev/mpic.c 	psrc = of_get_property(mpic->node, "protected-sources", &psize);
node             1470 arch/powerpc/sysdev/mpic.c 	of_property_read_u32(mpic->node, "last-interrupt-source", &last_irq);
node             1487 arch/powerpc/sysdev/mpic.c 	mpic->irqhost = irq_domain_add_linear(mpic->node,
node             1530 arch/powerpc/sysdev/mpic.c 	of_node_put(node);
node             1650 arch/powerpc/sysdev/mpic.c 		int virq = irq_of_parse_and_map(mpic->node, 0);
node             1653 arch/powerpc/sysdev/mpic.c 					mpic->node, virq);
node              133 arch/powerpc/sysdev/mpic_msgr.c static int mpic_msgr_block_number(struct device_node *node)
node              149 arch/powerpc/sysdev/mpic_msgr.c 		if (node == of_find_node_by_path(prop->value))
node               66 arch/powerpc/sysdev/mpic_timer.c 	struct list_head		node;
node              202 arch/powerpc/sysdev/mpic_timer.c 	list_for_each_entry(priv, &timer_group_list, node) {
node              509 arch/powerpc/sysdev/mpic_timer.c 	list_add_tail(&priv->node, &timer_group_list);
node              527 arch/powerpc/sysdev/mpic_timer.c 	list_for_each_entry(priv, &timer_group_list, node) {
node               23 arch/powerpc/sysdev/of_rtc.c 	struct device_node *node;
node               30 arch/powerpc/sysdev/of_rtc.c 		for_each_compatible_node(node, NULL,
node               38 arch/powerpc/sysdev/of_rtc.c 				       node);
node               42 arch/powerpc/sysdev/of_rtc.c 			err = of_address_to_resource(node, 0, res);
node               46 arch/powerpc/sysdev/of_rtc.c 				       node);
node               51 arch/powerpc/sysdev/of_rtc.c 			       node, plat_name,
node              104 arch/powerpc/sysdev/pmi.c 	list_for_each_entry(handler, &data->handler, node) {
node              184 arch/powerpc/sysdev/pmi.c 	list_for_each_entry_safe(handler, tmp, &data->handler, node)
node              185 arch/powerpc/sysdev/pmi.c 		list_del(&handler->node);
node              246 arch/powerpc/sysdev/pmi.c 	list_add_tail(&handler->node, &data->handler);
node              261 arch/powerpc/sysdev/pmi.c 	list_del(&handler->node);
node              403 arch/powerpc/sysdev/tsi108_pci.c void __init tsi108_pci_int_init(struct device_node *node)
node              407 arch/powerpc/sysdev/tsi108_pci.c 	pci_irq_host = irq_domain_add_legacy_isa(node, &pci_irq_domain_ops, NULL);
node              164 arch/powerpc/sysdev/xics/ics-opal.c static int ics_opal_host_match(struct ics *ics, struct device_node *node)
node               30 arch/powerpc/sysdev/xics/ics-rtas.c static int ics_rtas_host_match(struct ics *ics, struct device_node *node);
node              207 arch/powerpc/sysdev/xics/ics-rtas.c static int ics_rtas_host_match(struct ics *ics, struct device_node *node)
node              213 arch/powerpc/sysdev/xics/ics-rtas.c 	return !of_device_is_compatible(node, "chrp,iic");
node              307 arch/powerpc/sysdev/xics/xics-common.c static int xics_host_match(struct irq_domain *h, struct device_node *node,
node              313 arch/powerpc/sysdev/xics/xics-common.c 		if (ics->host_match(ics, node))
node             1293 arch/powerpc/sysdev/xive/common.c static int xive_irq_domain_match(struct irq_domain *h, struct device_node *node,
node             1296 arch/powerpc/sysdev/xive/common.c 	return xive_ops->match(node);
node              243 arch/powerpc/sysdev/xive/native.c static bool xive_native_match(struct device_node *node)
node              245 arch/powerpc/sysdev/xive/native.c 	return of_device_is_compatible(node, "ibm,opal-xive-vc");
node              541 arch/powerpc/sysdev/xive/spapr.c static bool xive_spapr_match(struct device_node *node)
node             4072 arch/powerpc/xmon/xmon.c 	DUMP_FIELD(spu, "0x%x", node);
node               77 arch/riscv/include/asm/processor.h int riscv_of_processor_hartid(struct device_node *node);
node               12 arch/riscv/kernel/cacheinfo.c 			 struct device_node *node,
node               15 arch/riscv/kernel/cpu.c int riscv_of_processor_hartid(struct device_node *node)
node               20 arch/riscv/kernel/cpu.c 	if (!of_device_is_compatible(node, "riscv")) {
node               25 arch/riscv/kernel/cpu.c 	if (of_property_read_u32(node, "reg", &hart)) {
node               30 arch/riscv/kernel/cpu.c 	if (!of_device_is_available(node)) {
node               35 arch/riscv/kernel/cpu.c 	if (of_property_read_string(node, "riscv,isa", &isa)) {
node              131 arch/riscv/kernel/cpu.c 	struct device_node *node = of_get_cpu_node(cpu_id, NULL);
node              136 arch/riscv/kernel/cpu.c 	if (!of_property_read_string(node, "riscv,isa", &isa))
node              138 arch/riscv/kernel/cpu.c 	if (!of_property_read_string(node, "mmu-type", &mmu))
node              140 arch/riscv/kernel/cpu.c 	if (!of_property_read_string(node, "compatible", &compat)
node              144 arch/riscv/kernel/cpu.c 	of_node_put(node);
node               22 arch/riscv/kernel/cpufeature.c 	struct device_node *node;
node               36 arch/riscv/kernel/cpufeature.c 	for_each_of_cpu_node(node) {
node               39 arch/riscv/kernel/cpufeature.c 		if (riscv_of_processor_hartid(node) < 0)
node               42 arch/riscv/kernel/cpufeature.c 		if (of_property_read_string(node, "riscv,isa", &isa)) {
node              469 arch/riscv/kernel/perf_event.c 	struct device_node *node = of_find_node_by_type(NULL, "pmu");
node              474 arch/riscv/kernel/perf_event.c 	if (node) {
node              475 arch/riscv/kernel/perf_event.c 		of_id = of_match_node(riscv_pmu_of_ids, node);
node              479 arch/riscv/kernel/perf_event.c 		of_node_put(node);
node              464 arch/riscv/mm/init.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              467 arch/riscv/mm/init.c 	return vmemmap_populate_basepages(start, end, node);
node               79 arch/s390/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node)
node               81 arch/s390/include/asm/topology.h 	return &node_to_cpumask_map[node];
node               62 arch/s390/lib/spinlock.c 	struct spin_wait *node;
node               65 arch/s390/lib/spinlock.c 	node = per_cpu_ptr(&spin_wait[0], cpu);
node               66 arch/s390/lib/spinlock.c 	for (ix = 0; ix < 4; ix++, node++) {
node               67 arch/s390/lib/spinlock.c 		memset(node, 0, sizeof(*node));
node               68 arch/s390/lib/spinlock.c 		node->node_id = ((cpu + 1) << _Q_TAIL_CPU_OFFSET) +
node              106 arch/s390/lib/spinlock.c static inline int arch_spin_yield_target(int lock, struct spin_wait *node)
node              110 arch/s390/lib/spinlock.c 	if (node == NULL || node->prev == NULL)
node              112 arch/s390/lib/spinlock.c 	while (node->prev)
node              113 arch/s390/lib/spinlock.c 		node = node->prev;
node              114 arch/s390/lib/spinlock.c 	return node->node_id >> _Q_TAIL_CPU_OFFSET;
node              119 arch/s390/lib/spinlock.c 	struct spin_wait *node, *next;
node              125 arch/s390/lib/spinlock.c 	node = this_cpu_ptr(&spin_wait[ix]);
node              126 arch/s390/lib/spinlock.c 	node->prev = node->next = NULL;
node              127 arch/s390/lib/spinlock.c 	node_id = node->node_id;
node              156 arch/s390/lib/spinlock.c 		node->prev = arch_spin_decode_tail(tail_id);
node              157 arch/s390/lib/spinlock.c 		WRITE_ONCE(node->prev->next, node);
node              161 arch/s390/lib/spinlock.c 	owner = arch_spin_yield_target(old, node);
node              168 arch/s390/lib/spinlock.c 		while (READ_ONCE(node->prev) != NULL) {
node              173 arch/s390/lib/spinlock.c 			owner = arch_spin_yield_target(old, node);
node              202 arch/s390/lib/spinlock.c 		while ((next = READ_ONCE(node->next)) == NULL)
node              213 arch/s390/mm/vmem.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node              267 arch/s390/mm/vmem.c 				new_page = vmemmap_alloc_block(PMD_SIZE, node);
node              287 arch/s390/mm/vmem.c 			new_page = vmemmap_alloc_block(PAGE_SIZE, node);
node               86 arch/s390/numa/mode_emu.c static int cores_pinned(struct toptree *node)
node               88 arch/s390/numa/mode_emu.c 	return emu_cores->per_node[node->id];
node              164 arch/s390/numa/mode_emu.c static int dist_node_to_core(struct toptree *node, struct toptree *core)
node              169 arch/s390/numa/mode_emu.c 	toptree_for_each(core_node, node, CORE)
node              193 arch/s390/numa/mode_emu.c 	struct toptree *node, *node_best = NULL;
node              199 arch/s390/numa/mode_emu.c 	toptree_for_each(node, numa, NODE) {
node              201 arch/s390/numa/mode_emu.c 		if (core_pinned_to_node_id(core) == node->id) {
node              202 arch/s390/numa/mode_emu.c 			node_best = node;
node              206 arch/s390/numa/mode_emu.c 		if (cores_pinned(node) >= cores_target)
node              208 arch/s390/numa/mode_emu.c 		dist_cur = dist_node_to_core(node, core);
node              211 arch/s390/numa/mode_emu.c 			node_best = node;
node              223 arch/s390/numa/mode_emu.c 	struct toptree *node, *core, *tmp;
node              226 arch/s390/numa/mode_emu.c 		node = node_for_core(numa, core, extra);
node              227 arch/s390/numa/mode_emu.c 		if (!node)
node              229 arch/s390/numa/mode_emu.c 		toptree_move(core, node);
node              230 arch/s390/numa/mode_emu.c 		pin_core_to_node(core->id, node->id);
node              237 arch/s390/numa/mode_emu.c static void move_level_to_numa_node(struct toptree *node, struct toptree *phys,
node              244 arch/s390/numa/mode_emu.c 		cores_free = cores_target - toptree_count(node, CORE);
node              247 arch/s390/numa/mode_emu.c 				toptree_move(cur, node);
node              250 arch/s390/numa/mode_emu.c 				toptree_move(cur, node);
node              263 arch/s390/numa/mode_emu.c 	struct toptree *node;
node              265 arch/s390/numa/mode_emu.c 	toptree_for_each(node, numa, NODE)
node              266 arch/s390/numa/mode_emu.c 		move_level_to_numa_node(node, phys, level, perfect);
node              353 arch/s390/numa/mode_emu.c 	struct toptree *phys, *node, *drawer, *book, *mc, *core;
node              361 arch/s390/numa/mode_emu.c 		node = toptree_get_child(phys, 0);
node              362 arch/s390/numa/mode_emu.c 		drawer = toptree_get_child(node, top->drawer_id);
node              248 arch/sh/include/asm/dwarf.h 	struct rb_node node;
node              266 arch/sh/include/asm/dwarf.h 	struct rb_node node;
node                9 arch/sh/include/asm/topology.h #define cpumask_of_node(node)	((void)node, cpu_online_mask)
node               18 arch/sh/kernel/cpu/sh2/probe.c static int __init scan_cache(unsigned long node, const char *uname,
node               21 arch/sh/kernel/cpu/sh2/probe.c 	if (!of_flat_dt_is_compatible(node, "jcore,cache"))
node               24 arch/sh/kernel/cpu/sh2/probe.c 	j2_ccr_base = (u32 __iomem *)of_flat_dt_translate_address(node);
node              323 arch/sh/kernel/dwarf.c 		cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node);
node              359 arch/sh/kernel/dwarf.c 		fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node);
node              840 arch/sh/kernel/dwarf.c 		cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node);
node              852 arch/sh/kernel/dwarf.c 	rb_link_node(&cie->node, parent, rb_node);
node              853 arch/sh/kernel/dwarf.c 	rb_insert_color(&cie->node, &cie_root);
node              926 arch/sh/kernel/dwarf.c 		fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node);
node              944 arch/sh/kernel/dwarf.c 	rb_link_node(&fde->node, parent, rb_node);
node              945 arch/sh/kernel/dwarf.c 	rb_insert_color(&fde->node, &fde_root);
node             1004 arch/sh/kernel/dwarf.c 	rbtree_postorder_for_each_entry_safe(fde, next_fde, &fde_root, node)
node             1007 arch/sh/kernel/dwarf.c 	rbtree_postorder_for_each_entry_safe(cie, next_cie, &cie_root, node)
node             1140 arch/sh/kernel/dwarf.c 		rb_erase(&cie->node, &cie_root);
node             1150 arch/sh/kernel/dwarf.c 		rb_erase(&fde->node, &fde_root);
node               21 arch/sparc/include/asm/fb.h 	struct device_node *node;
node               26 arch/sparc/include/asm/fb.h 	node = dev->of_node;
node               27 arch/sparc/include/asm/fb.h 	if (node &&
node               28 arch/sparc/include/asm/fb.h 	    node == of_console_device)
node               40 arch/sparc/include/asm/mdesc.h 			       u64 node, const char *name, int *lenp);
node               41 arch/sparc/include/asm/mdesc.h const char *mdesc_node_name(struct mdesc_handle *hp, u64 node);
node               67 arch/sparc/include/asm/mdesc.h 	void (*add)(struct mdesc_handle *handle, u64 node,
node               69 arch/sparc/include/asm/mdesc.h 	void (*remove)(struct mdesc_handle *handle, u64 node,
node               90 arch/sparc/include/asm/mdesc.h int mdesc_get_node_info(struct mdesc_handle *hp, u64 node,
node              174 arch/sparc/include/asm/openprom.h 	phandle (*no_nextnode)(phandle node);
node              175 arch/sparc/include/asm/openprom.h 	phandle (*no_child)(phandle node);
node              176 arch/sparc/include/asm/openprom.h 	int (*no_proplen)(phandle node, const char *name);
node              177 arch/sparc/include/asm/openprom.h 	int (*no_getprop)(phandle node, const char *name, char *val);
node              178 arch/sparc/include/asm/openprom.h 	int (*no_setprop)(phandle node, const char *name, char *val, int len);
node              179 arch/sparc/include/asm/openprom.h 	char * (*no_nextprop)(phandle node, char *name);
node              120 arch/sparc/include/asm/oplib_32.h phandle prom_getsibling(phandle node);
node              134 arch/sparc/include/asm/oplib_32.h int prom_getint(phandle node, char *property);
node              137 arch/sparc/include/asm/oplib_32.h int prom_getintdefault(phandle node, char *property, int defval);
node              140 arch/sparc/include/asm/oplib_32.h int prom_getbool(phandle node, char *prop);
node              143 arch/sparc/include/asm/oplib_32.h void prom_getstring(phandle node, char *prop, char *buf, int bufsize);
node              153 arch/sparc/include/asm/oplib_32.h char *prom_nextprop(phandle node, char *prev_property, char *buffer);
node              161 arch/sparc/include/asm/oplib_32.h int prom_setprop(phandle node, const char *prop_name, char *prop_value,
node              172 arch/sparc/include/asm/oplib_32.h void prom_apply_generic_ranges(phandle node, phandle parent,
node              188 arch/sparc/include/asm/oplib_64.h phandle prom_getsibling(phandle node);
node              202 arch/sparc/include/asm/oplib_64.h int prom_getint(phandle node, const char *property);
node              205 arch/sparc/include/asm/oplib_64.h int prom_getintdefault(phandle node, const char *property, int defval);
node              208 arch/sparc/include/asm/oplib_64.h int prom_getbool(phandle node, const char *prop);
node              211 arch/sparc/include/asm/oplib_64.h void prom_getstring(phandle node, const char *prop, char *buf,
node              225 arch/sparc/include/asm/oplib_64.h char *prom_firstprop(phandle node, char *buffer);
node              230 arch/sparc/include/asm/oplib_64.h char *prom_nextprop(phandle node, const char *prev_property, char *buf);
node              233 arch/sparc/include/asm/oplib_64.h int prom_node_has_property(phandle node, const char *property);
node              241 arch/sparc/include/asm/oplib_64.h int prom_setprop(phandle node, const char *prop_name, char *prop_value,
node               33 arch/sparc/include/asm/prom.h int of_set_property(struct device_node *node, const char *name, void *val, int len);
node               14 arch/sparc/include/asm/topology_64.h #define cpumask_of_node(node) ((node) == -1 ?				\
node               16 arch/sparc/include/asm/topology_64.h 			       &numa_cpumask_lookup_table[node])
node              362 arch/sparc/include/asm/vio.h 	struct list_head		node;
node               32 arch/sparc/kernel/auxio_32.c 	phandle node, auxio_nd;
node               43 arch/sparc/kernel/auxio_32.c 	node = prom_getchild(prom_root_node);
node               44 arch/sparc/kernel/auxio_32.c 	auxio_nd = prom_searchsiblings(node, "auxiliary-io");
node               46 arch/sparc/kernel/auxio_32.c 		node = prom_searchsiblings(node, "obio");
node               47 arch/sparc/kernel/auxio_32.c 		node = prom_getchild(node);
node               48 arch/sparc/kernel/auxio_32.c 		auxio_nd = prom_searchsiblings(node, "auxio");
node               54 arch/sparc/kernel/auxio_32.c 			if(prom_searchsiblings(node, "leds")) {
node              115 arch/sparc/kernel/auxio_32.c 	phandle node;
node              119 arch/sparc/kernel/auxio_32.c 	node = prom_getchild(prom_root_node);
node              120 arch/sparc/kernel/auxio_32.c 	node = prom_searchsiblings(node, "obio");
node              121 arch/sparc/kernel/auxio_32.c 	node = prom_getchild(node);
node              122 arch/sparc/kernel/auxio_32.c 	node = prom_searchsiblings(node, "power");
node              123 arch/sparc/kernel/auxio_32.c 	if (node == 0 || (s32)node == -1)
node              127 arch/sparc/kernel/auxio_32.c 	if (prom_getproperty(node, "reg", (char *)&regs, sizeof(regs)) <= 0)
node               43 arch/sparc/kernel/btext.c static int __init btext_initialize(phandle node)
node               49 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "width", (char *)&width, 4) < 0)
node               51 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "height", (char *)&height, 4) < 0)
node               53 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "depth", (char *)&depth, 4) < 0)
node               57 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "linebytes", (char *)&prop, 4) >= 0 &&
node               64 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "address", (char *)&prop, 4) >= 0)
node              312 arch/sparc/kernel/btext.c 	phandle node;
node              316 arch/sparc/kernel/btext.c 	node = prom_inst2pkg(prom_stdout);
node              317 arch/sparc/kernel/btext.c 	if (prom_getproperty(node, "device_type", type, 32) < 0)
node              322 arch/sparc/kernel/btext.c 	ret = btext_initialize(node);
node              188 arch/sparc/kernel/cpumap.c 	struct cpuinfo_node *node;
node              211 arch/sparc/kernel/cpumap.c 		node = &new_tree->nodes[n];
node              218 arch/sparc/kernel/cpumap.c 		node->id = id;
node              219 arch/sparc/kernel/cpumap.c 		node->level = level;
node              220 arch/sparc/kernel/cpumap.c 		node->num_cpus = 1;
node              222 arch/sparc/kernel/cpumap.c 		node->parent_index = (level > CPUINFO_LVL_ROOT)
node              225 arch/sparc/kernel/cpumap.c 		node->child_start = node->child_end = node->rover =
node              229 arch/sparc/kernel/cpumap.c 		prev_id[level] = node->id;
node              252 arch/sparc/kernel/cpumap.c 				node = &new_tree->nodes[level_rover[level]];
node              253 arch/sparc/kernel/cpumap.c 				node->num_cpus = num_cpus[level];
node              257 arch/sparc/kernel/cpumap.c 					node->num_cpus++;
node              261 arch/sparc/kernel/cpumap.c 					node->parent_index = -1;
node              263 arch/sparc/kernel/cpumap.c 					node->parent_index =
node              267 arch/sparc/kernel/cpumap.c 					node->child_end =
node              270 arch/sparc/kernel/cpumap.c 					node->child_end =
node              277 arch/sparc/kernel/cpumap.c 					node = &new_tree->nodes[n];
node              278 arch/sparc/kernel/cpumap.c 					node->id = id;
node              279 arch/sparc/kernel/cpumap.c 					node->level = level;
node              282 arch/sparc/kernel/cpumap.c 					node->child_start = node->child_end =
node              283 arch/sparc/kernel/cpumap.c 					node->rover =
node              299 arch/sparc/kernel/cpumap.c 	struct cpuinfo_node *node = &t->nodes[node_index];
node              303 arch/sparc/kernel/cpumap.c 	for (level = node->level; level >= top_level; level--) {
node              304 arch/sparc/kernel/cpumap.c 		node->rover++;
node              305 arch/sparc/kernel/cpumap.c 		if (node->rover <= node->child_end)
node              308 arch/sparc/kernel/cpumap.c 		node->rover = node->child_start;
node              314 arch/sparc/kernel/cpumap.c 		node = &t->nodes[node->parent_index];
node               92 arch/sparc/kernel/mdesc.c static int get_vdev_port_node_info(struct mdesc_handle *md, u64 node,
node               98 arch/sparc/kernel/mdesc.c static int get_ds_port_node_info(struct mdesc_handle *md, u64 node,
node              290 arch/sparc/kernel/mdesc.c 	u64 node;
node              312 arch/sparc/kernel/mdesc.c 	mdesc_for_each_node_by_name(cur_mdesc, node, client->node_name)
node              313 arch/sparc/kernel/mdesc.c 		client->add(cur_mdesc, node, client->node_name);
node              318 arch/sparc/kernel/mdesc.c static const u64 *parent_cfg_handle(struct mdesc_handle *hp, u64 node)
node              324 arch/sparc/kernel/mdesc.c 	mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) {
node              337 arch/sparc/kernel/mdesc.c static int get_vdev_port_node_info(struct mdesc_handle *md, u64 node,
node              350 arch/sparc/kernel/mdesc.c 	idp = mdesc_get_property(md, node, "id", NULL);
node              351 arch/sparc/kernel/mdesc.c 	name = mdesc_get_property(md, node, "name", NULL);
node              352 arch/sparc/kernel/mdesc.c 	parent_cfg_hdlp = parent_cfg_handle(md, node);
node              391 arch/sparc/kernel/mdesc.c static int get_ds_port_node_info(struct mdesc_handle *md, u64 node,
node              397 arch/sparc/kernel/mdesc.c 	idp = mdesc_get_property(md, node, "id", NULL);
node              583 arch/sparc/kernel/mdesc.c int mdesc_get_node_info(struct mdesc_handle *hp, u64 node,
node              589 arch/sparc/kernel/mdesc.c 	if (hp == NULL || node == MDESC_NODE_NULL ||
node              602 arch/sparc/kernel/mdesc.c 	rv = get_info_func(hp, node, node_info);
node              657 arch/sparc/kernel/mdesc.c const void *mdesc_get_property(struct mdesc_handle *hp, u64 node,
node              665 arch/sparc/kernel/mdesc.c 	if (node == MDESC_NODE_NULL || node >= last_node)
node              668 arch/sparc/kernel/mdesc.c 	ep = node_block(&hp->mdesc) + node;
node              739 arch/sparc/kernel/mdesc.c const char *mdesc_node_name(struct mdesc_handle *hp, u64 node)
node              745 arch/sparc/kernel/mdesc.c 	if (node == MDESC_NODE_NULL || node >= last_node)
node              748 arch/sparc/kernel/mdesc.c 	ep = base + node;
node              860 arch/sparc/kernel/mdesc.c static void find_back_node_value(struct mdesc_handle *hp, u64 node,
node              871 arch/sparc/kernel/mdesc.c 	mdesc_for_each_arc(arc, hp, node, MDESC_ARC_TYPE_BACK) {
node              882 arch/sparc/kernel/mdesc.c static void __mark_core_id(struct mdesc_handle *hp, u64 node,
node              885 arch/sparc/kernel/mdesc.c 	const u64 *id = mdesc_get_property(hp, node, "id", NULL);
node              891 arch/sparc/kernel/mdesc.c static void __mark_max_cache_id(struct mdesc_handle *hp, u64 node,
node              894 arch/sparc/kernel/mdesc.c 	const u64 *id = mdesc_get_property(hp, node, "id", NULL);
node               16 arch/sparc/kernel/of_device_common.c unsigned int irq_of_parse_and_map(struct device_node *node, int index)
node               18 arch/sparc/kernel/of_device_common.c 	struct platform_device *op = of_find_device_by_node(node);
node               27 arch/sparc/kernel/of_device_common.c int of_address_to_resource(struct device_node *node, int index,
node               30 arch/sparc/kernel/of_device_common.c 	struct platform_device *op = of_find_device_by_node(node);
node               40 arch/sparc/kernel/of_device_common.c void __iomem *of_iomap(struct device_node *node, int index)
node               42 arch/sparc/kernel/of_device_common.c 	struct platform_device *op = of_find_device_by_node(node);
node              206 arch/sparc/kernel/pci.c 			       struct device_node *node,
node              213 arch/sparc/kernel/pci.c 	addrs = of_get_property(node, "assigned-addresses", &proplen);
node              264 arch/sparc/kernel/pci.c 					 struct device_node *node,
node              276 arch/sparc/kernel/pci.c 	op = of_find_device_by_node(node);
node              285 arch/sparc/kernel/pci.c 	if (of_node_name_eq(node, "ebus"))
node              290 arch/sparc/kernel/pci.c 			 devfn, of_node_get_device_type(node));
node              292 arch/sparc/kernel/pci.c 	dev->sysdata = node;
node              295 arch/sparc/kernel/pci.c 	dev->dev.of_node = of_node_get(node);
node              301 arch/sparc/kernel/pci.c 	dev->vendor = of_getintprop_default(node, "vendor-id", 0xffff);
node              302 arch/sparc/kernel/pci.c 	dev->device = of_getintprop_default(node, "device-id", 0xffff);
node              304 arch/sparc/kernel/pci.c 		of_getintprop_default(node, "subsystem-vendor-id", 0);
node              306 arch/sparc/kernel/pci.c 		of_getintprop_default(node, "subsystem-id", 0);
node              334 arch/sparc/kernel/pci.c 	if (of_node_name_eq(node, "pci")) {
node              338 arch/sparc/kernel/pci.c 	} else if (of_node_is_type(node, "cardbus")) {
node              352 arch/sparc/kernel/pci.c 	pci_parse_of_addrs(sd->op, node, dev);
node              411 arch/sparc/kernel/pci.c 			    struct device_node *node,
node              417 arch/sparc/kernel/pci.c 			       struct device_node *node,
node              429 arch/sparc/kernel/pci.c 		pci_info(dev, "of_scan_pci_bridge(%pOF)\n", node);
node              432 arch/sparc/kernel/pci.c 	busrange = of_get_property(node, "bus-range", &len);
node              435 arch/sparc/kernel/pci.c 		       node);
node              443 arch/sparc/kernel/pci.c 	ranges = of_get_property(node, "ranges", &len);
node              446 arch/sparc/kernel/pci.c 		const char *model = of_get_property(node, "model", NULL);
node              454 arch/sparc/kernel/pci.c 			node);
node              510 arch/sparc/kernel/pci.c 					" for bridge %pOF\n", node);
node              516 arch/sparc/kernel/pci.c 					" for bridge %pOF\n", node);
node              539 arch/sparc/kernel/pci.c 	pci_of_scan_bus(pbm, node, bus);
node              543 arch/sparc/kernel/pci.c 			    struct device_node *node,
node              553 arch/sparc/kernel/pci.c 			 node, bus->number);
node              557 arch/sparc/kernel/pci.c 	while ((child = of_get_next_child(node, child)) != NULL) {
node              618 arch/sparc/kernel/pci.c 	list_for_each_entry(child_bus, &bus->children, node)
node              685 arch/sparc/kernel/pci.c 	list_for_each_entry(child_bus, &bus->children, node)
node              693 arch/sparc/kernel/pci.c 	struct device_node *node = pbm->op->dev.of_node;
node              696 arch/sparc/kernel/pci.c 	printk("PCI: Scanning PBM %pOF\n", node);
node              712 arch/sparc/kernel/pci.c 		printk(KERN_ERR "Failed to create bus for %pOF\n", node);
node              717 arch/sparc/kernel/pci.c 	pci_of_scan_bus(pbm, node, bus);
node             1070 arch/sparc/kernel/pci.c 	list_for_each_entry(bus, &pbus->children, node)
node             1074 arch/sparc/kernel/pci.c static void pci_bus_slot_names(struct device_node *node, struct pci_bus *bus)
node             1084 arch/sparc/kernel/pci.c 	prop = of_get_property(node, "slot-names", &len);
node             1093 arch/sparc/kernel/pci.c 			 node, mask);
node             1124 arch/sparc/kernel/pci.c 		struct device_node *node;
node             1137 arch/sparc/kernel/pci.c 				node = pbus->self->dev.of_node;
node             1143 arch/sparc/kernel/pci.c 				node = pbm->op->dev.of_node;
node             1146 arch/sparc/kernel/pci.c 			pci_bus_slot_names(node, pbus);
node              497 arch/sparc/kernel/pci_common.c 	list_for_each_entry(bus, &pbus->children, node)
node              520 arch/sparc/kernel/pci_common.c 	list_for_each_entry(bus, &pbus->children, node)
node              544 arch/sparc/kernel/pci_common.c 	list_for_each_entry(bus, &pbus->children, node)
node              293 arch/sparc/kernel/pcic.c 	phandle node;
node              302 arch/sparc/kernel/pcic.c 	node = prom_getchild (prom_root_node);
node              303 arch/sparc/kernel/pcic.c 	node = prom_searchsiblings (node, "pci");
node              304 arch/sparc/kernel/pcic.c 	if (node == 0)
node              309 arch/sparc/kernel/pcic.c 	err = prom_getproperty(node, "reg", (char*)regs, sizeof(regs));
node              353 arch/sparc/kernel/pcic.c 	pbm->prom_node = node;
node              354 arch/sparc/kernel/pcic.c 	prom_getstring(node, "name", namebuf, 63);  namebuf[63] = 0;
node              451 arch/sparc/kernel/pcic.c 	phandle node = prom_getchild(pbm->prom_node);
node              453 arch/sparc/kernel/pcic.c 	while(node) {
node              454 arch/sparc/kernel/pcic.c 		err = prom_getproperty(node, "reg", 
node              459 arch/sparc/kernel/pcic.c 				return node;
node              461 arch/sparc/kernel/pcic.c 		node = prom_getsibling(node);
node              472 arch/sparc/kernel/pcic.c     struct pci_dev *dev, int node)
node              479 arch/sparc/kernel/pcic.c 	if (node == 0 || node == -1) {
node              482 arch/sparc/kernel/pcic.c 		prom_getstring(node, "name", namebuf, 63); namebuf[63] = 0;
node              531 arch/sparc/kernel/pcic.c pcic_fill_irq(struct linux_pcic *pcic, struct pci_dev *dev, int node)
node              538 arch/sparc/kernel/pcic.c 	if (node == 0 || node == -1) {
node              541 arch/sparc/kernel/pcic.c 		prom_getstring(node, "name", namebuf, sizeof(namebuf));
node              607 arch/sparc/kernel/pcic.c 	int node;
node              626 arch/sparc/kernel/pcic.c 		node = pdev_to_pnode(&pcic->pbm, dev);
node              627 arch/sparc/kernel/pcic.c 		if(node == 0)
node              628 arch/sparc/kernel/pcic.c 			node = -1;
node              633 arch/sparc/kernel/pcic.c 		pcp->prom_node = of_find_node_by_phandle(node);
node              638 arch/sparc/kernel/pcic.c 			pcic_map_pci_device(pcic, dev, node);
node              640 arch/sparc/kernel/pcic.c 		pcic_fill_irq(pcic, dev, node);
node              222 arch/sparc/kernel/prom_32.c 	phandle node;
node              273 arch/sparc/kernel/prom_32.c 		node = (*romvec->pv_v2devops.v2_inst2pkg)(fd);
node              277 arch/sparc/kernel/prom_32.c 		if (!node) {
node              282 arch/sparc/kernel/prom_32.c 		dp = of_find_node_by_phandle(node);
node              601 arch/sparc/kernel/prom_64.c 	phandle node;
node              615 arch/sparc/kernel/prom_64.c 	node = prom_inst2pkg(prom_stdout);
node              616 arch/sparc/kernel/prom_64.c 	if (!node) {
node              622 arch/sparc/kernel/prom_64.c 	dp = of_find_node_by_phandle(node);
node              132 arch/sparc/kernel/prom_common.c static int __init prom_common_nextprop(phandle node, char *prev, char *buf)
node              137 arch/sparc/kernel/prom_common.c 	name = prom_nextprop(node, prev, buf);
node              620 arch/sparc/kernel/setup_64.c 	unsigned int i, node;
node              623 arch/sparc/kernel/setup_64.c 		node = cpu_to_node(i);
node              626 arch/sparc/kernel/setup_64.c 						       THREAD_SIZE, node);
node              629 arch/sparc/kernel/setup_64.c 			      __func__, THREAD_SIZE, THREAD_SIZE, node);
node              631 arch/sparc/kernel/setup_64.c 						       THREAD_SIZE, node);
node              634 arch/sparc/kernel/setup_64.c 			      __func__, THREAD_SIZE, THREAD_SIZE, node);
node             1587 arch/sparc/kernel/smp_64.c 	int node = cpu_to_node(cpu);
node             1590 arch/sparc/kernel/smp_64.c 	if (!node_online(node) || !NODE_DATA(node)) {
node             1593 arch/sparc/kernel/smp_64.c 			cpu, node);
node             1598 arch/sparc/kernel/smp_64.c 					     MEMBLOCK_ALLOC_ACCESSIBLE, node);
node             1600 arch/sparc/kernel/smp_64.c 			 "%016lx\n", cpu, size, node, __pa(ptr));
node              166 arch/sparc/kernel/time_64.c static unsigned long cpuid_to_freq(phandle node, int cpuid)
node              172 arch/sparc/kernel/time_64.c 	if (!node)
node              175 arch/sparc/kernel/time_64.c 	if (prom_getproperty(node, "device_type", type, sizeof(type)) != -1)
node              179 arch/sparc/kernel/time_64.c 	if (is_cpu_node && (prom_getint(node, "upa-portid") == cpuid ||
node              180 arch/sparc/kernel/time_64.c 			    prom_getint(node, "cpuid") == cpuid))
node              181 arch/sparc/kernel/time_64.c 		freq = prom_getintdefault(node, "clock-frequency", 0);
node              183 arch/sparc/kernel/time_64.c 		freq = cpuid_to_freq(prom_getchild(node), cpuid);
node              185 arch/sparc/kernel/time_64.c 		freq = cpuid_to_freq(prom_getsibling(node), cpuid);
node              207 arch/sparc/kernel/vio.c static const u64 *vio_cfg_handle(struct mdesc_handle *hp, u64 node)
node              212 arch/sparc/kernel/vio.c 	mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) {
node              240 arch/sparc/kernel/vio.c 	u64 node;
node              245 arch/sparc/kernel/vio.c 	node = mdesc_get_node(hp, (const char *)vdev->node_name,
node              248 arch/sparc/kernel/vio.c 	return node;
node              413 arch/sparc/kernel/vio.c static void vio_add(struct mdesc_handle *hp, u64 node,
node              416 arch/sparc/kernel/vio.c 	(void) vio_create_one(hp, node, node_name, &root_vdev->dev);
node              421 arch/sparc/kernel/vio.c 	u64 node;
node              428 arch/sparc/kernel/vio.c 	u64 node;
node              432 arch/sparc/kernel/vio.c 	node = vio_vdev_node(node_data->hp, vdev);
node              434 arch/sparc/kernel/vio.c 	if (node == node_data->node)
node              440 arch/sparc/kernel/vio.c static void vio_remove(struct mdesc_handle *hp, u64 node, const char *node_name)
node              446 arch/sparc/kernel/vio.c 	node_data.node = node;
node              471 arch/sparc/kernel/vio.c static void vio_add_ds(struct mdesc_handle *hp, u64 node,
node              478 arch/sparc/kernel/vio.c 	mdesc_for_each_arc(a, hp, node, MDESC_ARC_TYPE_BACK) {
node              489 arch/sparc/kernel/vio.c 		(void) vio_create_one(hp, node, node_name, &root_vdev->dev);
node              113 arch/sparc/mm/init_64.c 	phandle node = prom_finddevice("/memory");
node              114 arch/sparc/mm/init_64.c 	int prop_size = prom_getproplen(node, property);
node              125 arch/sparc/mm/init_64.c 	ret = prom_getproperty(node, property, (char *) regs, prop_size);
node              614 arch/sparc/mm/init_64.c 	int n, node, ents, first, last, i;
node              616 arch/sparc/mm/init_64.c 	node = prom_finddevice("/virtual-memory");
node              617 arch/sparc/mm/init_64.c 	n = prom_getproplen(node, "translations");
node              627 arch/sparc/mm/init_64.c 	if ((n = prom_getproperty(node, "translations",
node              938 arch/sparc/mm/init_64.c 	u64	node;
node             1259 arch/sparc/mm/init_64.c 	u64 node;
node             1261 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "memory-latency-group")
node             1275 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "memory-latency-group") {
node             1279 arch/sparc/mm/init_64.c 		m->node = node;
node             1281 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node, "latency", NULL);
node             1283 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node, "address-match", NULL);
node             1285 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node, "address-mask", NULL);
node             1290 arch/sparc/mm/init_64.c 			count - 1, m->node, m->latency, m->match, m->mask);
node             1300 arch/sparc/mm/init_64.c 	u64 node;
node             1302 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "mblock")
node             1316 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "mblock") {
node             1320 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node, "base", NULL);
node             1322 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node, "size", NULL);
node             1324 arch/sparc/mm/init_64.c 		val = mdesc_get_property(md, node,
node             1362 arch/sparc/mm/init_64.c static struct mdesc_mlgroup * __init find_mlgroup(u64 node)
node             1368 arch/sparc/mm/init_64.c 		if (m->node == node)
node             1481 arch/sparc/mm/init_64.c 	u64 node;
node             1483 arch/sparc/mm/init_64.c 	node = mdesc_node_by_name(md, MDESC_NODE_NULL, "latency-groups");
node             1484 arch/sparc/mm/init_64.c 	if (node == MDESC_NODE_NULL) {
node             1498 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "group") {
node             1499 arch/sparc/mm/init_64.c 		err = numa_parse_mdesc_group(md, node, count);
node             1506 arch/sparc/mm/init_64.c 	mdesc_for_each_node_by_name(md, node, "group") {
node             1507 arch/sparc/mm/init_64.c 		find_numa_latencies_for_group(md, node, count);
node             2597 arch/sparc/mm/init_64.c 			       int node, struct vmem_altmap *altmap)
node             2613 arch/sparc/mm/init_64.c 		pgd_t *pgd = vmemmap_pgd_populate(vstart, node);
node             2621 arch/sparc/mm/init_64.c 		pud = vmemmap_pud_populate(pgd, vstart, node);
node             2628 arch/sparc/mm/init_64.c 			void *block = vmemmap_alloc_block(PMD_SIZE, node);
node               34 arch/sparc/prom/init_64.c 	phandle node;
node               44 arch/sparc/prom/init_64.c 	node = prom_finddevice("/openprom");
node               45 arch/sparc/prom/init_64.c 	if (!node || (s32)node == -1)
node               48 arch/sparc/prom/init_64.c 	prom_getstring(node, "version", prom_version, sizeof(prom_version));
node               35 arch/sparc/prom/memory.c 	phandle node;
node               38 arch/sparc/prom/memory.c 	node = prom_searchsiblings(prom_getchild(prom_root_node), "memory");
node               39 arch/sparc/prom/memory.c 	size = prom_getproperty(node, "available", (char *) reg, sizeof(reg));
node              167 arch/sparc/prom/misc_64.c 	phandle node;
node              173 arch/sparc/prom/misc_64.c 	node = prom_finddevice(prom_chosen_path);
node              174 arch/sparc/prom/misc_64.c 	ret = prom_getint(node, prom_mmu_name);
node              186 arch/sparc/prom/misc_64.c 	phandle node;
node              192 arch/sparc/prom/misc_64.c 	node = prom_finddevice("/chosen");
node              193 arch/sparc/prom/misc_64.c 	ret = prom_getint(node, "memory");
node               67 arch/sparc/prom/ranges.c 	phandle node, obio_node;
node               73 arch/sparc/prom/ranges.c 	node = prom_getchild(prom_root_node);
node               74 arch/sparc/prom/ranges.c 	obio_node = prom_searchsiblings(node, "obio");
node               88 arch/sparc/prom/ranges.c void prom_apply_generic_ranges(phandle node, phandle parent,
node               95 arch/sparc/prom/ranges.c 	success = prom_getproperty(node, "ranges",
node               24 arch/sparc/prom/tree_32.c static phandle __prom_getchild(phandle node)
node               30 arch/sparc/prom/tree_32.c 	cnode = prom_nodeops->no_child(node);
node               40 arch/sparc/prom/tree_32.c phandle prom_getchild(phandle node)
node               44 arch/sparc/prom/tree_32.c 	if ((s32)node == -1)
node               47 arch/sparc/prom/tree_32.c 	cnode = __prom_getchild(node);
node               56 arch/sparc/prom/tree_32.c static phandle __prom_getsibling(phandle node)
node               62 arch/sparc/prom/tree_32.c 	cnode = prom_nodeops->no_nextnode(node);
node               72 arch/sparc/prom/tree_32.c phandle prom_getsibling(phandle node)
node               76 arch/sparc/prom/tree_32.c 	if ((s32)node == -1)
node               79 arch/sparc/prom/tree_32.c 	sibnode = __prom_getsibling(node);
node               90 arch/sparc/prom/tree_32.c int prom_getproplen(phandle node, const char *prop)
node               95 arch/sparc/prom/tree_32.c 	if((!node) || (!prop))
node               99 arch/sparc/prom/tree_32.c 	ret = prom_nodeops->no_proplen(node, prop);
node              110 arch/sparc/prom/tree_32.c int prom_getproperty(phandle node, const char *prop, char *buffer, int bufsize)
node              115 arch/sparc/prom/tree_32.c 	plen = prom_getproplen(node, prop);
node              120 arch/sparc/prom/tree_32.c 	ret = prom_nodeops->no_getprop(node, prop, buffer);
node              130 arch/sparc/prom/tree_32.c int prom_getint(phandle node, char *prop)
node              134 arch/sparc/prom/tree_32.c 	if(prom_getproperty(node, prop, (char *) &intprop, sizeof(int)) != -1)
node              144 arch/sparc/prom/tree_32.c int prom_getintdefault(phandle node, char *property, int deflt)
node              148 arch/sparc/prom/tree_32.c 	retval = prom_getint(node, property);
node              156 arch/sparc/prom/tree_32.c int prom_getbool(phandle node, char *prop)
node              160 arch/sparc/prom/tree_32.c 	retval = prom_getproplen(node, prop);
node              170 arch/sparc/prom/tree_32.c void prom_getstring(phandle node, char *prop, char *user_buf, int ubuf_size)
node              174 arch/sparc/prom/tree_32.c 	len = prom_getproperty(node, prop, user_buf, ubuf_size);
node              204 arch/sparc/prom/tree_32.c static char *__prom_nextprop(phandle node, char * oprop)
node              210 arch/sparc/prom/tree_32.c 	prop = prom_nodeops->no_nextprop(node, oprop);
node              221 arch/sparc/prom/tree_32.c char *prom_nextprop(phandle node, char *oprop, char *buffer)
node              223 arch/sparc/prom/tree_32.c 	if (node == 0 || (s32)node == -1)
node              226 arch/sparc/prom/tree_32.c 	return __prom_nextprop(node, oprop);
node              234 arch/sparc/prom/tree_32.c 	phandle node = prom_root_node, node2;
node              239 arch/sparc/prom/tree_32.c 		if (!*s) return node; /* path '.../' is legal */
node              240 arch/sparc/prom/tree_32.c 		node = prom_getchild(node);
node              246 arch/sparc/prom/tree_32.c 		node = prom_searchsiblings(node, nbuf);
node              247 arch/sparc/prom/tree_32.c 		if (!node)
node              256 arch/sparc/prom/tree_32.c 					node2 = node;
node              260 arch/sparc/prom/tree_32.c 								node = node2;
node              274 arch/sparc/prom/tree_32.c 	return node;
node              281 arch/sparc/prom/tree_32.c int prom_setprop(phandle node, const char *pname, char *value, int size)
node              291 arch/sparc/prom/tree_32.c 	ret = prom_nodeops->no_setprop(node, pname, value, size);
node              300 arch/sparc/prom/tree_32.c 	phandle node;
node              304 arch/sparc/prom/tree_32.c 	node = (*romvec->pv_v2devops.v2_inst2pkg)(inst);
node              307 arch/sparc/prom/tree_32.c 	if ((s32)node == -1)
node              309 arch/sparc/prom/tree_32.c 	return node;
node               20 arch/sparc/prom/tree_64.c static phandle prom_node_to_node(const char *type, phandle node)
node               27 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node               38 arch/sparc/prom/tree_64.c inline phandle __prom_getchild(phandle node)
node               40 arch/sparc/prom/tree_64.c 	return prom_node_to_node("child", node);
node               43 arch/sparc/prom/tree_64.c phandle prom_getchild(phandle node)
node               47 arch/sparc/prom/tree_64.c 	if ((s32)node == -1)
node               49 arch/sparc/prom/tree_64.c 	cnode = __prom_getchild(node);
node               56 arch/sparc/prom/tree_64.c inline phandle prom_getparent(phandle node)
node               60 arch/sparc/prom/tree_64.c 	if ((s32)node == -1)
node               62 arch/sparc/prom/tree_64.c 	cnode = prom_node_to_node("parent", node);
node               71 arch/sparc/prom/tree_64.c inline phandle __prom_getsibling(phandle node)
node               73 arch/sparc/prom/tree_64.c 	return prom_node_to_node(prom_peer_name, node);
node               76 arch/sparc/prom/tree_64.c phandle prom_getsibling(phandle node)
node               80 arch/sparc/prom/tree_64.c 	if ((s32)node == -1)
node               82 arch/sparc/prom/tree_64.c 	sibnode = __prom_getsibling(node);
node               93 arch/sparc/prom/tree_64.c int prom_getproplen(phandle node, const char *prop)
node               97 arch/sparc/prom/tree_64.c 	if (!node || !prop)
node              103 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node              117 arch/sparc/prom/tree_64.c int prom_getproperty(phandle node, const char *prop,
node              123 arch/sparc/prom/tree_64.c 	plen = prom_getproplen(node, prop);
node              130 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node              145 arch/sparc/prom/tree_64.c int prom_getint(phandle node, const char *prop)
node              149 arch/sparc/prom/tree_64.c 	if (prom_getproperty(node, prop, (char *) &intprop, sizeof(int)) != -1)
node              160 arch/sparc/prom/tree_64.c int prom_getintdefault(phandle node, const char *property, int deflt)
node              164 arch/sparc/prom/tree_64.c 	retval = prom_getint(node, property);
node              173 arch/sparc/prom/tree_64.c int prom_getbool(phandle node, const char *prop)
node              177 arch/sparc/prom/tree_64.c 	retval = prom_getproplen(node, prop);
node              188 arch/sparc/prom/tree_64.c void prom_getstring(phandle node, const char *prop, char *user_buf,
node              193 arch/sparc/prom/tree_64.c 	len = prom_getproperty(node, prop, user_buf, ubuf_size);
node              203 arch/sparc/prom/tree_64.c int prom_nodematch(phandle node, const char *name)
node              206 arch/sparc/prom/tree_64.c 	prom_getproperty(node, "name", namebuf, sizeof(namebuf));
node              239 arch/sparc/prom/tree_64.c char *prom_firstprop(phandle node, char *buffer)
node              244 arch/sparc/prom/tree_64.c 	if ((s32)node == -1)
node              250 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node              265 arch/sparc/prom/tree_64.c char *prom_nextprop(phandle node, const char *oprop, char *buffer)
node              270 arch/sparc/prom/tree_64.c 	if ((s32)node == -1) {
node              282 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node              311 arch/sparc/prom/tree_64.c int prom_node_has_property(phandle node, const char *prop)
node              317 arch/sparc/prom/tree_64.c 		prom_nextprop(node, buf, buf);
node              329 arch/sparc/prom/tree_64.c prom_setprop(phandle node, const char *pname, char *value, int size)
node              347 arch/sparc/prom/tree_64.c 	args[3] = (unsigned int) node;
node              362 arch/sparc/prom/tree_64.c 	phandle node;
node              372 arch/sparc/prom/tree_64.c 	node = (int) args[4];
node              373 arch/sparc/prom/tree_64.c 	if ((s32)node == -1)
node              375 arch/sparc/prom/tree_64.c 	return node;
node               28 arch/unicore32/kernel/clock.c 	struct list_head	node;
node               62 arch/unicore32/kernel/clock.c 	list_for_each_entry(p, &clocks, node) {
node              217 arch/unicore32/kernel/clock.c 	list_add(&clk->node, &clocks);
node              228 arch/unicore32/kernel/clock.c 	list_del(&clk->node);
node               14 arch/x86/entry/vdso/vgetcpu.c __vdso_getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *unused)
node               16 arch/x86/entry/vdso/vgetcpu.c 	vdso_read_cpunode(cpu, node);
node               21 arch/x86/entry/vdso/vgetcpu.c long getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache)
node               50 arch/x86/events/amd/uncore.c 	struct hlist_node node;
node              384 arch/x86/events/amd/uncore.c 			hlist_add_head(&this->node, &uncore_unused_list);
node              424 arch/x86/events/amd/uncore.c 	hlist_for_each_entry_safe(uncore, n, &uncore_unused_list, node) {
node              425 arch/x86/events/amd/uncore.c 		hlist_del(&uncore->node);
node               86 arch/x86/events/intel/bts.c 	int node = (cpu == -1) ? cpu : cpu_to_node(cpu);
node              104 arch/x86/events/intel/bts.c 	buf = kzalloc_node(offsetof(struct bts_buffer, buf[nbuf]), GFP_KERNEL, node);
node              321 arch/x86/events/intel/ds.c 	int node = cpu_to_node(cpu);
node              324 arch/x86/events/intel/ds.c 	page = __alloc_pages_node(node, flags | __GFP_ZERO, order);
node              339 arch/x86/events/intel/ds.c 	int max, node = cpu_to_node(cpu);
node              354 arch/x86/events/intel/ds.c 		insn_buff = kzalloc_node(PEBS_FIXUP_SIZE, GFP_KERNEL, node);
node              614 arch/x86/events/intel/pt.c 	int node = cpu_to_node(cpu);
node              618 arch/x86/events/intel/pt.c 	p = alloc_pages_node(node, gfp | __GFP_ZERO, 0);
node             1196 arch/x86/events/intel/pt.c 	int node, ret, cpu = event->cpu;
node             1203 arch/x86/events/intel/pt.c 	node = cpu_to_node(cpu);
node             1205 arch/x86/events/intel/pt.c 	buf = kzalloc_node(sizeof(struct pt_buffer), GFP_KERNEL, node);
node             1240 arch/x86/events/intel/pt.c 	int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu);
node             1245 arch/x86/events/intel/pt.c 	filters = kzalloc_node(sizeof(struct pt_filters), GFP_KERNEL, node);
node              314 arch/x86/events/intel/uncore.c 						 int node)
node              321 arch/x86/events/intel/uncore.c 	box = kzalloc_node(size, GFP_KERNEL, node);
node               26 arch/x86/include/asm/amd_nb.h extern int amd_smn_read(u16 node, u32 address, u32 *value);
node               27 arch/x86/include/asm/amd_nb.h extern int amd_smn_write(u16 node, u32 address, u32 value);
node               28 arch/x86/include/asm/amd_nb.h extern int amd_df_indirect_read(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo);
node               85 arch/x86/include/asm/amd_nb.h struct amd_northbridge *node_to_amd_nb(int node);
node               16 arch/x86/include/asm/device.h 	struct list_head node;
node              178 arch/x86/include/asm/io_apic.h 				  int node, int trigger, int polarity);
node              489 arch/x86/include/asm/kvm_host.h 	struct list_head node;
node               23 arch/x86/include/asm/kvm_page_track.h 	struct hlist_node node;
node               36 arch/x86/include/asm/kvm_page_track.h 			    int bytes, struct kvm_page_track_notifier_node *node);
node               46 arch/x86/include/asm/kvm_page_track.h 			    struct kvm_page_track_notifier_node *node);
node               37 arch/x86/include/asm/numa.h static inline void set_apicid_to_node(int apicid, s16 node)
node               39 arch/x86/include/asm/numa.h 	__apicid_to_node[apicid] = node;
node               45 arch/x86/include/asm/numa.h static inline void set_apicid_to_node(int apicid, s16 node)
node               60 arch/x86/include/asm/numa.h extern void numa_set_node(int cpu, int node);
node               66 arch/x86/include/asm/numa.h static inline void numa_set_node(int cpu, int node)	{ }
node               74 arch/x86/include/asm/numa.h void debug_cpumask_set_cpu(int cpu, int node, bool enable);
node               19 arch/x86/include/asm/pci.h 	int		node;		/* NUMA node */
node              136 arch/x86/include/asm/pci.h 	return sd->node;
node              142 arch/x86/include/asm/pci.h 	int node;
node              144 arch/x86/include/asm/pci.h 	node = __pcibus_to_node(bus);
node              145 arch/x86/include/asm/pci.h 	return (node == NUMA_NO_NODE) ? cpu_online_mask :
node              146 arch/x86/include/asm/pci.h 			      cpumask_of_node(node);
node              249 arch/x86/include/asm/segment.h static inline unsigned long vdso_encode_cpunode(int cpu, unsigned long node)
node              251 arch/x86/include/asm/segment.h 	return (node << VDSO_CPUNODE_BITS) | cpu;
node              254 arch/x86/include/asm/segment.h static inline void vdso_read_cpunode(unsigned *cpu, unsigned *node)
node              273 arch/x86/include/asm/segment.h 	if (node)
node              274 arch/x86/include/asm/segment.h 		*node = (p >> VDSO_CPUNODE_BITS);
node               67 arch/x86/include/asm/topology.h extern const struct cpumask *cpumask_of_node(int node);
node               70 arch/x86/include/asm/topology.h static inline const struct cpumask *cpumask_of_node(int node)
node               72 arch/x86/include/asm/topology.h 	return node_to_cpumask_map[node];
node              205 arch/x86/include/asm/uv/uv_hub.h static inline struct uv_hub_info_s *uv_hub_info_list(int node)
node              207 arch/x86/include/asm/uv/uv_hub.h 	return (struct uv_hub_info_s *)__uv_hub_info_list[node];
node              591 arch/x86/include/asm/uv/uv_hub.h 	unsigned short sockid, node, *p2s;
node              598 arch/x86/include/asm/uv/uv_hub.h 	node = uv_socket_to_node(sockid);
node              601 arch/x86/include/asm/uv/uv_hub.h 	if (!node)
node              604 arch/x86/include/asm/uv/uv_hub.h 	base = (unsigned long)(uv_hub_info->gr_table[node - 1].limit);
node              174 arch/x86/include/asm/x86_init.h 	void (*fixup_cpu_id)(struct cpuinfo_x86 *c, int node);
node              659 arch/x86/kernel/acpi/boot.c 	int node;
node              662 arch/x86/kernel/acpi/boot.c 	node = dev ? dev_to_node(dev) : NUMA_NO_NODE;
node              665 arch/x86/kernel/acpi/boot.c 	ioapic_set_alloc_attr(&info, node, trigger, polarity);
node              109 arch/x86/kernel/amd_nb.c struct amd_northbridge *node_to_amd_nb(int node)
node              111 arch/x86/kernel/amd_nb.c 	return (node < amd_northbridges.num) ? &amd_northbridges.nb[node] : NULL;
node              126 arch/x86/kernel/amd_nb.c static int __amd_smn_rw(u16 node, u32 address, u32 *value, bool write)
node              131 arch/x86/kernel/amd_nb.c 	if (node >= amd_northbridges.num)
node              134 arch/x86/kernel/amd_nb.c 	root = node_to_amd_nb(node)->root;
node              159 arch/x86/kernel/amd_nb.c int amd_smn_read(u16 node, u32 address, u32 *value)
node              161 arch/x86/kernel/amd_nb.c 	return __amd_smn_rw(node, address, value, false);
node              165 arch/x86/kernel/amd_nb.c int amd_smn_write(u16 node, u32 address, u32 value)
node              167 arch/x86/kernel/amd_nb.c 	return __amd_smn_rw(node, address, &value, true);
node              181 arch/x86/kernel/amd_nb.c int amd_df_indirect_read(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo)
node              187 arch/x86/kernel/amd_nb.c 	if (node >= amd_northbridges.num)
node              190 arch/x86/kernel/amd_nb.c 	F4 = node_to_amd_nb(node)->link;
node              391 arch/x86/kernel/aperture_64.c 	int i, node;
node              406 arch/x86/kernel/aperture_64.c 	node = 0;
node              442 arch/x86/kernel/aperture_64.c 				node, aper_base, aper_base + aper_size - 1,
node              444 arch/x86/kernel/aperture_64.c 			node++;
node              170 arch/x86/kernel/apic/apic_numachip.c static void fixup_cpu_id(struct cpuinfo_x86 *c, int node)
node              175 arch/x86/kernel/apic/apic_numachip.c 	this_cpu_write(cpu_llc_id, node);
node              183 arch/x86/kernel/apic/apic_numachip.c 	c->phys_proc_id = node / nodes;
node              375 arch/x86/kernel/apic/io_apic.c 				 int node, int apic, int pin)
node              384 arch/x86/kernel/apic/io_apic.c 	entry = kzalloc_node(sizeof(struct irq_pin_list), GFP_ATOMIC, node);
node              387 arch/x86/kernel/apic/io_apic.c 		       node, apic, pin);
node              410 arch/x86/kernel/apic/io_apic.c 				int node, int apic, int pin)
node              412 arch/x86/kernel/apic/io_apic.c 	if (__add_pin_to_irq_node(data, node, apic, pin))
node              419 arch/x86/kernel/apic/io_apic.c static void __init replace_pin_at_irq_node(struct mp_chip_data *data, int node,
node              435 arch/x86/kernel/apic/io_apic.c 	add_pin_to_irq_node(data, node, newapic, newpin);
node              871 arch/x86/kernel/apic/io_apic.c void ioapic_set_alloc_attr(struct irq_alloc_info *info, int node,
node              876 arch/x86/kernel/apic/io_apic.c 	info->ioapic_node = node;
node             1007 arch/x86/kernel/apic/io_apic.c 	int node = ioapic_alloc_attr_node(info);
node             1017 arch/x86/kernel/apic/io_apic.c 		if (__add_pin_to_irq_node(irq_data->chip_data, node, ioapic,
node             1022 arch/x86/kernel/apic/io_apic.c 		irq = __irq_domain_alloc_irqs(domain, irq, 1, node, info, true,
node             2131 arch/x86/kernel/apic/io_apic.c 	int node = cpu_to_node(0);
node             2220 arch/x86/kernel/apic/io_apic.c 		replace_pin_at_irq_node(data, node, apic1, pin1, apic2, pin2);
node              373 arch/x86/kernel/apic/msi.c int dmar_alloc_hwirq(int id, int node, void *arg)
node              386 arch/x86/kernel/apic/msi.c 	return irq_domain_alloc_irqs(domain, 1, node, &info);
node              102 arch/x86/kernel/apic/vector.c static struct apic_chip_data *alloc_apic_chip_data(int node)
node              106 arch/x86/kernel/apic/vector.c 	apicd = kzalloc_node(sizeof(*apicd), GFP_KERNEL, node);
node              273 arch/x86/kernel/apic/vector.c 	int node = irq_data_get_node(irqd);
node              275 arch/x86/kernel/apic/vector.c 	if (node == NUMA_NO_NODE)
node              278 arch/x86/kernel/apic/vector.c 	cpumask_and(vector_searchmask, cpumask_of_node(node), affmsk);
node              282 arch/x86/kernel/apic/vector.c 	if (!assign_vector_locked(irqd, cpumask_of_node(node)))
node              534 arch/x86/kernel/apic/vector.c 	int i, err, node;
node              546 arch/x86/kernel/apic/vector.c 		node = irq_data_get_node(irqd);
node              548 arch/x86/kernel/apic/vector.c 		apicd = alloc_apic_chip_data(node);
node               14 arch/x86/kernel/apic/x2apic_cluster.c 	int		node;
node              124 arch/x86/kernel/apic/x2apic_cluster.c static int alloc_clustermask(unsigned int cpu, int node)
node              133 arch/x86/kernel/apic/x2apic_cluster.c 		if (cluster_hotplug_mask->node == node)
node              139 arch/x86/kernel/apic/x2apic_cluster.c 					    GFP_KERNEL, node);
node              142 arch/x86/kernel/apic/x2apic_cluster.c 	cluster_hotplug_mask->node = node;
node              291 arch/x86/kernel/cpu/amd.c 	int i, node;
node              294 arch/x86/kernel/cpu/amd.c 		node = __apicid_to_node[i];
node              295 arch/x86/kernel/cpu/amd.c 		if (node != NUMA_NO_NODE && node_online(node))
node              296 arch/x86/kernel/cpu/amd.c 			return node;
node              299 arch/x86/kernel/cpu/amd.c 		node = __apicid_to_node[i];
node              300 arch/x86/kernel/cpu/amd.c 		if (node != NUMA_NO_NODE && node_online(node))
node              301 arch/x86/kernel/cpu/amd.c 			return node;
node              420 arch/x86/kernel/cpu/amd.c 	int node;
node              423 arch/x86/kernel/cpu/amd.c 	node = numa_cpu_node(cpu);
node              424 arch/x86/kernel/cpu/amd.c 	if (node == NUMA_NO_NODE)
node              425 arch/x86/kernel/cpu/amd.c 		node = per_cpu(cpu_llc_id, cpu);
node              433 arch/x86/kernel/cpu/amd.c 		x86_cpuinit.fixup_cpu_id(c, node);
node              435 arch/x86/kernel/cpu/amd.c 	if (!node_online(node)) {
node              458 arch/x86/kernel/cpu/amd.c 			node = __apicid_to_node[ht_nodeid];
node              460 arch/x86/kernel/cpu/amd.c 		if (!node_online(node))
node              461 arch/x86/kernel/cpu/amd.c 			node = nearby_node(apicid);
node              463 arch/x86/kernel/cpu/amd.c 	numa_set_node(cpu, node);
node              577 arch/x86/kernel/cpu/cacheinfo.c 	int node;
node              583 arch/x86/kernel/cpu/cacheinfo.c 	node = amd_get_nb_id(smp_processor_id());
node              584 arch/x86/kernel/cpu/cacheinfo.c 	this_leaf->nb = node_to_amd_nb(node);
node               37 arch/x86/kernel/cpu/hygon.c 	int i, node;
node               40 arch/x86/kernel/cpu/hygon.c 		node = __apicid_to_node[i];
node               41 arch/x86/kernel/cpu/hygon.c 		if (node != NUMA_NO_NODE && node_online(node))
node               42 arch/x86/kernel/cpu/hygon.c 			return node;
node               45 arch/x86/kernel/cpu/hygon.c 		node = __apicid_to_node[i];
node               46 arch/x86/kernel/cpu/hygon.c 		if (node != NUMA_NO_NODE && node_online(node))
node               47 arch/x86/kernel/cpu/hygon.c 			return node;
node              132 arch/x86/kernel/cpu/hygon.c 	int node;
node              135 arch/x86/kernel/cpu/hygon.c 	node = numa_cpu_node(cpu);
node              136 arch/x86/kernel/cpu/hygon.c 	if (node == NUMA_NO_NODE)
node              137 arch/x86/kernel/cpu/hygon.c 		node = per_cpu(cpu_llc_id, cpu);
node              145 arch/x86/kernel/cpu/hygon.c 		x86_cpuinit.fixup_cpu_id(c, node);
node              147 arch/x86/kernel/cpu/hygon.c 	if (!node_online(node)) {
node              169 arch/x86/kernel/cpu/hygon.c 			node = __apicid_to_node[ht_nodeid];
node              171 arch/x86/kernel/cpu/hygon.c 		if (!node_online(node))
node              172 arch/x86/kernel/cpu/hygon.c 			node = nearby_node(apicid);
node              174 arch/x86/kernel/cpu/hygon.c 	numa_set_node(cpu, node);
node              483 arch/x86/kernel/cpu/intel.c 	unsigned node;
node              488 arch/x86/kernel/cpu/intel.c 	node = numa_cpu_node(cpu);
node              489 arch/x86/kernel/cpu/intel.c 	if (node == NUMA_NO_NODE || !node_online(node)) {
node              491 arch/x86/kernel/cpu/intel.c 		node = cpu_to_node(cpu);
node              493 arch/x86/kernel/cpu/intel.c 	numa_set_node(cpu, node);
node               34 arch/x86/kernel/cpu/mce/genpool.c 	struct mce_evt_llist *node;
node               39 arch/x86/kernel/cpu/mce/genpool.c 	llist_for_each_entry(node, &l->llnode, llnode) {
node               40 arch/x86/kernel/cpu/mce/genpool.c 		m2 = &node->mce;
node               59 arch/x86/kernel/cpu/mce/genpool.c 	struct mce_evt_llist *node, *t;
node               66 arch/x86/kernel/cpu/mce/genpool.c 	llist_for_each_entry_safe(node, t, head, llnode) {
node               67 arch/x86/kernel/cpu/mce/genpool.c 		if (!is_duplicate_mce_record(node, t))
node               68 arch/x86/kernel/cpu/mce/genpool.c 			llist_add(&node->llnode, &new_head);
node               77 arch/x86/kernel/cpu/mce/genpool.c 	struct mce_evt_llist *node, *tmp;
node               85 arch/x86/kernel/cpu/mce/genpool.c 	llist_for_each_entry_safe(node, tmp, head, llnode) {
node               86 arch/x86/kernel/cpu/mce/genpool.c 		mce = &node->mce;
node               88 arch/x86/kernel/cpu/mce/genpool.c 		gen_pool_free(mce_evt_pool, (unsigned long)node, sizeof(*node));
node               99 arch/x86/kernel/cpu/mce/genpool.c 	struct mce_evt_llist *node;
node              107 arch/x86/kernel/cpu/mce/genpool.c 	node = (void *)gen_pool_alloc(mce_evt_pool, sizeof(*node));
node              108 arch/x86/kernel/cpu/mce/genpool.c 	if (!node) {
node              113 arch/x86/kernel/cpu/mce/genpool.c 	memcpy(&node->mce, mce, sizeof(*mce));
node              114 arch/x86/kernel/cpu/mce/genpool.c 	llist_add(&node->llnode, &mce_event_llist);
node               33 arch/x86/kernel/devicetree.c void __init early_init_dt_scan_chosen_arch(unsigned long node)
node              138 arch/x86/kernel/espfix_64.c 	int n, node;
node              161 arch/x86/kernel/espfix_64.c 	node = cpu_to_node(cpu);
node              167 arch/x86/kernel/espfix_64.c 		struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0);
node              179 arch/x86/kernel/espfix_64.c 		struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0);
node              189 arch/x86/kernel/espfix_64.c 	stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0));
node              114 arch/x86/kernel/irq_32.c 	int node = cpu_to_node(cpu);
node              120 arch/x86/kernel/irq_32.c 	ph = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER);
node              123 arch/x86/kernel/irq_32.c 	ps = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER);
node               32 arch/x86/kernel/kdebugfs.c 	struct setup_data_node *node = file->private_data;
node               41 arch/x86/kernel/kdebugfs.c 	if (pos >= node->len)
node               44 arch/x86/kernel/kdebugfs.c 	if (count > node->len - pos)
node               45 arch/x86/kernel/kdebugfs.c 		count = node->len - pos;
node               47 arch/x86/kernel/kdebugfs.c 	pa = node->paddr + sizeof(struct setup_data) + pos;
node               72 arch/x86/kernel/kdebugfs.c 		       struct setup_data_node *node)
node               80 arch/x86/kernel/kdebugfs.c 	debugfs_create_x32("type", S_IRUGO, d, &node->type);
node               81 arch/x86/kernel/kdebugfs.c 	debugfs_create_file("data", S_IRUGO, d, node, &fops_setup_data);
node               86 arch/x86/kernel/kdebugfs.c 	struct setup_data_node *node;
node               98 arch/x86/kernel/kdebugfs.c 		node = kmalloc(sizeof(*node), GFP_KERNEL);
node               99 arch/x86/kernel/kdebugfs.c 		if (!node) {
node              106 arch/x86/kernel/kdebugfs.c 			kfree(node);
node              111 arch/x86/kernel/kdebugfs.c 		node->paddr = pa_data;
node              112 arch/x86/kernel/kdebugfs.c 		node->type = data->type;
node              113 arch/x86/kernel/kdebugfs.c 		node->len = data->len;
node              114 arch/x86/kernel/kdebugfs.c 		create_setup_data_node(d, no, node);
node              540 arch/x86/kernel/quirks.c 	u32 node;
node              549 arch/x86/kernel/quirks.c 	node = pcibus_to_node(dev->bus) | (val & 7);
node              554 arch/x86/kernel/quirks.c 	if (node_online(node))
node              555 arch/x86/kernel/quirks.c 		set_dev_node(&dev->dev, node);
node               74 arch/x86/kernel/setup_percpu.c 		int node = early_cpu_to_node(cpu);
node               76 arch/x86/kernel/setup_percpu.c 		if (node_online(node) && NODE_DATA(node) &&
node               77 arch/x86/kernel/setup_percpu.c 		    last && last != NODE_DATA(node))
node               80 arch/x86/kernel/setup_percpu.c 		last = NODE_DATA(node);
node              105 arch/x86/kernel/setup_percpu.c 	int node = early_cpu_to_node(cpu);
node              108 arch/x86/kernel/setup_percpu.c 	if (!node_online(node) || !NODE_DATA(node)) {
node              111 arch/x86/kernel/setup_percpu.c 			cpu, node);
node              117 arch/x86/kernel/setup_percpu.c 					     node);
node              120 arch/x86/kernel/setup_percpu.c 			 cpu, size, node, __pa(ptr));
node              905 arch/x86/kernel/smpboot.c 	int node = early_cpu_to_node(cpu);
node              918 arch/x86/kernel/smpboot.c 		if (node != current_node) {
node              921 arch/x86/kernel/smpboot.c 			current_node = node;
node              924 arch/x86/kernel/smpboot.c 			       node_width - num_digits(node), " ", node);
node              935 arch/x86/kernel/smpboot.c 			node, cpu, apicid);
node             5433 arch/x86/kvm/mmu.c 			      struct kvm_page_track_notifier_node *node)
node             5831 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *node;
node             5835 arch/x86/kvm/mmu.c 	list_for_each_entry_safe_reverse(sp, node,
node             5926 arch/x86/kvm/mmu.c 			struct kvm_page_track_notifier_node *node)
node             5933 arch/x86/kvm/mmu.c 	struct kvm_page_track_notifier_node *node = &kvm->arch.mmu_sp_tracker;
node             5935 arch/x86/kvm/mmu.c 	node->track_write = kvm_mmu_pte_write;
node             5936 arch/x86/kvm/mmu.c 	node->track_flush_slot = kvm_mmu_invalidate_zap_pages_in_memslot;
node             5937 arch/x86/kvm/mmu.c 	kvm_page_track_register_notifier(kvm, node);
node             5942 arch/x86/kvm/mmu.c 	struct kvm_page_track_notifier_node *node = &kvm->arch.mmu_sp_tracker;
node             5944 arch/x86/kvm/mmu.c 	kvm_page_track_unregister_notifier(kvm, node);
node             6123 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *node;
node             6129 arch/x86/kvm/mmu.c 	list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link) {
node              356 arch/x86/kvm/mtrr.c 		list_del(&mtrr_state->var_ranges[index].node);
node              369 arch/x86/kvm/mtrr.c 		list_for_each_entry(tmp, &mtrr_state->head, node)
node              372 arch/x86/kvm/mtrr.c 		list_add_tail(&cur->node, &tmp->node);
node              522 arch/x86/kvm/mtrr.c 	list_for_each_entry_continue(iter->range, &mtrr_state->head, node)
node              537 arch/x86/kvm/mtrr.c 	iter->range = list_prepare_entry(iter->range, &mtrr_state->head, node);
node              192 arch/x86/kvm/page_track.c 	hlist_add_head_rcu(&n->node, &head->track_notifier_list);
node              210 arch/x86/kvm/page_track.c 	hlist_del_rcu(&n->node);
node              236 arch/x86/kvm/page_track.c 	hlist_for_each_entry_rcu(n, &head->track_notifier_list, node)
node              261 arch/x86/kvm/page_track.c 	hlist_for_each_entry_rcu(n, &head->track_notifier_list, node)
node              255 arch/x86/kvm/svm.c 	struct list_head node;	/* Used by SVM for per-vcpu ir_list */
node             2066 arch/x86/kvm/svm.c 	list_for_each_entry(ir, &svm->ir_list, node) {
node             5221 arch/x86/kvm/svm.c 	list_for_each_entry(cur, &svm->ir_list, node) {
node             5224 arch/x86/kvm/svm.c 		list_del(&cur->node);
node             5269 arch/x86/kvm/svm.c 	list_add(&ir->node, &svm->ir_list);
node             2504 arch/x86/kvm/vmx/vmx.c 	int node = cpu_to_node(cpu);
node             2508 arch/x86/kvm/vmx/vmx.c 	pages = __alloc_pages_node(node, flags, vmcs_config.order);
node             1450 arch/x86/mm/init_64.c 		unsigned long end, int node, struct vmem_altmap *altmap)
node             1462 arch/x86/mm/init_64.c 		pgd = vmemmap_pgd_populate(addr, node);
node             1466 arch/x86/mm/init_64.c 		p4d = vmemmap_p4d_populate(pgd, addr, node);
node             1470 arch/x86/mm/init_64.c 		pud = vmemmap_pud_populate(p4d, addr, node);
node             1481 arch/x86/mm/init_64.c 				p = vmemmap_alloc_block_buf(PMD_SIZE, node);
node             1490 arch/x86/mm/init_64.c 				if (p_end != p || node_start != node) {
node             1495 arch/x86/mm/init_64.c 					node_start = node;
node             1505 arch/x86/mm/init_64.c 			vmemmap_verify((pte_t *)pmd, node, addr, next);
node             1508 arch/x86/mm/init_64.c 		if (vmemmap_populate_basepages(addr, next, node))
node             1514 arch/x86/mm/init_64.c int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
node             1520 arch/x86/mm/init_64.c 		err = vmemmap_populate_basepages(start, end, node);
node             1522 arch/x86/mm/init_64.c 		err = vmemmap_populate_hugepages(start, end, node, altmap);
node             1528 arch/x86/mm/init_64.c 		err = vmemmap_populate_basepages(start, end, node);
node               80 arch/x86/mm/numa.c void numa_set_node(int cpu, int node)
node               86 arch/x86/mm/numa.c 		cpu_to_node_map[cpu] = node;
node               97 arch/x86/mm/numa.c 	per_cpu(x86_cpu_to_node_map, cpu) = node;
node               99 arch/x86/mm/numa.c 	set_cpu_numa_node(cpu, node);
node              116 arch/x86/mm/numa.c 	unsigned int node;
node              123 arch/x86/mm/numa.c 	for (node = 0; node < nr_node_ids; node++)
node              124 arch/x86/mm/numa.c 		alloc_bootmem_cpumask_var(&node_to_cpumask_map[node]);
node              758 arch/x86/mm/numa.c 		int node = numa_cpu_node(cpu);
node              760 arch/x86/mm/numa.c 		if (node == NUMA_NO_NODE)
node              763 arch/x86/mm/numa.c 		if (!node_online(node))
node              764 arch/x86/mm/numa.c 			init_memory_less_node(node);
node              766 arch/x86/mm/numa.c 		numa_set_node(cpu, node);
node              816 arch/x86/mm/numa.c void debug_cpumask_set_cpu(int cpu, int node, bool enable)
node              820 arch/x86/mm/numa.c 	if (node == NUMA_NO_NODE) {
node              824 arch/x86/mm/numa.c 	mask = node_to_cpumask_map[node];
node              826 arch/x86/mm/numa.c 		pr_err("node_to_cpumask_map[%i] NULL\n", node);
node              838 arch/x86/mm/numa.c 		cpu, node, cpumask_pr_args(mask));
node              862 arch/x86/mm/numa.c const struct cpumask *cpumask_of_node(int node)
node              864 arch/x86/mm/numa.c 	if ((unsigned)node >= nr_node_ids) {
node              867 arch/x86/mm/numa.c 			node, nr_node_ids);
node              871 arch/x86/mm/numa.c 	if (node_to_cpumask_map[node] == NULL) {
node              874 arch/x86/mm/numa.c 			node);
node              878 arch/x86/mm/numa.c 	return node_to_cpumask_map[node];
node               39 arch/x86/mm/pat_rbtree.c static int is_node_overlap(struct memtype *node, u64 start, u64 end)
node               41 arch/x86/mm/pat_rbtree.c 	if (node->start >= end || node->end <= start)
node               47 arch/x86/mm/pat_rbtree.c static u64 get_subtree_max_end(struct rb_node *node)
node               50 arch/x86/mm/pat_rbtree.c 	if (node) {
node               51 arch/x86/mm/pat_rbtree.c 		struct memtype *data = rb_entry(node, struct memtype, rb);
node               57 arch/x86/mm/pat_rbtree.c #define NODE_END(node) ((node)->end)
node               66 arch/x86/mm/pat_rbtree.c 	struct rb_node *node = root->rb_node;
node               69 arch/x86/mm/pat_rbtree.c 	while (node) {
node               70 arch/x86/mm/pat_rbtree.c 		struct memtype *data = rb_entry(node, struct memtype, rb);
node               72 arch/x86/mm/pat_rbtree.c 		if (get_subtree_max_end(node->rb_left) > start) {
node               74 arch/x86/mm/pat_rbtree.c 			node = node->rb_left;
node               80 arch/x86/mm/pat_rbtree.c 			node = node->rb_right;
node              100 arch/x86/mm/pat_rbtree.c 		struct rb_node *node;
node              110 arch/x86/mm/pat_rbtree.c 		node = rb_next(&match->rb);
node              111 arch/x86/mm/pat_rbtree.c 		if (node)
node              112 arch/x86/mm/pat_rbtree.c 			match = rb_entry(node, struct memtype, rb);
node              125 arch/x86/mm/pat_rbtree.c 	struct rb_node *node;
node              139 arch/x86/mm/pat_rbtree.c 	node = rb_next(&match->rb);
node              140 arch/x86/mm/pat_rbtree.c 	while (node) {
node              141 arch/x86/mm/pat_rbtree.c 		match = rb_entry(node, struct memtype, rb);
node              151 arch/x86/mm/pat_rbtree.c 		node = rb_next(&match->rb);
node              168 arch/x86/mm/pat_rbtree.c 	struct rb_node **node = &(root->rb_node);
node              171 arch/x86/mm/pat_rbtree.c 	while (*node) {
node              172 arch/x86/mm/pat_rbtree.c 		struct memtype *data = rb_entry(*node, struct memtype, rb);
node              174 arch/x86/mm/pat_rbtree.c 		parent = *node;
node              178 arch/x86/mm/pat_rbtree.c 			node = &((*node)->rb_left);
node              180 arch/x86/mm/pat_rbtree.c 			node = &((*node)->rb_right);
node              184 arch/x86/mm/pat_rbtree.c 	rb_link_node(&newdata->rb, parent, node);
node              251 arch/x86/mm/pat_rbtree.c 	struct rb_node *node;
node              254 arch/x86/mm/pat_rbtree.c 	node = rb_first(&memtype_rbroot);
node              255 arch/x86/mm/pat_rbtree.c 	while (node && pos != i) {
node              256 arch/x86/mm/pat_rbtree.c 		node = rb_next(node);
node              260 arch/x86/mm/pat_rbtree.c 	if (node) { /* pos == i */
node              261 arch/x86/mm/pat_rbtree.c 		struct memtype *this = rb_entry(node, struct memtype, rb);
node               30 arch/x86/mm/srat.c 	int pxm, node;
node               48 arch/x86/mm/srat.c 	node = acpi_map_pxm_to_node(pxm);
node               49 arch/x86/mm/srat.c 	if (node < 0) {
node               56 arch/x86/mm/srat.c 		printk(KERN_INFO "SRAT: PXM %u -> APIC 0x%04x -> Node %u skipped apicid that is too big\n", pxm, apic_id, node);
node               59 arch/x86/mm/srat.c 	set_apicid_to_node(apic_id, node);
node               60 arch/x86/mm/srat.c 	node_set(node, numa_nodes_parsed);
node               62 arch/x86/mm/srat.c 	       pxm, apic_id, node);
node               69 arch/x86/mm/srat.c 	int pxm, node;
node               83 arch/x86/mm/srat.c 	node = acpi_map_pxm_to_node(pxm);
node               84 arch/x86/mm/srat.c 	if (node < 0) {
node               96 arch/x86/mm/srat.c 		printk(KERN_INFO "SRAT: PXM %u -> APIC 0x%02x -> Node %u skipped apicid that is too big\n", pxm, apic_id, node);
node              100 arch/x86/mm/srat.c 	set_apicid_to_node(apic_id, node);
node              101 arch/x86/mm/srat.c 	node_set(node, numa_nodes_parsed);
node              103 arch/x86/mm/srat.c 	       pxm, apic_id, node);
node              249 arch/x86/pci/acpi.c 	int node = acpi_get_node(device->handle);
node              251 arch/x86/pci/acpi.c 	if (node == NUMA_NO_NODE) {
node              252 arch/x86/pci/acpi.c 		node = x86_pci_root_bus_node(busnum);
node              253 arch/x86/pci/acpi.c 		if (node != 0 && node != NUMA_NO_NODE)
node              255 arch/x86/pci/acpi.c 				node);
node              257 arch/x86/pci/acpi.c 	if (node != NUMA_NO_NODE && !node_online(node))
node              258 arch/x86/pci/acpi.c 		node = NUMA_NO_NODE;
node              260 arch/x86/pci/acpi.c 	return node;
node              330 arch/x86/pci/acpi.c 	int node = pci_acpi_root_get_node(root);
node              351 arch/x86/pci/acpi.c 			.node = node,
node              366 arch/x86/pci/acpi.c 			info->sd.node = node;
node              378 arch/x86/pci/acpi.c 		list_for_each_entry(child, &bus->children, node)
node               42 arch/x86/pci/amd_bus.c static struct pci_root_info __init *find_pci_root_info(int node, int link)
node               48 arch/x86/pci/amd_bus.c 		if (info->node == node && info->link == link)
node               65 arch/x86/pci/amd_bus.c 	int node;
node              126 arch/x86/pci/amd_bus.c 		node = (reg >> 4) & 0x07;
node              129 arch/x86/pci/amd_bus.c 		info = alloc_pci_root_info(min_bus, max_bus, node, link);
node              159 arch/x86/pci/amd_bus.c 		node = reg & 0x07;
node              163 arch/x86/pci/amd_bus.c 		info = find_pci_root_info(node, link);
node              168 arch/x86/pci/amd_bus.c 		       node, link, start, end);
node              226 arch/x86/pci/amd_bus.c 		node = reg & 0x07;
node              232 arch/x86/pci/amd_bus.c 		info = find_pci_root_info(node, link);
node              238 arch/x86/pci/amd_bus.c 		       node, link, start, end);
node              321 arch/x86/pci/amd_bus.c 		       &info->busn, info->node, info->link);
node               28 arch/x86/pci/bus_numa.c 	return info->node;
node               71 arch/x86/pci/bus_numa.c 						 int node, int link)
node               87 arch/x86/pci/bus_numa.c 	info->node = node;
node               18 arch/x86/pci/bus_numa.h 	int node;
node               24 arch/x86/pci/bus_numa.h 						int node, int link);
node              468 arch/x86/pci/common.c 	sd->node = x86_pci_root_bus_node(busnum);
node              635 arch/x86/pci/common.c 	list_add(&domain->node, &dma_domain_list);
node              643 arch/x86/pci/common.c 	list_del(&domain->node);
node              653 arch/x86/pci/common.c 	list_for_each_entry(domain, &dma_domain_list, node) {
node              240 arch/x86/pci/i386.c 	list_for_each_entry(child, &bus->children, node)
node              360 arch/x86/pci/i386.c 		list_for_each_entry(bus, &pci_root_buses, node)
node              394 arch/x86/pci/i386.c 	list_for_each_entry(bus, &pci_root_buses, node)
node              397 arch/x86/pci/i386.c 	list_for_each_entry(bus, &pci_root_buses, node)
node              399 arch/x86/pci/i386.c 	list_for_each_entry(bus, &pci_root_buses, node)
node               59 arch/x86/platform/olpc/olpc-xo1-rtc.c 	struct device_node *node;
node               61 arch/x86/platform/olpc/olpc-xo1-rtc.c 	node = of_find_compatible_node(NULL, NULL, "olpc,xo1-rtc");
node               62 arch/x86/platform/olpc/olpc-xo1-rtc.c 	if (!node)
node               64 arch/x86/platform/olpc/olpc-xo1-rtc.c 	of_node_put(node);
node               22 arch/x86/platform/olpc/olpc_dt.c static phandle __init olpc_dt_getsibling(phandle node)
node               24 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node };
node               25 arch/x86/platform/olpc/olpc_dt.c 	void *res[] = { &node };
node               27 arch/x86/platform/olpc/olpc_dt.c 	if ((s32)node == -1)
node               30 arch/x86/platform/olpc/olpc_dt.c 	if (olpc_ofw("peer", args, res) || (s32)node == -1)
node               33 arch/x86/platform/olpc/olpc_dt.c 	return node;
node               36 arch/x86/platform/olpc/olpc_dt.c static phandle __init olpc_dt_getchild(phandle node)
node               38 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node };
node               39 arch/x86/platform/olpc/olpc_dt.c 	void *res[] = { &node };
node               41 arch/x86/platform/olpc/olpc_dt.c 	if ((s32)node == -1)
node               44 arch/x86/platform/olpc/olpc_dt.c 	if (olpc_ofw("child", args, res) || (s32)node == -1) {
node               49 arch/x86/platform/olpc/olpc_dt.c 	return node;
node               52 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_getproplen(phandle node, const char *prop)
node               54 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node, prop };
node               58 arch/x86/platform/olpc/olpc_dt.c 	if ((s32)node == -1)
node               69 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_getproperty(phandle node, const char *prop,
node               74 arch/x86/platform/olpc/olpc_dt.c 	plen = olpc_dt_getproplen(node, prop);
node               78 arch/x86/platform/olpc/olpc_dt.c 		const void *args[] = { (void *)node, prop, buf, (void *)plen };
node               90 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_nextprop(phandle node, char *prev, char *buf)
node               92 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node, prev, buf };
node               98 arch/x86/platform/olpc/olpc_dt.c 	if ((s32)node == -1)
node              107 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_pkg2path(phandle node, char *buf,
node              110 arch/x86/platform/olpc/olpc_dt.c 	const void *args[] = { (void *)node, buf, (void *)buflen };
node              113 arch/x86/platform/olpc/olpc_dt.c 	if ((s32)node == -1)
node              168 arch/x86/platform/olpc/olpc_dt.c 	phandle node;
node              170 arch/x86/platform/olpc/olpc_dt.c 	void *res[] = { &node };
node              177 arch/x86/platform/olpc/olpc_dt.c 	if ((s32) node == -1)
node              180 arch/x86/platform/olpc/olpc_dt.c 	return node;
node              203 arch/x86/platform/olpc/olpc_dt.c 	phandle node;
node              207 arch/x86/platform/olpc/olpc_dt.c 	node = olpc_dt_finddevice("/");
node              208 arch/x86/platform/olpc/olpc_dt.c 	if (!node)
node              211 arch/x86/platform/olpc/olpc_dt.c 	r = olpc_dt_getproperty(node, "board-revision-int",
node              219 arch/x86/platform/olpc/olpc_dt.c static int __init olpc_dt_compatible_match(phandle node, const char *compat)
node              224 arch/x86/platform/olpc/olpc_dt.c 	plen = olpc_dt_getproperty(node, "compatible", buf, sizeof(buf));
node              239 arch/x86/platform/olpc/olpc_dt.c 	phandle node;
node              242 arch/x86/platform/olpc/olpc_dt.c 	node = olpc_dt_finddevice("/battery@0");
node              243 arch/x86/platform/olpc/olpc_dt.c 	if (!node)
node              253 arch/x86/platform/olpc/olpc_dt.c 		if (olpc_dt_compatible_match(node, "olpc,xo1.5-battery"))
node              261 arch/x86/platform/olpc/olpc_dt.c 		if (olpc_dt_compatible_match(node, "olpc,xo1-battery")) {
node              280 arch/x86/platform/olpc/olpc_dt.c 		if (olpc_dt_compatible_match(node, "olpc,xo1-battery")) {
node              176 arch/x86/platform/uv/tlb_uv.c 	int node, b;
node              178 arch/x86/platform/uv/tlb_uv.c 	for_each_online_node(node) {
node              179 arch/x86/platform/uv/tlb_uv.c 		b = uv_node_to_blade_id(node);
node              181 arch/x86/platform/uv/tlb_uv.c 			return node;
node             1710 arch/x86/platform/uv/tlb_uv.c static void activation_descriptor_init(int node, int pnode, int base_pnode)
node             1730 arch/x86/platform/uv/tlb_uv.c 	bau_desc = kmalloc_node(dsize, GFP_KERNEL, node);
node             1795 arch/x86/platform/uv/tlb_uv.c static void pq_init(int node, int pnode)
node             1806 arch/x86/platform/uv/tlb_uv.c 	vp = kmalloc_node(plsize, GFP_KERNEL, node);
node             1850 arch/x86/platform/uv/tlb_uv.c 	int node;
node             1854 arch/x86/platform/uv/tlb_uv.c 	node = uvhub_to_first_node(uvhub);
node             1857 arch/x86/platform/uv/tlb_uv.c 	activation_descriptor_init(node, pnode, base_pnode);
node             1859 arch/x86/platform/uv/tlb_uv.c 	pq_init(node, pnode);
node               53 arch/x86/um/vdso/um_vdso.c __vdso_getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *unused)
node               61 arch/x86/um/vdso/um_vdso.c 	if (node)
node               62 arch/x86/um/vdso/um_vdso.c 		*node = 0;
node               67 arch/x86/um/vdso/um_vdso.c long getcpu(unsigned *cpu, unsigned *node, struct getcpu_cache *tcache)
node              193 arch/xtensa/kernel/setup.c static int __init xtensa_dt_io_area(unsigned long node, const char *uname,
node              202 arch/xtensa/kernel/setup.c 	if (!of_flat_dt_is_compatible(node, "simple-bus"))
node              205 arch/xtensa/kernel/setup.c 	ranges = of_get_flat_dt_prop(node, "ranges", &len);
node              220 arch/xtensa/kernel/setup.c static int __init xtensa_dt_io_area(unsigned long node, const char *uname,
node              114 arch/xtensa/platforms/xtfpga/setup.c static void __init update_local_mac(struct device_node *node)
node              120 arch/xtensa/platforms/xtfpga/setup.c 	macaddr = of_get_property(node, "local-mac-address", &prop_len);
node              138 arch/xtensa/platforms/xtfpga/setup.c 	of_update_property(node, newmac);
node              509 block/bfq-cgroup.c 	bfqg = kzalloc_node(sizeof(*bfqg), gfp, q->node);
node             1211 block/bfq-cgroup.c struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node)
node             1413 block/bfq-cgroup.c struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node)
node             1418 block/bfq-cgroup.c 	bfqg = kmalloc_node(sizeof(*bfqg), GFP_KERNEL | __GFP_ZERO, node);
node             2415 block/bfq-iosched.c 	struct rb_node *parent, *node;
node             2439 block/bfq-iosched.c 		node = rb_next(&__bfqq->pos_node);
node             2441 block/bfq-iosched.c 		node = rb_prev(&__bfqq->pos_node);
node             2442 block/bfq-iosched.c 	if (!node)
node             2445 block/bfq-iosched.c 	__bfqq = rb_entry(node, struct bfq_queue, pos_node);
node             5152 block/bfq-iosched.c 				     bfqd->queue->node);
node             6432 block/bfq-iosched.c 	bfqd = kzalloc_node(sizeof(*bfqd), GFP_KERNEL, q->node);
node             6540 block/bfq-iosched.c 	bfqd->root_group = bfq_create_group_hierarchy(bfqd, q->node);
node              982 block/bfq-iosched.h struct bfq_group *bfq_create_group_hierarchy(struct bfq_data *bfqd, int node);
node             1027 block/bfq-iosched.h struct bfq_entity *bfq_entity_of(struct rb_node *node);
node               25 block/bfq-wf2q.c 	struct rb_node *node = tree->rb_node;
node               27 block/bfq-wf2q.c 	return rb_entry(node, struct bfq_entity, rb_node);
node              318 block/bfq-wf2q.c struct bfq_entity *bfq_entity_of(struct rb_node *node)
node              322 block/bfq-wf2q.c 	if (node)
node              323 block/bfq-wf2q.c 		entity = rb_entry(node, struct bfq_entity, rb_node);
node              377 block/bfq-wf2q.c 	struct rb_node **node = &root->rb_node;
node              380 block/bfq-wf2q.c 	while (*node) {
node              381 block/bfq-wf2q.c 		parent = *node;
node              385 block/bfq-wf2q.c 			node = &parent->rb_left;
node              387 block/bfq-wf2q.c 			node = &parent->rb_right;
node              390 block/bfq-wf2q.c 	rb_link_node(&entity->rb_node, parent, node);
node              406 block/bfq-wf2q.c static void bfq_update_min(struct bfq_entity *entity, struct rb_node *node)
node              410 block/bfq-wf2q.c 	if (node) {
node              411 block/bfq-wf2q.c 		child = rb_entry(node, struct bfq_entity, rb_node);
node              425 block/bfq-wf2q.c static void bfq_update_active_node(struct rb_node *node)
node              427 block/bfq-wf2q.c 	struct bfq_entity *entity = rb_entry(node, struct bfq_entity, rb_node);
node              430 block/bfq-wf2q.c 	bfq_update_min(entity, node->rb_right);
node              431 block/bfq-wf2q.c 	bfq_update_min(entity, node->rb_left);
node              444 block/bfq-wf2q.c static void bfq_update_active_tree(struct rb_node *node)
node              449 block/bfq-wf2q.c 	bfq_update_active_node(node);
node              451 block/bfq-wf2q.c 	parent = rb_parent(node);
node              455 block/bfq-wf2q.c 	if (node == parent->rb_left && parent->rb_right)
node              460 block/bfq-wf2q.c 	node = parent;
node              479 block/bfq-wf2q.c 	struct rb_node *node = &entity->rb_node;
node              488 block/bfq-wf2q.c 	if (node->rb_left)
node              489 block/bfq-wf2q.c 		node = node->rb_left;
node              490 block/bfq-wf2q.c 	else if (node->rb_right)
node              491 block/bfq-wf2q.c 		node = node->rb_right;
node              493 block/bfq-wf2q.c 	bfq_update_active_tree(node);
node              553 block/bfq-wf2q.c static struct rb_node *bfq_find_deepest(struct rb_node *node)
node              557 block/bfq-wf2q.c 	if (!node->rb_right && !node->rb_left)
node              558 block/bfq-wf2q.c 		deepest = rb_parent(node);
node              559 block/bfq-wf2q.c 	else if (!node->rb_right)
node              560 block/bfq-wf2q.c 		deepest = node->rb_left;
node              561 block/bfq-wf2q.c 	else if (!node->rb_left)
node              562 block/bfq-wf2q.c 		deepest = node->rb_right;
node              564 block/bfq-wf2q.c 		deepest = rb_next(node);
node              567 block/bfq-wf2q.c 		else if (rb_parent(deepest) != node)
node              583 block/bfq-wf2q.c 	struct rb_node *node;
node              590 block/bfq-wf2q.c 	node = bfq_find_deepest(&entity->rb_node);
node              593 block/bfq-wf2q.c 	if (node)
node              594 block/bfq-wf2q.c 		bfq_update_active_tree(node);
node             1362 block/bfq-wf2q.c 	struct rb_node *node = st->active.rb_node;
node             1364 block/bfq-wf2q.c 	while (node) {
node             1365 block/bfq-wf2q.c 		entry = rb_entry(node, struct bfq_entity, rb_node);
node             1370 block/bfq-wf2q.c 		if (node->rb_left) {
node             1371 block/bfq-wf2q.c 			entry = rb_entry(node->rb_left,
node             1374 block/bfq-wf2q.c 				node = node->rb_left;
node             1380 block/bfq-wf2q.c 		node = node->rb_right;
node              152 block/blk-cgroup.c 	blkg = kzalloc_node(sizeof(*blkg), gfp_mask, q->node);
node              507 block/blk-core.c 	q->node = node_id;
node              960 block/blk-core.c 	create_io_context(GFP_ATOMIC, q->node);
node              476 block/blk-flush.c 		int node, int cmd_size, gfp_t flags)
node              481 block/blk-flush.c 	fq = kzalloc_node(sizeof(*fq), flags, node);
node              488 block/blk-flush.c 	fq->flush_rq = kzalloc_node(rq_sz, flags, node);
node              250 block/blk-ioc.c int create_task_io_context(struct task_struct *task, gfp_t gfp_flags, int node)
node              256 block/blk-ioc.c 				    node);
node              304 block/blk-ioc.c 				       gfp_t gfp_flags, int node)
node              319 block/blk-ioc.c 	} while (!create_task_io_context(task, gfp_flags, node));
node              380 block/blk-ioc.c 				    q->node);
node             2009 block/blk-iocost.c 			    gfp, q->node);
node              943 block/blk-iolatency.c 	iolat = kzalloc_node(sizeof(*iolat), gfp, q->node);
node              437 block/blk-mq-tag.c 		    bool round_robin, int node)
node              440 block/blk-mq-tag.c 				       node);
node              444 block/blk-mq-tag.c 						   int node, int alloc_policy)
node              449 block/blk-mq-tag.c 	if (bt_alloc(&tags->bitmap_tags, depth, round_robin, node))
node              452 block/blk-mq-tag.c 		     node))
node              465 block/blk-mq-tag.c 				     int node, int alloc_policy)
node              474 block/blk-mq-tag.c 	tags = kzalloc_node(sizeof(*tags), GFP_KERNEL, node);
node              481 block/blk-mq-tag.c 	return blk_mq_init_bitmap_tags(tags, node, alloc_policy);
node               25 block/blk-mq-tag.h extern struct blk_mq_tags *blk_mq_init_tags(unsigned int nr_tags, unsigned int reserved_tags, int node, int alloc_policy);
node             2106 block/blk-mq.c 	int node;
node             2108 block/blk-mq.c 	node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], hctx_idx);
node             2109 block/blk-mq.c 	if (node == NUMA_NO_NODE)
node             2110 block/blk-mq.c 		node = set->numa_node;
node             2112 block/blk-mq.c 	tags = blk_mq_init_tags(nr_tags, reserved_tags, node,
node             2119 block/blk-mq.c 				 node);
node             2127 block/blk-mq.c 					node);
node             2143 block/blk-mq.c 			       unsigned int hctx_idx, int node)
node             2148 block/blk-mq.c 		ret = set->ops->init_request(set, rq, hctx_idx, node);
node             2162 block/blk-mq.c 	int node;
node             2164 block/blk-mq.c 	node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], hctx_idx);
node             2165 block/blk-mq.c 	if (node == NUMA_NO_NODE)
node             2166 block/blk-mq.c 		node = set->numa_node;
node             2188 block/blk-mq.c 			page = alloc_pages_node(node,
node             2218 block/blk-mq.c 			if (blk_mq_init_request(set, rq, hctx_idx, node)) {
node             2239 block/blk-mq.c static int blk_mq_hctx_notify_dead(unsigned int cpu, struct hlist_node *node)
node             2246 block/blk-mq.c 	hctx = hlist_entry_safe(node, struct blk_mq_hw_ctx, cpuhp_dead);
node             2352 block/blk-mq.c 		int node)
node             2357 block/blk-mq.c 	hctx = kzalloc_node(blk_mq_hw_ctx_size(set), gfp, node);
node             2361 block/blk-mq.c 	if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node))
node             2365 block/blk-mq.c 	if (node == NUMA_NO_NODE)
node             2366 block/blk-mq.c 		node = set->numa_node;
node             2367 block/blk-mq.c 	hctx->numa_node = node;
node             2382 block/blk-mq.c 			gfp, node);
node             2387 block/blk-mq.c 				gfp, node))
node             2760 block/blk-mq.c 		int hctx_idx, int node)
node             2767 block/blk-mq.c 		if (tmp->numa_node == node) {
node             2777 block/blk-mq.c 		hctx = blk_mq_alloc_hctx(q, set, node);
node             2801 block/blk-mq.c 		int node;
node             2804 block/blk-mq.c 		node = blk_mq_hw_queue_to_node(&set->map[HCTX_TYPE_DEFAULT], i);
node             2810 block/blk-mq.c 		if (hctxs[i] && (hctxs[i]->numa_node == node))
node             2813 block/blk-mq.c 		hctx = blk_mq_alloc_and_init_hctx(set, q, i, node);
node             2822 block/blk-mq.c 						node, hctxs[i]->numa_node);
node             3207 block/blk-mq.c 	struct list_head node;
node             3228 block/blk-mq.c 	INIT_LIST_HEAD(&qe->node);
node             3231 block/blk-mq.c 	list_add(&qe->node, head);
node             3254 block/blk-mq.c 	list_for_each_entry(qe, head, node)
node             3263 block/blk-mq.c 	list_del(&qe->node);
node               68 block/blk-throttle.c 	struct list_head	node;		/* service_queue->queued[] */
node               98 block/blk-throttle.c #define rb_entry_tg(node)	rb_entry((node), struct throtl_grp, rb_node)
node              391 block/blk-throttle.c 	INIT_LIST_HEAD(&qn->node);
node              410 block/blk-throttle.c 	if (list_empty(&qn->node)) {
node              411 block/blk-throttle.c 		list_add_tail(&qn->node, queued);
node              422 block/blk-throttle.c 	struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node);
node              450 block/blk-throttle.c 	struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node);
node              460 block/blk-throttle.c 		list_del_init(&qn->node);
node              466 block/blk-throttle.c 		list_move_tail(&qn->node, queued);
node              488 block/blk-throttle.c 	tg = kzalloc_node(sizeof(*tg), gfp, q->node);
node              654 block/blk-throttle.c 	struct rb_node **node = &parent_sq->pending_tree.rb_root.rb_node;
node              660 block/blk-throttle.c 	while (*node != NULL) {
node              661 block/blk-throttle.c 		parent = *node;
node              665 block/blk-throttle.c 			node = &parent->rb_left;
node              667 block/blk-throttle.c 			node = &parent->rb_right;
node              672 block/blk-throttle.c 	rb_link_node(&tg->rb_node, parent, node);
node             2376 block/blk-throttle.c 	td = kzalloc_node(sizeof(*td), GFP_KERNEL, q->node);
node              404 block/blk-zoned.c static inline unsigned long *blk_alloc_zone_bitmap(int node,
node              408 block/blk-zoned.c 			    GFP_NOIO, node);
node              486 block/blk-zoned.c 	seq_zones_wlock = blk_alloc_zone_bitmap(q->node, nr_zones);
node              489 block/blk-zoned.c 	seq_zones_bitmap = blk_alloc_zone_bitmap(q->node, nr_zones);
node               59 block/blk.h    		int node, int cmd_size, gfp_t flags);
node              286 block/blk.h    int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node);
node              300 block/blk.h    static inline struct io_context *create_io_context(gfp_t gfp_mask, int node)
node              304 block/blk.h    		create_task_io_context(current, gfp_mask, node);
node              169 block/elevator.c 	eq = kzalloc_node(sizeof(*eq), GFP_KERNEL, q->node);
node             1622 block/genhd.c  	struct list_head	node;		/* all disk_event's */
node             1995 block/genhd.c  	list_for_each_entry(ev, &disk_events, node)
node             2030 block/genhd.c  	INIT_LIST_HEAD(&ev->node);
node             2052 block/genhd.c  	list_add_tail(&disk->ev->node, &disk_events);
node             2068 block/genhd.c  		list_del_init(&disk->ev->node);
node              372 block/kyber-iosched.c 	kqd = kzalloc_node(sizeof(*kqd), GFP_KERNEL, q->node);
node              390 block/kyber-iosched.c 					      GFP_KERNEL, q->node);
node               79 block/mq-deadline.c 	struct rb_node *node = rb_next(&rq->rb_node);
node               81 block/mq-deadline.c 	if (node)
node               82 block/mq-deadline.c 		return rb_entry_rq(node);
node              415 block/mq-deadline.c 	dd = kzalloc_node(sizeof(*dd), GFP_KERNEL, q->node);
node              218 block/sed-opal.c 	struct list_head node;
node             1073 block/sed-opal.c 	list_for_each_entry(iter, &dev->unlk_lst, node) {
node             1075 block/sed-opal.c 			list_del(&iter->node);
node             1080 block/sed-opal.c 	list_add_tail(&sus->node, &dev->unlk_lst);
node             2009 block/sed-opal.c 	list_for_each_entry_safe(suspend, next, &dev->unlk_lst, node) {
node             2010 block/sed-opal.c 		list_del(&suspend->node);
node             2422 block/sed-opal.c 	list_for_each_entry(suspend, &dev->unlk_lst, node) {
node               42 crypto/af_alg.c 	struct alg_type_list *node;
node               45 crypto/af_alg.c 	list_for_each_entry(node, &alg_types, list) {
node               46 crypto/af_alg.c 		if (strcmp(node->type->name, name))
node               49 crypto/af_alg.c 		if (try_module_get(node->type->owner))
node               50 crypto/af_alg.c 			type = node->type;
node               60 crypto/af_alg.c 	struct alg_type_list *node;
node               64 crypto/af_alg.c 	list_for_each_entry(node, &alg_types, list) {
node               65 crypto/af_alg.c 		if (!strcmp(node->type->name, type->name))
node               69 crypto/af_alg.c 	node = kmalloc(sizeof(*node), GFP_KERNEL);
node               71 crypto/af_alg.c 	if (!node)
node               77 crypto/af_alg.c 	node->type = type;
node               78 crypto/af_alg.c 	list_add(&node->list, &alg_types);
node               90 crypto/af_alg.c 	struct alg_type_list *node;
node               94 crypto/af_alg.c 	list_for_each_entry(node, &alg_types, list) {
node               95 crypto/af_alg.c 		if (strcmp(node->type->name, type->name))
node               98 crypto/af_alg.c 		list_del(&node->list);
node               99 crypto/af_alg.c 		kfree(node);
node               74 drivers/acpi/acpi_amba.c 	list_for_each_entry(rentry, &resource_list, node) {
node              101 drivers/acpi/acpi_apd.c 	list_for_each_entry(rentry, &resource_list, node) {
node              657 drivers/acpi/acpi_lpss.c 	list_for_each_entry(rentry, &resource_list, node)
node              187 drivers/acpi/acpi_memhotplug.c 	int node;
node              189 drivers/acpi/acpi_memhotplug.c 	node = acpi_get_node(handle);
node              207 drivers/acpi/acpi_memhotplug.c 		if (node < 0)
node              208 drivers/acpi/acpi_memhotplug.c 			node = memory_add_physaddr_to_nid(info->start_addr);
node              210 drivers/acpi/acpi_memhotplug.c 		result = __add_memory(node, info->start_addr, info->length);
node              129 drivers/acpi/acpi_platform.c 		list_for_each_entry(rentry, &resource_list, node)
node             1535 drivers/acpi/acpi_video.c 	list_for_each_entry(dev, &device->children, node) {
node              142 drivers/acpi/acpica/acdispat.h 			      union acpi_operand_object ***node);
node              163 drivers/acpi/acpica/acdispat.h 			     struct acpi_namespace_node **node);
node              171 drivers/acpi/acpica/acdispat.h acpi_ds_auto_serialize_method(struct acpi_namespace_node *node,
node              229 drivers/acpi/acpica/acdispat.h 		    struct acpi_namespace_node *node,
node              277 drivers/acpi/acpica/acdispat.h acpi_ds_scope_stack_push(struct acpi_namespace_node *node,
node               39 drivers/acpi/acpica/acevents.h u8 acpi_ev_is_notify_object(struct acpi_namespace_node *node);
node               44 drivers/acpi/acpica/acevents.h acpi_ev_queue_notify_request(struct acpi_namespace_node *node,
node              159 drivers/acpi/acpica/acevents.h acpi_ev_has_default_handler(struct acpi_namespace_node *node,
node              165 drivers/acpi/acpica/acevents.h acpi_ev_install_space_handler(struct acpi_namespace_node *node,
node              191 drivers/acpi/acpica/acevents.h acpi_ev_execute_reg_methods(struct acpi_namespace_node *node,
node              233 drivers/acpi/acpica/acevents.h u8 acpi_ev_is_pci_root_bridge(struct acpi_namespace_node *node);
node              393 drivers/acpi/acpica/acinterp.h void acpi_ex_dump_namespace_node(struct acpi_namespace_node *node, u32 flags);
node              413 drivers/acpi/acpica/acinterp.h 			     struct acpi_namespace_node *node,
node              455 drivers/acpi/acpica/acinterp.h 				 struct acpi_namespace_node *node);
node              207 drivers/acpi/acpica/aclocal.h 	struct acpi_namespace_node *node;
node              474 drivers/acpi/acpica/aclocal.h 	struct acpi_namespace_node *node;
node              598 drivers/acpi/acpica/aclocal.h 	ACPI_STATE_COMMON struct acpi_namespace_node *node;
node              649 drivers/acpi/acpica/aclocal.h 	struct acpi_namespace_node *node;
node              737 drivers/acpi/acpica/aclocal.h 	struct acpi_namespace_node      *node;              /* For use by interpreter */\
node               71 drivers/acpi/acpica/acnamesp.h acpi_ns_load_table(u32 table_index, struct acpi_namespace_node *node);
node              131 drivers/acpi/acpica/acnamesp.h void acpi_ns_delete_node(struct acpi_namespace_node *node);
node              133 drivers/acpi/acpica/acnamesp.h void acpi_ns_remove_node(struct acpi_namespace_node *node);
node              140 drivers/acpi/acpica/acnamesp.h void acpi_ns_detach_object(struct acpi_namespace_node *node);
node              215 drivers/acpi/acpica/acnamesp.h 			     struct acpi_namespace_node *node,
node              221 drivers/acpi/acpica/acnamesp.h 			      struct acpi_namespace_node *node,
node              230 drivers/acpi/acpica/acnamesp.h acpi_ns_check_return_value(struct acpi_namespace_node *node,
node              253 drivers/acpi/acpica/acnamesp.h char *acpi_ns_get_external_pathname(struct acpi_namespace_node *node);
node              256 drivers/acpi/acpica/acnamesp.h acpi_ns_build_normalized_path(struct acpi_namespace_node *node,
node              261 drivers/acpi/acpica/acnamesp.h char *acpi_ns_get_normalized_pathname(struct acpi_namespace_node *node,
node              289 drivers/acpi/acpica/acnamesp.h acpi_size acpi_ns_get_pathname_length(struct acpi_namespace_node *node);
node              295 drivers/acpi/acpica/acnamesp.h acpi_ns_attach_object(struct acpi_namespace_node *node,
node              300 drivers/acpi/acpica/acnamesp.h 						       *node);
node              307 drivers/acpi/acpica/acnamesp.h acpi_ns_attach_data(struct acpi_namespace_node *node,
node              311 drivers/acpi/acpica/acnamesp.h acpi_ns_detach_data(struct acpi_namespace_node *node,
node              315 drivers/acpi/acpica/acnamesp.h acpi_ns_get_attached_data(struct acpi_namespace_node *node,
node              350 drivers/acpi/acpica/acnamesp.h 			struct acpi_namespace_node *node,
node              360 drivers/acpi/acpica/acnamesp.h 			 struct acpi_namespace_node *node,
node              367 drivers/acpi/acpica/acnamesp.h 			 struct acpi_namespace_node *node,
node              374 drivers/acpi/acpica/acnamesp.h 		     struct acpi_namespace_node *node, acpi_object_type type);
node              379 drivers/acpi/acpica/acnamesp.h acpi_object_type acpi_ns_get_type(struct acpi_namespace_node *node);
node              384 drivers/acpi/acpica/acnamesp.h acpi_ns_print_node_pathname(struct acpi_namespace_node *node, const char *msg);
node              101 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node *node;	/* Link back to parent node */
node              105 drivers/acpi/acpica/acobject.h 	ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node;	/* Link back to parent node */
node              130 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node *node;	/* Containing namespace node */
node              136 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node *node;	/* Containing namespace node */
node              148 drivers/acpi/acpica/acobject.h 	union acpi_operand_object *node;
node              223 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node      *node;              /* Link back to parent node */\
node              273 drivers/acpi/acpica/acobject.h 	ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node;	/* Parent device */
node              284 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node *node;	/* Parent device */
node              311 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node *node;	/* ref_of or Namepath */
node              398 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node node;
node              434 drivers/acpi/acpica/acobject.h 	struct acpi_namespace_node node;
node              168 drivers/acpi/acpica/acresrc.h acpi_rs_get_prt_method_data(struct acpi_namespace_node *node,
node              172 drivers/acpi/acpica/acresrc.h acpi_rs_get_crs_method_data(struct acpi_namespace_node *node,
node              176 drivers/acpi/acpica/acresrc.h acpi_rs_get_prs_method_data(struct acpi_namespace_node *node,
node              184 drivers/acpi/acpica/acresrc.h acpi_rs_set_srs_method_data(struct acpi_namespace_node *node,
node              188 drivers/acpi/acpica/acresrc.h acpi_rs_get_aei_method_data(struct acpi_namespace_node *node,
node              159 drivers/acpi/acpica/acstruct.h 	struct acpi_namespace_node *node;	/* Resolved node (prefix_node:relative_pathname) */
node              721 drivers/acpi/acpica/acutils.h 		     struct acpi_namespace_node *node,
node               27 drivers/acpi/acpica/dbcmds.c acpi_dm_test_resource_conversion(struct acpi_namespace_node *node, char *name);
node               55 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node               63 drivers/acpi/acpica/dbcmds.c 		node = ACPI_TO_POINTER(address);
node               64 drivers/acpi/acpica/dbcmds.c 		if (!acpi_os_readable(node, sizeof(struct acpi_namespace_node))) {
node               65 drivers/acpi/acpica/dbcmds.c 			acpi_os_printf("Address %p is invalid", node);
node               71 drivers/acpi/acpica/dbcmds.c 		if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) {
node               74 drivers/acpi/acpica/dbcmds.c 			     node, acpi_ut_get_descriptor_name(node));
node               82 drivers/acpi/acpica/dbcmds.c 		node = acpi_db_local_ns_lookup(in_string);
node               83 drivers/acpi/acpica/dbcmds.c 		if (!node) {
node               87 drivers/acpi/acpica/dbcmds.c 			node = acpi_gbl_root_node;
node               91 drivers/acpi/acpica/dbcmds.c 	return (node);
node              337 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node              342 drivers/acpi/acpica/dbcmds.c 	node = acpi_db_convert_to_node(object_name);
node              343 drivers/acpi/acpica/dbcmds.c 	if (!node) {
node              347 drivers/acpi/acpica/dbcmds.c 	status = acpi_unload_parent_table(ACPI_CAST_PTR(acpi_handle, node));
node              350 drivers/acpi/acpica/dbcmds.c 			       object_name, node);
node              373 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node              378 drivers/acpi/acpica/dbcmds.c 	node = acpi_db_convert_to_node(name);
node              379 drivers/acpi/acpica/dbcmds.c 	if (!node) {
node              385 drivers/acpi/acpica/dbcmds.c 	if (acpi_ev_is_notify_object(node)) {
node              386 drivers/acpi/acpica/dbcmds.c 		status = acpi_ev_queue_notify_request(node, value);
node              393 drivers/acpi/acpica/dbcmds.c 			       acpi_ut_get_node_name(node),
node              394 drivers/acpi/acpica/dbcmds.c 			       acpi_ut_get_type_name(node->type));
node              493 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node              499 drivers/acpi/acpica/dbcmds.c 	node = acpi_db_convert_to_node(buffer_arg);
node              500 drivers/acpi/acpica/dbcmds.c 	if (!node || (node == acpi_gbl_root_node)) {
node              507 drivers/acpi/acpica/dbcmds.c 	if (node->type != ACPI_TYPE_BUFFER) {
node              519 drivers/acpi/acpica/dbcmds.c 	status = acpi_rs_create_resource_list(node->object, &return_buffer);
node              538 drivers/acpi/acpica/dbcmds.c 	acpi_ut_debug_dump_buffer((u8 *)node->object->buffer.pointer,
node              539 drivers/acpi/acpica/dbcmds.c 				  node->object->buffer.length,
node              659 drivers/acpi/acpica/dbcmds.c acpi_dm_test_resource_conversion(struct acpi_namespace_node *node, char *name)
node              675 drivers/acpi/acpica/dbcmds.c 	status = acpi_evaluate_object(node, name, NULL, &return_buffer);
node              684 drivers/acpi/acpica/dbcmds.c 	status = acpi_get_current_resources(node, &resource_buffer);
node              755 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node              764 drivers/acpi/acpica/dbcmds.c 	node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle);
node              765 drivers/acpi/acpica/dbcmds.c 	parent_path = acpi_ns_get_normalized_pathname(node, TRUE);
node              772 drivers/acpi/acpica/dbcmds.c 	(void)acpi_get_handle(node, METHOD_NAME__PRT,
node              774 drivers/acpi/acpica/dbcmds.c 	(void)acpi_get_handle(node, METHOD_NAME__CRS,
node              776 drivers/acpi/acpica/dbcmds.c 	(void)acpi_get_handle(node, METHOD_NAME__PRS,
node              778 drivers/acpi/acpica/dbcmds.c 	(void)acpi_get_handle(node, METHOD_NAME__AEI,
node              808 drivers/acpi/acpica/dbcmds.c 		status = acpi_get_irq_routing_table(node, &return_buffer);
node              837 drivers/acpi/acpica/dbcmds.c 		status = acpi_walk_resources(node, METHOD_NAME__CRS,
node              850 drivers/acpi/acpica/dbcmds.c 		status = acpi_get_current_resources(node, &return_buffer);
node              879 drivers/acpi/acpica/dbcmds.c 		(void)acpi_dm_test_resource_conversion(node, METHOD_NAME__CRS);
node              885 drivers/acpi/acpica/dbcmds.c 		status = acpi_set_current_resources(node, &return_buffer);
node              916 drivers/acpi/acpica/dbcmds.c 		status = acpi_get_possible_resources(node, &return_buffer);
node              948 drivers/acpi/acpica/dbcmds.c 		status = acpi_get_event_resources(node, &return_buffer);
node              981 drivers/acpi/acpica/dbcmds.c 	struct acpi_namespace_node *node;
node              996 drivers/acpi/acpica/dbcmds.c 		node = acpi_db_convert_to_node(object_arg);
node              997 drivers/acpi/acpica/dbcmds.c 		if (node) {
node              998 drivers/acpi/acpica/dbcmds.c 			if (node->type != ACPI_TYPE_DEVICE) {
node             1001 drivers/acpi/acpica/dbcmds.c 				     node->name.ascii,
node             1002 drivers/acpi/acpica/dbcmds.c 				     acpi_ut_get_type_name(node->type));
node             1004 drivers/acpi/acpica/dbcmds.c 				(void)acpi_db_device_resources(node, 0, NULL,
node              142 drivers/acpi/acpica/dbdisply.c 	struct acpi_namespace_node *node;
node              196 drivers/acpi/acpica/dbdisply.c 			node = obj_ptr;
node              263 drivers/acpi/acpica/dbdisply.c 	node = acpi_db_local_ns_lookup(target);
node              264 drivers/acpi/acpica/dbdisply.c 	if (!node) {
node              271 drivers/acpi/acpica/dbdisply.c 	status = acpi_get_name(node, ACPI_FULL_PATHNAME_NO_TRAILING, &ret_buf);
node              278 drivers/acpi/acpica/dbdisply.c 			       node, (char *)ret_buf.pointer);
node              281 drivers/acpi/acpica/dbdisply.c 	if (!acpi_os_readable(node, sizeof(struct acpi_namespace_node))) {
node              282 drivers/acpi/acpica/dbdisply.c 		acpi_os_printf("Invalid Named object at address %p\n", node);
node              286 drivers/acpi/acpica/dbdisply.c 	acpi_ut_debug_dump_buffer((void *)node,
node              289 drivers/acpi/acpica/dbdisply.c 	acpi_ex_dump_namespace_node(node, 1);
node              291 drivers/acpi/acpica/dbdisply.c 	obj_desc = acpi_ns_get_attached_object(node);
node              352 drivers/acpi/acpica/dbdisply.c 	struct acpi_namespace_node *node;
node              371 drivers/acpi/acpica/dbdisply.c 	node = walk_state->method_node;
node              374 drivers/acpi/acpica/dbdisply.c 		       acpi_ut_get_node_name(node));
node              506 drivers/acpi/acpica/dbdisply.c 	struct acpi_namespace_node *node;
node              517 drivers/acpi/acpica/dbdisply.c 	node = walk_state->method_node;
node              524 drivers/acpi/acpica/dbdisply.c 		       acpi_ut_get_node_name(node), result_count);
node              560 drivers/acpi/acpica/dbdisply.c 	struct acpi_namespace_node *node;
node              568 drivers/acpi/acpica/dbdisply.c 	node = walk_state->method_node;
node              572 drivers/acpi/acpica/dbdisply.c 		node = walk_state->method_node;
node              573 drivers/acpi/acpica/dbdisply.c 		acpi_os_printf("  [%4.4s]\n", acpi_ut_get_node_name(node));
node              751 drivers/acpi/acpica/dbdisply.c 			status = acpi_get_name(gpe_block->node,
node              759 drivers/acpi/acpica/dbdisply.c 			if (gpe_block->node == acpi_gbl_fadt_gpe_device) {
node              767 drivers/acpi/acpica/dbdisply.c 			     block, gpe_block, gpe_block->node, buffer,
node             1059 drivers/acpi/acpica/dbdisply.c 	struct acpi_namespace_node *node =
node             1065 drivers/acpi/acpica/dbdisply.c 	obj_desc = acpi_ns_get_attached_object(node);
node             1070 drivers/acpi/acpica/dbdisply.c 	pathname = acpi_ns_get_normalized_pathname(node, TRUE);
node             1089 drivers/acpi/acpica/dbdisply.c 		acpi_os_printf(" Device Name: %s (%p)\n", pathname, node);
node              295 drivers/acpi/acpica/dbexec.c 	struct acpi_namespace_node *node =
node              300 drivers/acpi/acpica/dbexec.c 	obj_desc = acpi_ns_get_attached_object(node);
node              308 drivers/acpi/acpica/dbexec.c 	acpi_ns_print_node_pathname(node, "Evaluating");
node              315 drivers/acpi/acpica/dbexec.c 	status = acpi_evaluate_object(node, NULL, NULL, &return_obj);
node              318 drivers/acpi/acpica/dbexec.c 		       acpi_ut_get_node_name(node),
node              115 drivers/acpi/acpica/dbmethod.c 	struct acpi_namespace_node *node;
node              129 drivers/acpi/acpica/dbmethod.c 		node = acpi_db_convert_to_node(index_arg);
node              130 drivers/acpi/acpica/dbmethod.c 		if (!node) {
node              134 drivers/acpi/acpica/dbmethod.c 		if (node->type != ACPI_TYPE_INTEGER) {
node              138 drivers/acpi/acpica/dbmethod.c 		obj_desc = node->object;
node              360 drivers/acpi/acpica/dbmethod.c 	struct acpi_namespace_node *node =
node              373 drivers/acpi/acpica/dbmethod.c 	predefined = acpi_ut_match_predefined_method(node->name.ascii);
node              378 drivers/acpi/acpica/dbmethod.c 	if (node->type == ACPI_TYPE_LOCAL_SCOPE) {
node              382 drivers/acpi/acpica/dbmethod.c 	pathname = acpi_ns_get_external_pathname(node);
node              419 drivers/acpi/acpica/dbmethod.c 	status = acpi_evaluate_object(node, NULL, &param_objects, &return_obj);
node              103 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node;
node              117 drivers/acpi/acpica/dbnames.c 					  ACPI_NS_NO_UPSEARCH, &node);
node              127 drivers/acpi/acpica/dbnames.c 					  ACPI_NS_NO_UPSEARCH, &node);
node              147 drivers/acpi/acpica/dbnames.c 	acpi_gbl_db_scope_node = node;
node              399 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node =
node              407 drivers/acpi/acpica/dbnames.c 	predefined = acpi_ut_match_predefined_method(node->name.ascii);
node              412 drivers/acpi/acpica/dbnames.c 	pathname = acpi_ns_get_normalized_pathname(node, TRUE);
node              441 drivers/acpi/acpica/dbnames.c 	acpi_ns_check_acpi_compliance(pathname, node, predefined);
node              492 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node =
node              495 drivers/acpi/acpica/dbnames.c 	if (node->type > ACPI_TYPE_NS_NODE_MAX) {
node              497 drivers/acpi/acpica/dbnames.c 			       node->name.ascii, node->type);
node              499 drivers/acpi/acpica/dbnames.c 		info->types[node->type]++;
node              649 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node =
node              659 drivers/acpi/acpica/dbnames.c 		if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) {
node              662 drivers/acpi/acpica/dbnames.c 			     "is %2.2X should be %2.2X\n", node,
node              663 drivers/acpi/acpica/dbnames.c 			     acpi_ut_get_descriptor_name(node),
node              664 drivers/acpi/acpica/dbnames.c 			     ACPI_GET_DESCRIPTOR_TYPE(node),
node              669 drivers/acpi/acpica/dbnames.c 		if ((node->type == ACPI_TYPE_LOCAL_ALIAS) ||
node              670 drivers/acpi/acpica/dbnames.c 		    (node->type == ACPI_TYPE_LOCAL_METHOD_ALIAS)) {
node              671 drivers/acpi/acpica/dbnames.c 			node = (struct acpi_namespace_node *)node->object;
node              677 drivers/acpi/acpica/dbnames.c 	if (node->type > ACPI_TYPE_LOCAL_MAX) {
node              679 drivers/acpi/acpica/dbnames.c 			       node, node->type);
node              683 drivers/acpi/acpica/dbnames.c 	if (!acpi_ut_valid_nameseg(node->name.ascii)) {
node              684 drivers/acpi/acpica/dbnames.c 		acpi_os_printf("Invalid AcpiName for Node %p\n", node);
node              688 drivers/acpi/acpica/dbnames.c 	object = acpi_ns_get_attached_object(node);
node              749 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node =
node              754 drivers/acpi/acpica/dbnames.c 	if (node == (void *)obj_desc) {
node              756 drivers/acpi/acpica/dbnames.c 			       acpi_ut_get_node_name(node));
node              761 drivers/acpi/acpica/dbnames.c 	if (acpi_ns_get_attached_object(node) == obj_desc) {
node              763 drivers/acpi/acpica/dbnames.c 			       node, acpi_ut_get_node_name(node));
node              815 drivers/acpi/acpica/dbnames.c 	struct acpi_namespace_node *node =
node              823 drivers/acpi/acpica/dbnames.c 	if ((node->type != ACPI_TYPE_DEVICE) &&
node              824 drivers/acpi/acpica/dbnames.c 	    (node->type != ACPI_TYPE_PROCESSOR)) {
node              830 drivers/acpi/acpica/dbnames.c 	status = acpi_get_handle(node, METHOD_NAME__PRT,
node              853 drivers/acpi/acpica/dbnames.c 	acpi_os_printf("%-32s Type %X", (char *)buffer.pointer, node->type);
node               17 drivers/acpi/acpica/dbobject.c static void acpi_db_decode_node(struct acpi_namespace_node *node);
node               38 drivers/acpi/acpica/dbobject.c 	struct acpi_namespace_node *node;
node               40 drivers/acpi/acpica/dbobject.c 	node = walk_state->method_node;
node               44 drivers/acpi/acpica/dbobject.c 	if (node == acpi_gbl_root_node) {
node              156 drivers/acpi/acpica/dbobject.c static void acpi_db_decode_node(struct acpi_namespace_node *node)
node              160 drivers/acpi/acpica/dbobject.c 		       acpi_ut_get_node_name(node));
node              162 drivers/acpi/acpica/dbobject.c 	if (node->flags & ANOBJ_METHOD_ARG) {
node              165 drivers/acpi/acpica/dbobject.c 	if (node->flags & ANOBJ_METHOD_LOCAL) {
node              169 drivers/acpi/acpica/dbobject.c 	switch (node->type) {
node              186 drivers/acpi/acpica/dbobject.c 					       (node));
node              342 drivers/acpi/acpica/dbobject.c 				acpi_db_decode_node(obj_desc->reference.node);
node              393 drivers/acpi/acpica/dbobject.c 	struct acpi_namespace_node *node;
node              396 drivers/acpi/acpica/dbobject.c 	node = walk_state->method_node;
node              400 drivers/acpi/acpica/dbobject.c 	if (node == acpi_gbl_root_node) {
node              404 drivers/acpi/acpica/dbobject.c 	if (!node) {
node              410 drivers/acpi/acpica/dbobject.c 	if (node->type != ACPI_TYPE_METHOD) {
node              430 drivers/acpi/acpica/dbobject.c 		     acpi_ut_get_node_name(node));
node              443 drivers/acpi/acpica/dbobject.c 		     acpi_ut_get_node_name(node));
node              463 drivers/acpi/acpica/dbobject.c 	struct acpi_namespace_node *node;
node              466 drivers/acpi/acpica/dbobject.c 	node = walk_state->method_node;
node              471 drivers/acpi/acpica/dbobject.c 	if (node == acpi_gbl_root_node) {
node              475 drivers/acpi/acpica/dbobject.c 	if (!node) {
node              481 drivers/acpi/acpica/dbobject.c 	if (node->type != ACPI_TYPE_METHOD) {
node              501 drivers/acpi/acpica/dbobject.c 			       acpi_ut_get_node_name(node),
node              502 drivers/acpi/acpica/dbobject.c 			       node->object->method.param_count);
node              515 drivers/acpi/acpica/dbobject.c 		     acpi_ut_get_node_name(node));
node              218 drivers/acpi/acpica/dbstats.c 	struct acpi_namespace_node *node;
node              224 drivers/acpi/acpica/dbstats.c 	node = (struct acpi_namespace_node *)obj_handle;
node              225 drivers/acpi/acpica/dbstats.c 	obj_desc = acpi_ns_get_attached_object(node);
node              229 drivers/acpi/acpica/dbstats.c 	type = node->type;
node               26 drivers/acpi/acpica/dbtest.c acpi_db_test_integer_type(struct acpi_namespace_node *node, u32 bit_length);
node               29 drivers/acpi/acpica/dbtest.c acpi_db_test_buffer_type(struct acpi_namespace_node *node, u32 bit_length);
node               32 drivers/acpi/acpica/dbtest.c acpi_db_test_string_type(struct acpi_namespace_node *node, u32 byte_length);
node               34 drivers/acpi/acpica/dbtest.c static acpi_status acpi_db_test_package_type(struct acpi_namespace_node *node);
node               40 drivers/acpi/acpica/dbtest.c acpi_db_read_from_object(struct acpi_namespace_node *node,
node               45 drivers/acpi/acpica/dbtest.c acpi_db_write_to_object(struct acpi_namespace_node *node,
node              243 drivers/acpi/acpica/dbtest.c 	struct acpi_namespace_node *node;
node              250 drivers/acpi/acpica/dbtest.c 	node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle);
node              251 drivers/acpi/acpica/dbtest.c 	obj_desc = node->object;
node              257 drivers/acpi/acpica/dbtest.c 	switch (node->type) {
node              318 drivers/acpi/acpica/dbtest.c 		       acpi_ut_get_type_name(node->type), node->name.ascii);
node              330 drivers/acpi/acpica/dbtest.c 		status = acpi_db_test_integer_type(node, bit_length);
node              335 drivers/acpi/acpica/dbtest.c 		status = acpi_db_test_string_type(node, byte_length);
node              340 drivers/acpi/acpica/dbtest.c 		status = acpi_db_test_buffer_type(node, bit_length);
node              345 drivers/acpi/acpica/dbtest.c 		status = acpi_db_test_package_type(node);
node              388 drivers/acpi/acpica/dbtest.c acpi_db_test_integer_type(struct acpi_namespace_node *node, u32 bit_length)
node              405 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp1);
node              422 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              429 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp2);
node              443 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              450 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_INTEGER, &temp3);
node              490 drivers/acpi/acpica/dbtest.c acpi_db_test_buffer_type(struct acpi_namespace_node *node, u32 bit_length)
node              517 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp1);
node              549 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              556 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp2);
node              570 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              577 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_BUFFER, &temp3);
node              616 drivers/acpi/acpica/dbtest.c acpi_db_test_string_type(struct acpi_namespace_node *node, u32 byte_length)
node              627 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp1);
node              642 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              649 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp2);
node              664 drivers/acpi/acpica/dbtest.c 	status = acpi_db_write_to_object(node, &write_value);
node              671 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_STRING, &temp3);
node              706 drivers/acpi/acpica/dbtest.c static acpi_status acpi_db_test_package_type(struct acpi_namespace_node *node)
node              713 drivers/acpi/acpica/dbtest.c 	status = acpi_db_read_from_object(node, ACPI_TYPE_PACKAGE, &temp1);
node              785 drivers/acpi/acpica/dbtest.c 		     region_obj->region.node->name.ascii);
node              808 drivers/acpi/acpica/dbtest.c acpi_db_read_from_object(struct acpi_namespace_node *node,
node              819 drivers/acpi/acpica/dbtest.c 	params[0].reference.actual_type = node->type;
node              820 drivers/acpi/acpica/dbtest.c 	params[0].reference.handle = ACPI_CAST_PTR(acpi_handle, node);
node              892 drivers/acpi/acpica/dbtest.c acpi_db_write_to_object(struct acpi_namespace_node *node,
node              900 drivers/acpi/acpica/dbtest.c 	params[0].reference.actual_type = node->type;
node              901 drivers/acpi/acpica/dbtest.c 	params[0].reference.handle = ACPI_CAST_PTR(acpi_handle, node);
node              975 drivers/acpi/acpica/dbtest.c 	struct acpi_namespace_node *node =
node              994 drivers/acpi/acpica/dbtest.c 	predefined = acpi_ut_match_predefined_method(node->name.ascii);
node              999 drivers/acpi/acpica/dbtest.c 	if (node->type == ACPI_TYPE_LOCAL_SCOPE) {
node             1003 drivers/acpi/acpica/dbtest.c 	pathname = acpi_ns_get_normalized_pathname(node, TRUE);
node             1085 drivers/acpi/acpica/dbtest.c 	status = acpi_evaluate_object(node, NULL, &param_objects, &return_obj);
node              248 drivers/acpi/acpica/dbutils.c 	struct acpi_namespace_node *node = NULL;
node              267 drivers/acpi/acpica/dbutils.c 				NULL, &node);
node              274 drivers/acpi/acpica/dbutils.c 	return (node);
node               23 drivers/acpi/acpica/dsargs.c acpi_ds_execute_arguments(struct acpi_namespace_node *node,
node               43 drivers/acpi/acpica/dsargs.c acpi_ds_execute_arguments(struct acpi_namespace_node *node,
node               62 drivers/acpi/acpica/dsargs.c 	op->common.node = scope_node;
node               82 drivers/acpi/acpica/dsargs.c 	walk_state->deferred_node = node;
node               93 drivers/acpi/acpica/dsargs.c 	op->common.node = node;
node              103 drivers/acpi/acpica/dsargs.c 	op->common.node = scope_node;
node              124 drivers/acpi/acpica/dsargs.c 	walk_state->deferred_node = node;
node              149 drivers/acpi/acpica/dsargs.c 	struct acpi_namespace_node *node;
node              161 drivers/acpi/acpica/dsargs.c 	node = obj_desc->buffer_field.node;
node              164 drivers/acpi/acpica/dsargs.c 			(ACPI_TYPE_BUFFER_FIELD, node, NULL));
node              167 drivers/acpi/acpica/dsargs.c 			  acpi_ut_get_node_name(node)));
node              171 drivers/acpi/acpica/dsargs.c 	status = acpi_ds_execute_arguments(node, node->parent,
node              194 drivers/acpi/acpica/dsargs.c 	struct acpi_namespace_node *node;
node              206 drivers/acpi/acpica/dsargs.c 	node = obj_desc->bank_field.node;
node              209 drivers/acpi/acpica/dsargs.c 			(ACPI_TYPE_LOCAL_BANK_FIELD, node, NULL));
node              212 drivers/acpi/acpica/dsargs.c 			  acpi_ut_get_node_name(node)));
node              216 drivers/acpi/acpica/dsargs.c 	status = acpi_ds_execute_arguments(node, node->parent,
node              225 drivers/acpi/acpica/dsargs.c 					   obj_desc->region.length, node);
node              244 drivers/acpi/acpica/dsargs.c 	struct acpi_namespace_node *node;
node              255 drivers/acpi/acpica/dsargs.c 	node = obj_desc->buffer.node;
node              256 drivers/acpi/acpica/dsargs.c 	if (!node) {
node              267 drivers/acpi/acpica/dsargs.c 	status = acpi_ds_execute_arguments(node, node,
node              288 drivers/acpi/acpica/dsargs.c 	struct acpi_namespace_node *node;
node              299 drivers/acpi/acpica/dsargs.c 	node = obj_desc->package.node;
node              300 drivers/acpi/acpica/dsargs.c 	if (!node) {
node              312 drivers/acpi/acpica/dsargs.c 	status = acpi_ds_execute_arguments(node, node,
node              334 drivers/acpi/acpica/dsargs.c 	struct acpi_namespace_node *node;
node              351 drivers/acpi/acpica/dsargs.c 	node = obj_desc->region.node;
node              354 drivers/acpi/acpica/dsargs.c 			(ACPI_TYPE_REGION, node, NULL));
node              358 drivers/acpi/acpica/dsargs.c 			  acpi_ut_get_node_name(node),
node              363 drivers/acpi/acpica/dsargs.c 	status = acpi_ds_execute_arguments(node, extra_desc->extra.scope_node,
node              372 drivers/acpi/acpica/dsargs.c 					   obj_desc->region.length, node);
node               25 drivers/acpi/acpica/dsdebug.c acpi_ds_print_node_pathname(struct acpi_namespace_node *node,
node               41 drivers/acpi/acpica/dsdebug.c acpi_ds_print_node_pathname(struct acpi_namespace_node *node,
node               49 drivers/acpi/acpica/dsdebug.c 	if (!node) {
node               58 drivers/acpi/acpica/dsdebug.c 	status = acpi_ns_handle_to_pathname(node, &buffer, TRUE);
node               66 drivers/acpi/acpica/dsdebug.c 				      (char *)buffer.pointer, node));
node              145 drivers/acpi/acpica/dsdebug.c 						  method_desc->method.node,
node               33 drivers/acpi/acpica/dsfield.c 			       struct acpi_namespace_node **node);
node               64 drivers/acpi/acpica/dsfield.c 			       struct acpi_namespace_node **node)
node               82 drivers/acpi/acpica/dsfield.c 				walk_state, node);
node               94 drivers/acpi/acpica/dsfield.c 	obj_desc->region.node = *node;
node               95 drivers/acpi/acpica/dsfield.c 	status = acpi_ns_attach_object(*node, obj_desc, ACPI_TYPE_REGION);
node              124 drivers/acpi/acpica/dsfield.c 	struct acpi_namespace_node *node;
node              151 drivers/acpi/acpica/dsfield.c 		node = walk_state->deferred_node;
node              180 drivers/acpi/acpica/dsfield.c 					walk_state, &node);
node              193 drivers/acpi/acpica/dsfield.c 	op->common.node = node;
node              200 drivers/acpi/acpica/dsfield.c 	obj_desc = acpi_ns_get_attached_object(node);
node              225 drivers/acpi/acpica/dsfield.c 	obj_desc->buffer_field.node = node;
node              229 drivers/acpi/acpica/dsfield.c 	status = acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_BUFFER_FIELD);
node              390 drivers/acpi/acpica/dsfield.c 				arg->common.node = info->field_node;
node              552 drivers/acpi/acpica/dsfield.c 	struct acpi_namespace_node *node;
node              629 drivers/acpi/acpica/dsfield.c 						walk_state, &node);
node              643 drivers/acpi/acpica/dsfield.c 			arg->common.node = node;
node               51 drivers/acpi/acpica/dsinit.c 	struct acpi_namespace_node *node =
node               62 drivers/acpi/acpica/dsinit.c 	if (node->owner_id != info->owner_id) {
node               94 drivers/acpi/acpica/dsinit.c 		obj_desc = acpi_ns_get_attached_object(node);
node              110 drivers/acpi/acpica/dsinit.c 			acpi_ds_auto_serialize_method(node, obj_desc);
node               55 drivers/acpi/acpica/dsmethod.c acpi_ds_auto_serialize_method(struct acpi_namespace_node *node,
node               62 drivers/acpi/acpica/dsmethod.c 	ACPI_FUNCTION_TRACE_PTR(ds_auto_serialize_method, node);
node               66 drivers/acpi/acpica/dsmethod.c 			  acpi_ut_get_node_name(node), node));
node               75 drivers/acpi/acpica/dsmethod.c 	acpi_ps_set_name(op, node->name.integer);
node               76 drivers/acpi/acpica/dsmethod.c 	op->common.node = node;
node               81 drivers/acpi/acpica/dsmethod.c 	    acpi_ds_create_walk_state(node->owner_id, NULL, NULL, NULL);
node               87 drivers/acpi/acpica/dsmethod.c 	status = acpi_ds_init_aml_walk(walk_state, op, node,
node              839 drivers/acpi/acpica/dsmethod.c 				  method.node, method_desc, walk_state);
node              216 drivers/acpi/acpica/dsmthdat.c 			     struct acpi_namespace_node **node)
node              235 drivers/acpi/acpica/dsmthdat.c 		*node = &walk_state->local_variables[index];
node              249 drivers/acpi/acpica/dsmthdat.c 		*node = &walk_state->arguments[index];
node              285 drivers/acpi/acpica/dsmthdat.c 	struct acpi_namespace_node *node;
node              296 drivers/acpi/acpica/dsmthdat.c 	status = acpi_ds_method_data_get_node(type, index, walk_state, &node);
node              311 drivers/acpi/acpica/dsmthdat.c 	node->object = object;
node              339 drivers/acpi/acpica/dsmthdat.c 	struct acpi_namespace_node *node;
node              353 drivers/acpi/acpica/dsmthdat.c 	status = acpi_ds_method_data_get_node(type, index, walk_state, &node);
node              360 drivers/acpi/acpica/dsmthdat.c 	object = node->object;
node              381 drivers/acpi/acpica/dsmthdat.c 			node->object = object;
node              392 drivers/acpi/acpica/dsmthdat.c 					    index, node));
node              443 drivers/acpi/acpica/dsmthdat.c 	struct acpi_namespace_node *node;
node              450 drivers/acpi/acpica/dsmthdat.c 	status = acpi_ds_method_data_get_node(type, index, walk_state, &node);
node              457 drivers/acpi/acpica/dsmthdat.c 	object = acpi_ns_get_attached_object(node);
node              464 drivers/acpi/acpica/dsmthdat.c 	node->object = NULL;
node              504 drivers/acpi/acpica/dsmthdat.c 	struct acpi_namespace_node *node;
node              520 drivers/acpi/acpica/dsmthdat.c 	status = acpi_ds_method_data_get_node(type, index, walk_state, &node);
node              525 drivers/acpi/acpica/dsmthdat.c 	current_obj_desc = acpi_ns_get_attached_object(node);
node              655 drivers/acpi/acpica/dsmthdat.c 	struct acpi_namespace_node *node;
node              662 drivers/acpi/acpica/dsmthdat.c 	status = acpi_ds_method_data_get_node(opcode, index, walk_state, &node);
node              669 drivers/acpi/acpica/dsmthdat.c 	object = acpi_ns_get_attached_object(node);
node               52 drivers/acpi/acpica/dsobject.c 		if (!op->common.node) {
node               78 drivers/acpi/acpica/dsobject.c 							 &(op->common.node)));
node              118 drivers/acpi/acpica/dsobject.c 		    !obj_desc->reference.node) {
node              123 drivers/acpi/acpica/dsobject.c 			obj_desc->reference.node =
node              124 drivers/acpi/acpica/dsobject.c 			    walk_state->scope_info->scope.node;
node              234 drivers/acpi/acpica/dsobject.c 	op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc);
node              254 drivers/acpi/acpica/dsobject.c 		    struct acpi_namespace_node *node,
node              267 drivers/acpi/acpica/dsobject.c 	if (acpi_ns_get_attached_object(node)) {
node              289 drivers/acpi/acpica/dsobject.c 	node->type = obj_desc->common.type;
node              293 drivers/acpi/acpica/dsobject.c 	status = acpi_ns_attach_object(node, obj_desc, node->type);
node              346 drivers/acpi/acpica/dsobject.c 		obj_desc->buffer.node =
node              361 drivers/acpi/acpica/dsobject.c 		obj_desc->package.node =
node              511 drivers/acpi/acpica/dsobject.c 				obj_desc->reference.node = op->common.node;
node              513 drivers/acpi/acpica/dsobject.c 				if (op->common.node) {
node              515 drivers/acpi/acpica/dsobject.c 					    op->common.node->object;
node              271 drivers/acpi/acpica/dsopcode.c 	struct acpi_namespace_node *node;
node              280 drivers/acpi/acpica/dsopcode.c 	node = op->common.node;
node              293 drivers/acpi/acpica/dsopcode.c 	obj_desc = acpi_ns_get_attached_object(node);
node              357 drivers/acpi/acpica/dsopcode.c 	struct acpi_namespace_node *node;
node              367 drivers/acpi/acpica/dsopcode.c 	node = op->common.node;
node              394 drivers/acpi/acpica/dsopcode.c 	obj_desc = acpi_ns_get_attached_object(node);
node              414 drivers/acpi/acpica/dsopcode.c 			      node->name.ascii, space_id));
node              434 drivers/acpi/acpica/dsopcode.c 					   obj_desc->region.length, node);
node              464 drivers/acpi/acpica/dsopcode.c 	struct acpi_namespace_node *node;
node              475 drivers/acpi/acpica/dsopcode.c 	node = op->common.node;
node              524 drivers/acpi/acpica/dsopcode.c 	obj_desc = acpi_ns_get_attached_object(node);
node              684 drivers/acpi/acpica/dsopcode.c 	struct acpi_namespace_node *node;
node              741 drivers/acpi/acpica/dsopcode.c 			node = arg->common.node;
node              743 drivers/acpi/acpica/dsopcode.c 			obj_desc = acpi_ns_get_attached_object(node);
node               98 drivers/acpi/acpica/dspkginit.c 		obj_desc->package.node = parent->common.node;
node              155 drivers/acpi/acpica/dspkginit.c 			if (!arg->common.node) {
node              181 drivers/acpi/acpica/dspkginit.c 			if (arg->common.node->type == ACPI_TYPE_METHOD) {
node              198 drivers/acpi/acpica/dspkginit.c 						  arg->common.node);
node              269 drivers/acpi/acpica/dspkginit.c 			if (arg->common.node) {
node              273 drivers/acpi/acpica/dspkginit.c 							  arg->common.node));
node              274 drivers/acpi/acpica/dspkginit.c 				arg->common.node = NULL;
node              307 drivers/acpi/acpica/dspkginit.c 	op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc);
node              406 drivers/acpi/acpica/dspkginit.c 	scope_info.scope.node = element->reference.node;	/* Prefix node */
node              457 drivers/acpi/acpica/dspkginit.c 			    scope_info.scope.node->name.ascii));
node              474 drivers/acpi/acpica/dspkginit.c 	element->reference.node = resolved_node;
node              475 drivers/acpi/acpica/dspkginit.c 	type = element->reference.node->type;
node              284 drivers/acpi/acpica/dswexec.c 				    acpi_ds_scope_stack_push(op->named.node,
node              285 drivers/acpi/acpica/dswexec.c 							     op->named.node->
node              478 drivers/acpi/acpica/dswexec.c 				op->common.node = (struct acpi_namespace_node *)
node              479 drivers/acpi/acpica/dswexec.c 				    op->asl.value.arg->asl.node;
node              481 drivers/acpi/acpica/dswexec.c 						      node->object);
node              556 drivers/acpi/acpica/dswexec.c 			    acpi_ns_get_external_pathname(op->common.node);
node              562 drivers/acpi/acpica/dswexec.c 								node->object,
node              590 drivers/acpi/acpica/dswexec.c 				    op->common.parent->common.node;
node              595 drivers/acpi/acpica/dswexec.c 							     common.node,
node              611 drivers/acpi/acpica/dswexec.c 								 node));
node              107 drivers/acpi/acpica/dswload.c 	struct acpi_namespace_node *node;
node              129 drivers/acpi/acpica/dswload.c 		if (op->common.node) {
node              155 drivers/acpi/acpica/dswload.c 				   walk_state, &(node));
node              169 drivers/acpi/acpica/dswload.c 					   &node);
node              182 drivers/acpi/acpica/dswload.c 		switch (node->type) {
node              209 drivers/acpi/acpica/dswload.c 					  acpi_ut_get_node_name(node),
node              210 drivers/acpi/acpica/dswload.c 					  acpi_ut_get_type_name(node->type)));
node              212 drivers/acpi/acpica/dswload.c 			node->type = ACPI_TYPE_ANY;
node              221 drivers/acpi/acpica/dswload.c 			if ((node == acpi_gbl_root_node) &&
node              236 drivers/acpi/acpica/dswload.c 				    acpi_ut_get_type_name(node->type),
node              237 drivers/acpi/acpica/dswload.c 				    acpi_ut_get_node_name(node)));
node              264 drivers/acpi/acpica/dswload.c 			node = walk_state->deferred_node;
node              274 drivers/acpi/acpica/dswload.c 			node = NULL;
node              310 drivers/acpi/acpica/dswload.c 				   &node);
node              316 drivers/acpi/acpica/dswload.c 				if (node->flags & ANOBJ_IS_EXTERNAL) {
node              321 drivers/acpi/acpica/dswload.c 					node->flags &= ~ANOBJ_IS_EXTERNAL;
node              322 drivers/acpi/acpica/dswload.c 					node->type = (u8) object_type;
node              329 drivers/acpi/acpica/dswload.c 						    (node, object_type,
node              368 drivers/acpi/acpica/dswload.c 	if (node) {
node              373 drivers/acpi/acpica/dswload.c 		op->common.node = node;
node              374 drivers/acpi/acpica/dswload.c 		op->named.name = node->name.integer;
node              499 drivers/acpi/acpica/dswload.c 			if (op->common.node) {
node              500 drivers/acpi/acpica/dswload.c 				op->common.node->type = (u8) object_type;
node              510 drivers/acpi/acpica/dswload.c 	    op->common.node && op->common.aml_opcode == AML_EXTERNAL_OP) {
node              519 drivers/acpi/acpica/dswload.c 		op->common.node->flags |= ANOBJ_IS_EXTERNAL;
node              520 drivers/acpi/acpica/dswload.c 		op->common.node->type = (u8)object_type;
node              523 drivers/acpi/acpica/dswload.c 						      &op->common.node,
node              553 drivers/acpi/acpica/dswload.c 					  walk_state, op, op->named.node));
node              555 drivers/acpi/acpica/dswload.c 			if (!acpi_ns_get_attached_object(op->named.node)) {
node              557 drivers/acpi/acpica/dswload.c 				    ACPI_CAST_PTR(void, op->named.node);
node               42 drivers/acpi/acpica/dswload2.c 	struct acpi_namespace_node *node;
node              111 drivers/acpi/acpica/dswload2.c 		node = NULL;
node              124 drivers/acpi/acpica/dswload2.c 				   ACPI_NS_SEARCH_PARENT, walk_state, &(node));
node              131 drivers/acpi/acpica/dswload2.c 		if (op && (op->named.node == acpi_gbl_root_node)) {
node              132 drivers/acpi/acpica/dswload2.c 			node = op->named.node;
node              135 drivers/acpi/acpica/dswload2.c 			    acpi_ds_scope_stack_push(node, object_type,
node              150 drivers/acpi/acpica/dswload2.c 					   &(node));
node              173 drivers/acpi/acpica/dswload2.c 		switch (node->type) {
node              198 drivers/acpi/acpica/dswload2.c 				      acpi_ut_get_node_name(node),
node              199 drivers/acpi/acpica/dswload2.c 				      acpi_ut_get_type_name(node->type)));
node              201 drivers/acpi/acpica/dswload2.c 			node->type = ACPI_TYPE_ANY;
node              211 drivers/acpi/acpica/dswload2.c 			if ((node == acpi_gbl_root_node) &&
node              226 drivers/acpi/acpica/dswload2.c 				    acpi_ut_get_type_name(node->type),
node              227 drivers/acpi/acpica/dswload2.c 				    acpi_ut_get_node_name(node)));
node              237 drivers/acpi/acpica/dswload2.c 		if (op && op->common.node) {
node              241 drivers/acpi/acpica/dswload2.c 			node = op->common.node;
node              245 drivers/acpi/acpica/dswload2.c 				    acpi_ds_scope_stack_push(node, object_type,
node              267 drivers/acpi/acpica/dswload2.c 			node = walk_state->deferred_node;
node              315 drivers/acpi/acpica/dswload2.c 				   walk_state, &node);
node              320 drivers/acpi/acpica/dswload2.c 					  acpi_ut_get_node_name(node), node));
node              342 drivers/acpi/acpica/dswload2.c 		if (node) {
node              343 drivers/acpi/acpica/dswload2.c 			op->named.name = node->name.integer;
node              352 drivers/acpi/acpica/dswload2.c 	op->common.node = node;
node              374 drivers/acpi/acpica/dswload2.c 	struct acpi_namespace_node *node;
node              408 drivers/acpi/acpica/dswload2.c 	node = op->common.node;
node              414 drivers/acpi/acpica/dswload2.c 	walk_state->operands[0] = (void *)node;
node              462 drivers/acpi/acpica/dswload2.c 			  walk_state, op, node));
node              497 drivers/acpi/acpica/dswload2.c 						       node, walk_state);
node              503 drivers/acpi/acpica/dswload2.c 			    acpi_ds_create_bank_field(op, arg->common.node,
node              510 drivers/acpi/acpica/dswload2.c 			    acpi_ds_create_field(op, arg->common.node,
node              613 drivers/acpi/acpica/dswload2.c 			    (acpi_ns_get_attached_object(node));
node              618 drivers/acpi/acpica/dswload2.c 			status = acpi_ds_create_node(walk_state, node, op);
node              627 drivers/acpi/acpica/dswload2.c 			namepath = acpi_ns_get_external_pathname(node);
node              633 drivers/acpi/acpica/dswload2.c 				if (node->object) {
node              634 drivers/acpi/acpica/dswload2.c 					acpi_ns_detach_object(node);
node              636 drivers/acpi/acpica/dswload2.c 				acpi_ns_attach_object(node, obj_desc,
node              655 drivers/acpi/acpica/dswload2.c 					  walk_state, op, op->named.node));
node              657 drivers/acpi/acpica/dswload2.c 			if (!acpi_ns_get_attached_object(op->named.node)) {
node              659 drivers/acpi/acpica/dswload2.c 				    ACPI_CAST_PTR(void, op->named.node);
node              700 drivers/acpi/acpica/dswload2.c 				  walk_state, op, node));
node              726 drivers/acpi/acpica/dswload2.c 			op->common.node = new_node;
node               67 drivers/acpi/acpica/dswscope.c acpi_ds_scope_stack_push(struct acpi_namespace_node *node,
node               76 drivers/acpi/acpica/dswscope.c 	if (!node) {
node              100 drivers/acpi/acpica/dswscope.c 	scope_info->scope.node = node;
node              114 drivers/acpi/acpica/dswscope.c 							    scope.node),
node              123 drivers/acpi/acpica/dswscope.c 			      acpi_ut_get_node_name(scope_info->scope.node),
node              164 drivers/acpi/acpica/dswscope.c 			  acpi_ut_get_node_name(scope_info->scope.node),
node              171 drivers/acpi/acpica/dswscope.c 							    scope.node),
node              630 drivers/acpi/acpica/dswstate.c 		while (extra_op && !extra_op->common.node) {
node              637 drivers/acpi/acpica/dswstate.c 			parser_state->start_node = extra_op->common.node;
node              378 drivers/acpi/acpica/evgpe.c 		gpe_device = gpe_block->node;
node              328 drivers/acpi/acpica/evgpeblk.c 	gpe_block->node = gpe_device;
node              230 drivers/acpi/acpica/evgpeinit.c 			walk_info.gpe_device = gpe_block->node;
node              103 drivers/acpi/acpica/evgpeutil.c 		if ((gpe_block->node)->type == ACPI_TYPE_DEVICE) {
node              104 drivers/acpi/acpica/evgpeutil.c 			info->gpe_device = gpe_block->node;
node              118 drivers/acpi/acpica/evhandler.c acpi_ev_has_default_handler(struct acpi_namespace_node *node,
node              126 drivers/acpi/acpica/evhandler.c 	obj_desc = acpi_ns_get_attached_object(node);
node              171 drivers/acpi/acpica/evhandler.c 	struct acpi_namespace_node *node;
node              186 drivers/acpi/acpica/evhandler.c 	node = acpi_ns_validate_handle(obj_handle);
node              187 drivers/acpi/acpica/evhandler.c 	if (!node) {
node              195 drivers/acpi/acpica/evhandler.c 	if ((node->type != ACPI_TYPE_DEVICE) &&
node              196 drivers/acpi/acpica/evhandler.c 	    (node->type != ACPI_TYPE_REGION) && (node != acpi_gbl_root_node)) {
node              202 drivers/acpi/acpica/evhandler.c 	obj_desc = acpi_ns_get_attached_object(node);
node              328 drivers/acpi/acpica/evhandler.c acpi_ev_install_space_handler(struct acpi_namespace_node *node,
node              345 drivers/acpi/acpica/evhandler.c 	if ((node->type != ACPI_TYPE_DEVICE) &&
node              346 drivers/acpi/acpica/evhandler.c 	    (node->type != ACPI_TYPE_PROCESSOR) &&
node              347 drivers/acpi/acpica/evhandler.c 	    (node->type != ACPI_TYPE_THERMAL) && (node != acpi_gbl_root_node)) {
node              407 drivers/acpi/acpica/evhandler.c 	obj_desc = acpi_ns_get_attached_object(node);
node              438 drivers/acpi/acpica/evhandler.c 				  node));
node              442 drivers/acpi/acpica/evhandler.c 		if (node->type == ACPI_TYPE_ANY) {
node              445 drivers/acpi/acpica/evhandler.c 			type = node->type;
node              460 drivers/acpi/acpica/evhandler.c 		status = acpi_ns_attach_object(node, obj_desc, type);
node              475 drivers/acpi/acpica/evhandler.c 			  acpi_ut_get_node_name(node), node, obj_desc));
node              495 drivers/acpi/acpica/evhandler.c 	handler_obj->address_space.node = node;
node              520 drivers/acpi/acpica/evhandler.c 	status = acpi_ns_walk_namespace(ACPI_TYPE_ANY, node,
node               35 drivers/acpi/acpica/evmisc.c u8 acpi_ev_is_notify_object(struct acpi_namespace_node *node)
node               38 drivers/acpi/acpica/evmisc.c 	switch (node->type) {
node               68 drivers/acpi/acpica/evmisc.c acpi_ev_queue_notify_request(struct acpi_namespace_node *node, u32 notify_value)
node               80 drivers/acpi/acpica/evmisc.c 	if (!acpi_ev_is_notify_object(node)) {
node               94 drivers/acpi/acpica/evmisc.c 	obj_desc = acpi_ns_get_attached_object(node);
node              111 drivers/acpi/acpica/evmisc.c 				  acpi_ut_get_node_name(node), notify_value,
node              112 drivers/acpi/acpica/evmisc.c 				  node));
node              126 drivers/acpi/acpica/evmisc.c 	info->notify.node = node;
node              134 drivers/acpi/acpica/evmisc.c 			  acpi_ut_get_node_name(node),
node              135 drivers/acpi/acpica/evmisc.c 			  acpi_ut_get_type_name(node->type), notify_value,
node              137 drivers/acpi/acpica/evmisc.c 			  node));
node              171 drivers/acpi/acpica/evmisc.c 		info->notify.global->handler(info->notify.node,
node              180 drivers/acpi/acpica/evmisc.c 		handler_obj->notify.handler(info->notify.node,
node              129 drivers/acpi/acpica/evregion.c 			    acpi_ut_get_node_name(region_obj->region.node),
node              488 drivers/acpi/acpica/evregion.c 			  acpi_ut_get_node_name(region_obj->region.node),
node              525 drivers/acpi/acpica/evregion.c 	struct acpi_namespace_node *node;
node              545 drivers/acpi/acpica/evregion.c 	node = region_obj->region.node->parent;
node              547 drivers/acpi/acpica/evregion.c 	    acpi_ns_search_one_scope(*reg_name_ptr, node, ACPI_TYPE_METHOD,
node              649 drivers/acpi/acpica/evregion.c acpi_ev_execute_reg_methods(struct acpi_namespace_node *node,
node              683 drivers/acpi/acpica/evregion.c 	(void)acpi_ns_walk_namespace(ACPI_TYPE_ANY, node, ACPI_UINT32_MAX,
node              690 drivers/acpi/acpica/evregion.c 		acpi_ev_orphan_ec_reg_method(node);
node              716 drivers/acpi/acpica/evregion.c 	struct acpi_namespace_node *node;
node              724 drivers/acpi/acpica/evregion.c 	node = acpi_ns_validate_handle(obj_handle);
node              725 drivers/acpi/acpica/evregion.c 	if (!node) {
node              733 drivers/acpi/acpica/evregion.c 	if ((node->type != ACPI_TYPE_REGION) && (node != acpi_gbl_root_node)) {
node              739 drivers/acpi/acpica/evregion.c 	obj_desc = acpi_ns_get_attached_object(node);
node              165 drivers/acpi/acpica/evrgnini.c 	parent_node = region_obj->region.node->parent;
node              180 drivers/acpi/acpica/evrgnini.c 	if (handler_obj->address_space.node == acpi_gbl_root_node) {
node              218 drivers/acpi/acpica/evrgnini.c 		pci_root_node = handler_obj->address_space.node;
node              243 drivers/acpi/acpica/evrgnini.c 	pci_device_node = region_obj->region.node;
node              289 drivers/acpi/acpica/evrgnini.c 				  region_obj->region.node);
node              312 drivers/acpi/acpica/evrgnini.c u8 acpi_ev_is_pci_root_bridge(struct acpi_namespace_node *node)
node              322 drivers/acpi/acpica/evrgnini.c 	status = acpi_ut_execute_HID(node, &hid);
node              336 drivers/acpi/acpica/evrgnini.c 	status = acpi_ut_execute_CID(node, &cid);
node              480 drivers/acpi/acpica/evrgnini.c 	struct acpi_namespace_node *node;
node              494 drivers/acpi/acpica/evrgnini.c 	node = region_obj->region.node->parent;
node              501 drivers/acpi/acpica/evrgnini.c 	while (node) {
node              506 drivers/acpi/acpica/evrgnini.c 		obj_desc = acpi_ns_get_attached_object(node);
node              511 drivers/acpi/acpica/evrgnini.c 			switch (node->type) {
node              554 drivers/acpi/acpica/evrgnini.c 		node = node->parent;
node               61 drivers/acpi/acpica/evxface.c 	struct acpi_namespace_node *node =
node              114 drivers/acpi/acpica/evxface.c 	if (!acpi_ev_is_notify_object(node)) {
node              121 drivers/acpi/acpica/evxface.c 	obj_desc = acpi_ns_get_attached_object(node);
node              126 drivers/acpi/acpica/evxface.c 		obj_desc = acpi_ut_create_internal_object(node->type);
node              134 drivers/acpi/acpica/evxface.c 		status = acpi_ns_attach_object(device, obj_desc, node->type);
node              165 drivers/acpi/acpica/evxface.c 	handler_obj->notify.node = node;
node              214 drivers/acpi/acpica/evxface.c 	struct acpi_namespace_node *node =
node              268 drivers/acpi/acpica/evxface.c 	if (!acpi_ev_is_notify_object(node)) {
node              274 drivers/acpi/acpica/evxface.c 	obj_desc = acpi_ns_get_attached_object(node);
node              859 drivers/acpi/acpica/evxfgpe.c 	struct acpi_namespace_node *node;
node              873 drivers/acpi/acpica/evxfgpe.c 	node = acpi_ns_validate_handle(gpe_device);
node              874 drivers/acpi/acpica/evxfgpe.c 	if (!node) {
node              881 drivers/acpi/acpica/evxfgpe.c 	if (node->type != ACPI_TYPE_DEVICE) {
node              886 drivers/acpi/acpica/evxfgpe.c 	if (node->object) {
node              895 drivers/acpi/acpica/evxfgpe.c 	status = acpi_ev_create_gpe_block(node, gpe_block_address->address,
node              905 drivers/acpi/acpica/evxfgpe.c 	obj_desc = acpi_ns_get_attached_object(node);
node              919 drivers/acpi/acpica/evxfgpe.c 		    acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_DEVICE);
node              956 drivers/acpi/acpica/evxfgpe.c 	struct acpi_namespace_node *node;
node              969 drivers/acpi/acpica/evxfgpe.c 	node = acpi_ns_validate_handle(gpe_device);
node              970 drivers/acpi/acpica/evxfgpe.c 	if (!node) {
node              977 drivers/acpi/acpica/evxfgpe.c 	if (node->type != ACPI_TYPE_DEVICE) {
node              984 drivers/acpi/acpica/evxfgpe.c 	obj_desc = acpi_ns_get_attached_object(node);
node               48 drivers/acpi/acpica/evxfregn.c 	struct acpi_namespace_node *node;
node               66 drivers/acpi/acpica/evxfregn.c 	node = acpi_ns_validate_handle(device);
node               67 drivers/acpi/acpica/evxfregn.c 	if (!node) {
node               75 drivers/acpi/acpica/evxfregn.c 	    acpi_ev_install_space_handler(node, space_id, handler, setup,
node               83 drivers/acpi/acpica/evxfregn.c 	acpi_ev_execute_reg_methods(node, space_id, ACPI_REG_CONNECT);
node              114 drivers/acpi/acpica/evxfregn.c 	struct acpi_namespace_node *node;
node              132 drivers/acpi/acpica/evxfregn.c 	node = acpi_ns_validate_handle(device);
node              133 drivers/acpi/acpica/evxfregn.c 	if (!node ||
node              134 drivers/acpi/acpica/evxfregn.c 	    ((node->type != ACPI_TYPE_DEVICE) &&
node              135 drivers/acpi/acpica/evxfregn.c 	     (node->type != ACPI_TYPE_PROCESSOR) &&
node              136 drivers/acpi/acpica/evxfregn.c 	     (node->type != ACPI_TYPE_THERMAL) &&
node              137 drivers/acpi/acpica/evxfregn.c 	     (node != acpi_gbl_root_node))) {
node              144 drivers/acpi/acpica/evxfregn.c 	obj_desc = acpi_ns_get_attached_object(node);
node              174 drivers/acpi/acpica/evxfregn.c 					  node, obj_desc));
node              219 drivers/acpi/acpica/evxfregn.c 			  node, obj_desc));
node              120 drivers/acpi/acpica/exconfig.c 	start_node = walk_state->scope_info->scope.node;
node              181 drivers/acpi/acpica/excreate.c 	obj_desc->mutex.node =
node              185 drivers/acpi/acpica/excreate.c 	    acpi_ns_attach_object(obj_desc->mutex.node, obj_desc,
node              219 drivers/acpi/acpica/excreate.c 	struct acpi_namespace_node *node;
node              226 drivers/acpi/acpica/excreate.c 	node = walk_state->op->common.node;
node              232 drivers/acpi/acpica/excreate.c 	if (acpi_ns_get_attached_object(node)) {
node              272 drivers/acpi/acpica/excreate.c 		    walk_state->scope_info->scope.node;
node              274 drivers/acpi/acpica/excreate.c 		region_obj2->extra.scope_node = node;
node              282 drivers/acpi/acpica/excreate.c 	obj_desc->region.node = node;
node              290 drivers/acpi/acpica/excreate.c 	status = acpi_ns_attach_object(node, obj_desc, ACPI_TYPE_REGION);
node              430 drivers/acpi/acpica/excreate.c 	obj_desc->method.node = operand[0];
node              205 drivers/acpi/acpica/exdebug.c 		if (source_desc->reference.node) {
node              207 drivers/acpi/acpica/exdebug.c 			    (source_desc->reference.node) !=
node              211 drivers/acpi/acpica/exdebug.c 				     source_desc->reference.node);
node              214 drivers/acpi/acpica/exdebug.c 					       source_desc->reference.node,
node              215 drivers/acpi/acpica/exdebug.c 					       (source_desc->reference.node)->
node              218 drivers/acpi/acpica/exdebug.c 				switch ((source_desc->reference.node)->type) {
node              234 drivers/acpi/acpica/exdebug.c 								 node)->object,
node               63 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(buffer.node), "Parent Node"},
node               69 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(package.node), "Parent Node"},
node              118 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(region.node), "Parent Node"},
node              199 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(reference.node), "Node"},
node              213 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(address_space.node), "Node"},
node              219 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(notify.node), "Node"},
node              267 drivers/acpi/acpica/exdump.c 	{ACPI_EXD_NODE, ACPI_EXD_OFFSET(common_field.node), "Parent Node"}
node              339 drivers/acpi/acpica/exdump.c 	struct acpi_namespace_node *node;
node              545 drivers/acpi/acpica/exdump.c 			node =
node              549 drivers/acpi/acpica/exdump.c 			acpi_os_printf("%20s : %p", name, node);
node              550 drivers/acpi/acpica/exdump.c 			if (node) {
node              551 drivers/acpi/acpica/exdump.c 				acpi_os_printf(" [%4.4s]", node->name.ascii);
node              668 drivers/acpi/acpica/exdump.c 			acpi_ut_repair_name(obj_desc->reference.node->name.
node              671 drivers/acpi/acpica/exdump.c 				       obj_desc->reference.node->name.ascii,
node              672 drivers/acpi/acpica/exdump.c 				       obj_desc->reference.node);
node              935 drivers/acpi/acpica/exdump.c void acpi_ex_dump_namespace_node(struct acpi_namespace_node *node, u32 flags)
node              949 drivers/acpi/acpica/exdump.c 	acpi_os_printf("%20s : %4.4s\n", "Name", acpi_ut_get_node_name(node));
node              951 drivers/acpi/acpica/exdump.c 		       node->type, acpi_ut_get_type_name(node->type));
node              953 drivers/acpi/acpica/exdump.c 	acpi_ex_dump_object(ACPI_CAST_PTR(union acpi_operand_object, node),
node              975 drivers/acpi/acpica/exdump.c 		acpi_os_printf(" %p ", obj_desc->reference.node);
node              977 drivers/acpi/acpica/exdump.c 		status = acpi_ns_handle_to_pathname(obj_desc->reference.node,
node              986 drivers/acpi/acpica/exdump.c 							     reference.node->
node              152 drivers/acpi/acpica/exfldio.c 							  common_field.node),
node              155 drivers/acpi/acpica/exfldio.c 							  node),
node              166 drivers/acpi/acpica/exfldio.c 			    acpi_ut_get_node_name(obj_desc->common_field.node),
node              170 drivers/acpi/acpica/exfldio.c 			    acpi_ut_get_node_name(rgn_desc->region.node),
node              208 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node)));
node              220 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node),
node              359 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node)));
node              368 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node)));
node              381 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node),
node              397 drivers/acpi/acpica/exmutex.c 			    acpi_ut_get_node_name(obj_desc->mutex.node),
node              474 drivers/acpi/acpica/exmutex.c 				  obj_desc->mutex.node->name.ascii,
node              861 drivers/acpi/acpica/exoparg1.c 							      node,
node               58 drivers/acpi/acpica/exoparg2.c 	struct acpi_namespace_node *node;
node               72 drivers/acpi/acpica/exoparg2.c 		node = (struct acpi_namespace_node *)operand[0];
node               80 drivers/acpi/acpica/exoparg2.c 		if (!acpi_ev_is_notify_object(node)) {
node               83 drivers/acpi/acpica/exoparg2.c 				    acpi_ut_get_type_name(node->type)));
node               96 drivers/acpi/acpica/exoparg2.c 		status = acpi_ev_queue_notify_request(node, value);
node              415 drivers/acpi/acpica/exprep.c 	obj_desc->common_field.node = info->field_node;
node               51 drivers/acpi/acpica/exresnte.c 	struct acpi_namespace_node *node;
node               60 drivers/acpi/acpica/exresnte.c 	node = *object_ptr;
node               61 drivers/acpi/acpica/exresnte.c 	source_desc = acpi_ns_get_attached_object(node);
node               62 drivers/acpi/acpica/exresnte.c 	entry_type = acpi_ns_get_type((acpi_handle)node);
node               65 drivers/acpi/acpica/exresnte.c 			  node, source_desc,
node               73 drivers/acpi/acpica/exresnte.c 		node = ACPI_CAST_PTR(struct acpi_namespace_node, node->object);
node               74 drivers/acpi/acpica/exresnte.c 		source_desc = acpi_ns_get_attached_object(node);
node               75 drivers/acpi/acpica/exresnte.c 		entry_type = acpi_ns_get_type((acpi_handle)node);
node               76 drivers/acpi/acpica/exresnte.c 		*object_ptr = node;
node               88 drivers/acpi/acpica/exresnte.c 	    (node->flags & (ANOBJ_METHOD_ARG | ANOBJ_METHOD_LOCAL))) {
node               94 drivers/acpi/acpica/exresnte.c 			    node->name.ascii, node));
node              174 drivers/acpi/acpica/exresnte.c 				  node, source_desc, entry_type));
node              200 drivers/acpi/acpica/exresnte.c 			    "Untyped entry %p, no attached object!", node));
node              235 drivers/acpi/acpica/exresnte.c 			    node, entry_type));
node              219 drivers/acpi/acpica/exresolv.c 			if ((stack_desc->reference.node->type ==
node              221 drivers/acpi/acpica/exresolv.c 			    || (stack_desc->reference.node->type ==
node              226 drivers/acpi/acpica/exresolv.c 				*stack_ptr = (void *)stack_desc->reference.node;
node              231 drivers/acpi/acpica/exresolv.c 				    (stack_desc->reference.node)->object;
node              308 drivers/acpi/acpica/exresolv.c 	struct acpi_namespace_node *node =
node              326 drivers/acpi/acpica/exresolv.c 		obj_desc = acpi_ns_get_attached_object(node);
node              351 drivers/acpi/acpica/exresolv.c 					    acpi_ut_get_node_name(node)));
node              382 drivers/acpi/acpica/exresolv.c 				node = obj_desc->reference.object;
node              385 drivers/acpi/acpica/exresolv.c 				node = obj_desc->reference.node;
node              390 drivers/acpi/acpica/exresolv.c 			if (ACPI_GET_DESCRIPTOR_TYPE(node) !=
node              394 drivers/acpi/acpica/exresolv.c 					    node,
node              395 drivers/acpi/acpica/exresolv.c 					    acpi_ut_get_descriptor_name(node)));
node              401 drivers/acpi/acpica/exresolv.c 			obj_desc = acpi_ns_get_attached_object(node);
node              406 drivers/acpi/acpica/exresolv.c 				type = acpi_ns_get_type(node);
node              474 drivers/acpi/acpica/exresolv.c 								 &node);
node              479 drivers/acpi/acpica/exresolv.c 				obj_desc = acpi_ns_get_attached_object(node);
node               28 drivers/acpi/acpica/exstore.c 			     struct acpi_namespace_node *node,
node              362 drivers/acpi/acpica/exstore.c 			     struct acpi_namespace_node *node,
node              375 drivers/acpi/acpica/exstore.c 	target_type = acpi_ns_get_type(node);
node              376 drivers/acpi/acpica/exstore.c 	target_desc = acpi_ns_get_attached_object(node);
node              380 drivers/acpi/acpica/exstore.c 			  acpi_ut_get_object_type_name(source_desc), node,
node              438 drivers/acpi/acpica/exstore.c 				    acpi_ut_get_type_name(node->type),
node              439 drivers/acpi/acpica/exstore.c 				    node->name.ascii));
node              476 drivers/acpi/acpica/exstore.c 			    acpi_ex_store_direct_to_node(source_desc, node,
node              501 drivers/acpi/acpica/exstore.c 			    acpi_ns_attach_object(node, new_desc,
node              537 drivers/acpi/acpica/exstore.c 		    acpi_ex_store_direct_to_node(source_desc, node, walk_state);
node              561 drivers/acpi/acpica/exstore.c 			     struct acpi_namespace_node *node,
node              573 drivers/acpi/acpica/exstore.c 			  source_desc, acpi_ut_get_type_name(node->type),
node              574 drivers/acpi/acpica/exstore.c 			  node));
node              586 drivers/acpi/acpica/exstore.c 	status = acpi_ns_attach_object(node, new_desc, new_desc->common.type);
node              191 drivers/acpi/acpica/nsaccess.c 				obj_desc->mutex.node = new_node;
node              322 drivers/acpi/acpica/nsaccess.c 	if ((!scope_info) || (!scope_info->scope.node)) {
node              329 drivers/acpi/acpica/nsaccess.c 		prefix_node = scope_info->scope.node;
node               28 drivers/acpi/acpica/nsalloc.c 	struct acpi_namespace_node *node;
node               35 drivers/acpi/acpica/nsalloc.c 	node = acpi_os_acquire_object(acpi_gbl_namespace_cache);
node               36 drivers/acpi/acpica/nsalloc.c 	if (!node) {
node               50 drivers/acpi/acpica/nsalloc.c 	node->name.integer = name;
node               51 drivers/acpi/acpica/nsalloc.c 	ACPI_SET_DESCRIPTOR_TYPE(node, ACPI_DESC_TYPE_NAMED);
node               52 drivers/acpi/acpica/nsalloc.c 	return_PTR(node);
node               70 drivers/acpi/acpica/nsalloc.c void acpi_ns_delete_node(struct acpi_namespace_node *node)
node               77 drivers/acpi/acpica/nsalloc.c 	if (!node) {
node               83 drivers/acpi/acpica/nsalloc.c 	acpi_ns_detach_object(node);
node               91 drivers/acpi/acpica/nsalloc.c 	obj_desc = node->object;
node               97 drivers/acpi/acpica/nsalloc.c 			obj_desc->data.handler(node, obj_desc->data.pointer);
node              107 drivers/acpi/acpica/nsalloc.c 	if (node == acpi_gbl_root_node) {
node              113 drivers/acpi/acpica/nsalloc.c 	(void)acpi_os_release_object(acpi_gbl_namespace_cache, node);
node              117 drivers/acpi/acpica/nsalloc.c 			  node, acpi_gbl_current_node_count));
node              132 drivers/acpi/acpica/nsalloc.c void acpi_ns_remove_node(struct acpi_namespace_node *node)
node              138 drivers/acpi/acpica/nsalloc.c 	ACPI_FUNCTION_TRACE_PTR(ns_remove_node, node);
node              140 drivers/acpi/acpica/nsalloc.c 	parent_node = node->parent;
node              147 drivers/acpi/acpica/nsalloc.c 	while (next_node != node) {
node              156 drivers/acpi/acpica/nsalloc.c 		prev_node->peer = node->peer;
node              162 drivers/acpi/acpica/nsalloc.c 		parent_node->child = node->peer;
node              167 drivers/acpi/acpica/nsalloc.c 	acpi_ns_delete_node(node);
node              192 drivers/acpi/acpica/nsalloc.c 			  struct acpi_namespace_node *node,	/* New Child */
node              222 drivers/acpi/acpica/nsalloc.c 	node->peer = NULL;
node              223 drivers/acpi/acpica/nsalloc.c 	node->parent = parent_node;
node              227 drivers/acpi/acpica/nsalloc.c 		parent_node->child = node;
node              235 drivers/acpi/acpica/nsalloc.c 		child_node->peer = node;
node              240 drivers/acpi/acpica/nsalloc.c 	node->owner_id = owner_id;
node              241 drivers/acpi/acpica/nsalloc.c 	node->type = (u8) type;
node              245 drivers/acpi/acpica/nsalloc.c 			  acpi_ut_get_node_name(node),
node              246 drivers/acpi/acpica/nsalloc.c 			  acpi_ut_get_type_name(node->type), node, owner_id,
node               45 drivers/acpi/acpica/nsarguments.c 	if (!info->predefined || (info->node->flags & ANOBJ_EVALUATED)) {
node               70 drivers/acpi/acpica/nsarguments.c 			info->node->flags |= ANOBJ_EVALUATED;
node               93 drivers/acpi/acpica/nsarguments.c 			      struct acpi_namespace_node *node,
node               99 drivers/acpi/acpica/nsarguments.c 	if (!predefined || (node->flags & ANOBJ_EVALUATED)) {
node              112 drivers/acpi/acpica/nsarguments.c 	if (node->type != ACPI_TYPE_METHOD) {
node              120 drivers/acpi/acpica/nsarguments.c 						    acpi_ut_get_type_name(node->
node              132 drivers/acpi/acpica/nsarguments.c 						    acpi_ut_get_type_name(node->
node              149 drivers/acpi/acpica/nsarguments.c 	aml_param_count = node->object->method.param_count;
node              186 drivers/acpi/acpica/nsarguments.c 			     struct acpi_namespace_node *node,
node              193 drivers/acpi/acpica/nsarguments.c 	if (node->flags & ANOBJ_EVALUATED) {
node              202 drivers/acpi/acpica/nsarguments.c 		if (node->type != ACPI_TYPE_METHOD) {
node              209 drivers/acpi/acpica/nsarguments.c 						      (node->type)));
node              227 drivers/acpi/acpica/nsarguments.c 		aml_param_count = node->object->method.param_count;
node              438 drivers/acpi/acpica/nsconvert.c 	struct acpi_namespace_node *node;
node              454 drivers/acpi/acpica/nsconvert.c 	scope_info.scope.node =
node              459 drivers/acpi/acpica/nsconvert.c 			   NULL, &node);
node              476 drivers/acpi/acpica/nsconvert.c 	new_object->reference.node = node;
node              477 drivers/acpi/acpica/nsconvert.c 	new_object->reference.object = node->object;
node              484 drivers/acpi/acpica/nsconvert.c 	acpi_ut_add_reference(node->object);
node              338 drivers/acpi/acpica/nsdump.c 			    obj_desc->buffer_field.buffer_obj->buffer.node) {
node              344 drivers/acpi/acpica/nsdump.c 								     node));
node              354 drivers/acpi/acpica/nsdump.c 							     node));
node              363 drivers/acpi/acpica/nsdump.c 							     node),
node              368 drivers/acpi/acpica/nsdump.c 							     node));
node              377 drivers/acpi/acpica/nsdump.c 							     common_field.node),
node              382 drivers/acpi/acpica/nsdump.c 							     node));
node              682 drivers/acpi/acpica/nsdump.c 	struct acpi_namespace_node *node;
node              689 drivers/acpi/acpica/nsdump.c 	node = acpi_ns_validate_handle(obj_handle);
node              690 drivers/acpi/acpica/nsdump.c 	if (!node) {
node              697 drivers/acpi/acpica/nsdump.c 	pathname = acpi_ns_get_normalized_pathname(node, TRUE);
node              705 drivers/acpi/acpica/nsdump.c 		       level, level, " ", acpi_ut_get_type_name(node->type),
node               52 drivers/acpi/acpica/nseval.c 	if (!info->node) {
node               63 drivers/acpi/acpica/nseval.c 				     ACPI_NS_NO_UPSEARCH, &info->node);
node               73 drivers/acpi/acpica/nseval.c 	if (acpi_ns_get_type(info->node) == ACPI_TYPE_LOCAL_METHOD_ALIAS) {
node               74 drivers/acpi/acpica/nseval.c 		info->node =
node               76 drivers/acpi/acpica/nseval.c 				  info->node->object);
node               82 drivers/acpi/acpica/nseval.c 	info->node_flags = info->node->flags;
node               83 drivers/acpi/acpica/nseval.c 	info->obj_desc = acpi_ns_get_attached_object(info->node);
node               86 drivers/acpi/acpica/nseval.c 			  info->relative_pathname, info->node,
node               87 drivers/acpi/acpica/nseval.c 			  acpi_ns_get_attached_object(info->node)));
node               92 drivers/acpi/acpica/nseval.c 	    acpi_ut_match_predefined_method(info->node->name.ascii);
node               96 drivers/acpi/acpica/nseval.c 	info->full_pathname = acpi_ns_get_normalized_pathname(info->node, TRUE);
node              106 drivers/acpi/acpica/nseval.c 			      acpi_ut_get_type_name(info->node->type)));
node              133 drivers/acpi/acpica/nseval.c 	acpi_ns_check_acpi_compliance(info->full_pathname, info->node,
node              140 drivers/acpi/acpica/nseval.c 	acpi_ns_check_argument_count(info->full_pathname, info->node,
node              154 drivers/acpi/acpica/nseval.c 	switch (acpi_ns_get_type(info->node)) {
node              170 drivers/acpi/acpica/nseval.c 			    acpi_ut_get_type_name(info->node->type)));
node              235 drivers/acpi/acpica/nseval.c 		    ACPI_CAST_PTR(union acpi_operand_object, info->node);
node              261 drivers/acpi/acpica/nseval.c 	(void)acpi_ns_check_return_value(info->node, info, info->param_count,
node              265 drivers/acpi/acpica/nsinit.c 	struct acpi_namespace_node *node =
node              268 drivers/acpi/acpica/nsinit.c 	obj_desc = acpi_ns_get_attached_object(node);
node              323 drivers/acpi/acpica/nsinit.c 	struct acpi_namespace_node *node =
node              334 drivers/acpi/acpica/nsinit.c 	obj_desc = acpi_ns_get_attached_object(node);
node              413 drivers/acpi/acpica/nsinit.c 				acpi_ut_get_node_name(node),
node              421 drivers/acpi/acpica/nsinit.c 				acpi_ut_get_node_name(node),
node              455 drivers/acpi/acpica/nsinit.c 	struct acpi_namespace_node *node;
node              460 drivers/acpi/acpica/nsinit.c 	node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_handle);
node              461 drivers/acpi/acpica/nsinit.c 	if ((node->type == ACPI_TYPE_DEVICE) ||
node              462 drivers/acpi/acpica/nsinit.c 	    (node->type == ACPI_TYPE_PROCESSOR) ||
node              463 drivers/acpi/acpica/nsinit.c 	    (node->type == ACPI_TYPE_THERMAL)) {
node              470 drivers/acpi/acpica/nsinit.c 	if (!ACPI_COMPARE_NAMESEG(node->name.ascii, METHOD_NAME__INI)) {
node              478 drivers/acpi/acpica/nsinit.c 	parent_node = node->parent;
node               41 drivers/acpi/acpica/nsload.c acpi_ns_load_table(u32 table_index, struct acpi_namespace_node *node)
node               71 drivers/acpi/acpica/nsload.c 	status = acpi_ns_parse_table(table_index, node);
node              106 drivers/acpi/acpica/nsload.c 	status = acpi_ds_initialize_objects(table_index, node);
node               30 drivers/acpi/acpica/nsnames.c char *acpi_ns_get_external_pathname(struct acpi_namespace_node *node)
node               34 drivers/acpi/acpica/nsnames.c 	ACPI_FUNCTION_TRACE_PTR(ns_get_external_pathname, node);
node               36 drivers/acpi/acpica/nsnames.c 	name_buffer = acpi_ns_get_normalized_pathname(node, FALSE);
node               52 drivers/acpi/acpica/nsnames.c acpi_size acpi_ns_get_pathname_length(struct acpi_namespace_node *node)
node               58 drivers/acpi/acpica/nsnames.c 	if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) {
node               61 drivers/acpi/acpica/nsnames.c 			    node, ACPI_GET_DESCRIPTOR_TYPE(node)));
node               65 drivers/acpi/acpica/nsnames.c 	size = acpi_ns_build_normalized_path(node, NULL, 0, FALSE);
node               87 drivers/acpi/acpica/nsnames.c 	struct acpi_namespace_node *node;
node               92 drivers/acpi/acpica/nsnames.c 	node = acpi_ns_validate_handle(target_handle);
node               93 drivers/acpi/acpica/nsnames.c 	if (!node) {
node              106 drivers/acpi/acpica/nsnames.c 	node_name = acpi_ut_get_node_name(node);
node              135 drivers/acpi/acpica/nsnames.c 	struct acpi_namespace_node *node;
node              140 drivers/acpi/acpica/nsnames.c 	node = acpi_ns_validate_handle(target_handle);
node              141 drivers/acpi/acpica/nsnames.c 	if (!node) {
node              148 drivers/acpi/acpica/nsnames.c 	    acpi_ns_build_normalized_path(node, NULL, 0, no_trailing);
node              162 drivers/acpi/acpica/nsnames.c 	(void)acpi_ns_build_normalized_path(node, buffer->pointer,
node              193 drivers/acpi/acpica/nsnames.c acpi_ns_build_normalized_path(struct acpi_namespace_node *node,
node              202 drivers/acpi/acpica/nsnames.c 	ACPI_FUNCTION_TRACE_PTR(ns_build_normalized_path, node);
node              221 drivers/acpi/acpica/nsnames.c 	if (!node) {
node              225 drivers/acpi/acpica/nsnames.c 	next_node = node;
node              227 drivers/acpi/acpica/nsnames.c 		if (next_node != node) {
node              289 drivers/acpi/acpica/nsnames.c char *acpi_ns_get_normalized_pathname(struct acpi_namespace_node *node,
node              295 drivers/acpi/acpica/nsnames.c 	ACPI_FUNCTION_TRACE_PTR(ns_get_normalized_pathname, node);
node              299 drivers/acpi/acpica/nsnames.c 	size = acpi_ns_build_normalized_path(node, NULL, 0, no_trailing);
node              314 drivers/acpi/acpica/nsnames.c 	(void)acpi_ns_build_normalized_path(node, name_buffer, size,
node              349 drivers/acpi/acpica/nsnames.c 	if (prefix_scope && prefix_scope->scope.node) {
node              351 drivers/acpi/acpica/nsnames.c 		    acpi_ns_get_normalized_pathname(prefix_scope->scope.node,
node               37 drivers/acpi/acpica/nsobject.c acpi_ns_attach_object(struct acpi_namespace_node *node,
node               49 drivers/acpi/acpica/nsobject.c 	if (!node) {
node               66 drivers/acpi/acpica/nsobject.c 	if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) {
node               71 drivers/acpi/acpica/nsobject.c 			    node, acpi_ut_get_descriptor_name(node)));
node               77 drivers/acpi/acpica/nsobject.c 	if (node->object == object) {
node               80 drivers/acpi/acpica/nsobject.c 				  object, node));
node              119 drivers/acpi/acpica/nsobject.c 			  obj_desc, node, acpi_ut_get_node_name(node)));
node              123 drivers/acpi/acpica/nsobject.c 	if (node->object) {
node              124 drivers/acpi/acpica/nsobject.c 		acpi_ns_detach_object(node);
node              145 drivers/acpi/acpica/nsobject.c 		last_obj_desc->common.next_object = node->object;
node              148 drivers/acpi/acpica/nsobject.c 	node->type = (u8) object_type;
node              149 drivers/acpi/acpica/nsobject.c 	node->object = obj_desc;
node              168 drivers/acpi/acpica/nsobject.c void acpi_ns_detach_object(struct acpi_namespace_node *node)
node              174 drivers/acpi/acpica/nsobject.c 	obj_desc = node->object;
node              180 drivers/acpi/acpica/nsobject.c 	if (node->flags & ANOBJ_ALLOCATED_BUFFER) {
node              190 drivers/acpi/acpica/nsobject.c 		acpi_ut_remove_address_range(obj_desc->region.space_id, node);
node              195 drivers/acpi/acpica/nsobject.c 	node->object = NULL;
node              200 drivers/acpi/acpica/nsobject.c 		node->object = obj_desc->common.next_object;
node              204 drivers/acpi/acpica/nsobject.c 		if (node->object &&
node              205 drivers/acpi/acpica/nsobject.c 		    (node->object->common.type != ACPI_TYPE_LOCAL_DATA)) {
node              206 drivers/acpi/acpica/nsobject.c 			node->object = node->object->common.next_object;
node              222 drivers/acpi/acpica/nsobject.c 	node->type = ACPI_TYPE_ANY;
node              225 drivers/acpi/acpica/nsobject.c 			  node, acpi_ut_get_node_name(node), obj_desc));
node              248 drivers/acpi/acpica/nsobject.c 						       *node)
node              250 drivers/acpi/acpica/nsobject.c 	ACPI_FUNCTION_TRACE_PTR(ns_get_attached_object, node);
node              252 drivers/acpi/acpica/nsobject.c 	if (!node) {
node              257 drivers/acpi/acpica/nsobject.c 	if (!node->object ||
node              258 drivers/acpi/acpica/nsobject.c 	    ((ACPI_GET_DESCRIPTOR_TYPE(node->object) != ACPI_DESC_TYPE_OPERAND)
node              259 drivers/acpi/acpica/nsobject.c 	     && (ACPI_GET_DESCRIPTOR_TYPE(node->object) !=
node              261 drivers/acpi/acpica/nsobject.c 	    || ((node->object)->common.type == ACPI_TYPE_LOCAL_DATA)) {
node              265 drivers/acpi/acpica/nsobject.c 	return_PTR(node->object);
node              313 drivers/acpi/acpica/nsobject.c acpi_ns_attach_data(struct acpi_namespace_node *node,
node              323 drivers/acpi/acpica/nsobject.c 	obj_desc = node->object;
node              349 drivers/acpi/acpica/nsobject.c 		node->object = data_desc;
node              370 drivers/acpi/acpica/nsobject.c acpi_ns_detach_data(struct acpi_namespace_node *node,
node              377 drivers/acpi/acpica/nsobject.c 	obj_desc = node->object;
node              385 drivers/acpi/acpica/nsobject.c 				node->object = obj_desc->common.next_object;
node              415 drivers/acpi/acpica/nsobject.c acpi_ns_get_attached_data(struct acpi_namespace_node *node,
node              420 drivers/acpi/acpica/nsobject.c 	obj_desc = node->object;
node              101 drivers/acpi/acpica/nsparse.c 	info->node = start_node;
node              103 drivers/acpi/acpica/nsparse.c 	info->node_flags = info->node->flags;
node              104 drivers/acpi/acpica/nsparse.c 	info->full_pathname = acpi_ns_get_normalized_pathname(info->node, TRUE);
node               65 drivers/acpi/acpica/nspredef.c acpi_ns_check_return_value(struct acpi_namespace_node *node,
node              154 drivers/acpi/acpica/nspredef.c 	status = acpi_ns_complex_repairs(info, node, status, return_object_ptr);
node              163 drivers/acpi/acpica/nspredef.c 		node->flags |= ANOBJ_EVALUATED;
node              204 drivers/acpi/acpica/nspredef.c 				      return_object->node.name.ascii,
node              205 drivers/acpi/acpica/nspredef.c 				      acpi_ut_get_type_name(return_object->node.
node               55 drivers/acpi/acpica/nsrepair.c 									 *node,
node              136 drivers/acpi/acpica/nsrepair.c 	predefined = acpi_ns_match_simple_repair(info->node,
node              146 drivers/acpi/acpica/nsrepair.c 		status = predefined->object_converter(info->node, return_object,
node              307 drivers/acpi/acpica/nsrepair.c 									 *node,
node              319 drivers/acpi/acpica/nsrepair.c 		if (ACPI_COMPARE_NAMESEG(node->name.ascii, this_name->name)) {
node               37 drivers/acpi/acpica/nsrepair2.c 								   *node);
node              151 drivers/acpi/acpica/nsrepair2.c 			struct acpi_namespace_node *node,
node              160 drivers/acpi/acpica/nsrepair2.c 	predefined = acpi_ns_match_complex_repair(node);
node              183 drivers/acpi/acpica/nsrepair2.c 								   *node)
node              191 drivers/acpi/acpica/nsrepair2.c 		if (ACPI_COMPARE_NAMESEG(node->name.ascii, this_name->name)) {
node              715 drivers/acpi/acpica/nsrepair2.c 	struct acpi_namespace_node *node;
node              725 drivers/acpi/acpica/nsrepair2.c 	status = acpi_ns_get_node(info->node, "^_PSS",
node              726 drivers/acpi/acpica/nsrepair2.c 				  ACPI_NS_NO_UPSEARCH, &node);
node               22 drivers/acpi/acpica/nssearch.c 			   struct acpi_namespace_node *node,
node               64 drivers/acpi/acpica/nssearch.c 	struct acpi_namespace_node *node;
node               89 drivers/acpi/acpica/nssearch.c 	node = parent_node->child;
node               90 drivers/acpi/acpica/nssearch.c 	while (node) {
node               94 drivers/acpi/acpica/nssearch.c 		if (node->name.integer == target_name) {
node               98 drivers/acpi/acpica/nssearch.c 			if (acpi_ns_get_type(node) ==
node              100 drivers/acpi/acpica/nssearch.c 				node =
node              102 drivers/acpi/acpica/nssearch.c 						  node->object);
node              110 drivers/acpi/acpica/nssearch.c 					  acpi_ut_get_type_name(node->type),
node              111 drivers/acpi/acpica/nssearch.c 					  node,
node              115 drivers/acpi/acpica/nssearch.c 			*return_node = node;
node              121 drivers/acpi/acpica/nssearch.c 		node = node->peer;
node              164 drivers/acpi/acpica/nssearch.c 			   struct acpi_namespace_node *node,
node              173 drivers/acpi/acpica/nssearch.c 	parent_node = node->parent;
node              253 drivers/acpi/acpica/nssearch.c 			 struct acpi_namespace_node *node,
node              265 drivers/acpi/acpica/nssearch.c 	if (!node || !target_name || !return_node) {
node              268 drivers/acpi/acpica/nssearch.c 			    node, target_name, return_node));
node              286 drivers/acpi/acpica/nssearch.c 	status = acpi_ns_search_one_scope(target_name, node, type, return_node);
node              354 drivers/acpi/acpica/nssearch.c 		    acpi_ns_search_parent_tree(target_name, node, type,
node              366 drivers/acpi/acpica/nssearch.c 				  ACPI_CAST_PTR(char, &target_name), node));
node              393 drivers/acpi/acpica/nssearch.c 	acpi_ns_install_node(walk_state, node, new_node, type);
node               37 drivers/acpi/acpica/nsutils.c acpi_ns_print_node_pathname(struct acpi_namespace_node *node,
node               43 drivers/acpi/acpica/nsutils.c 	if (!node) {
node               52 drivers/acpi/acpica/nsutils.c 	status = acpi_ns_handle_to_pathname(node, &buffer, TRUE);
node               75 drivers/acpi/acpica/nsutils.c acpi_object_type acpi_ns_get_type(struct acpi_namespace_node * node)
node               79 drivers/acpi/acpica/nsutils.c 	if (!node) {
node               84 drivers/acpi/acpica/nsutils.c 	return_UINT8(node->type);
node              673 drivers/acpi/acpica/nsutils.c 	scope_info.scope.node = prefix_node;
node              277 drivers/acpi/acpica/nsxfeval.c 	switch (acpi_ns_get_type(info->node)) {
node              475 drivers/acpi/acpica/nsxfeval.c 	struct acpi_namespace_node *node;
node              498 drivers/acpi/acpica/nsxfeval.c 		node = info->return_object->reference.object;
node              499 drivers/acpi/acpica/nsxfeval.c 		if (node) {
node              500 drivers/acpi/acpica/nsxfeval.c 			obj_desc = node->object;
node              641 drivers/acpi/acpica/nsxfeval.c 	struct acpi_namespace_node *node;
node              654 drivers/acpi/acpica/nsxfeval.c 	node = acpi_ns_validate_handle(obj_handle);
node              660 drivers/acpi/acpica/nsxfeval.c 	if (!node) {
node              679 drivers/acpi/acpica/nsxfeval.c 		status = acpi_ut_execute_HID(node, &hid);
node              694 drivers/acpi/acpica/nsxfeval.c 			status = acpi_ut_execute_CID(node, &cid);
node              723 drivers/acpi/acpica/nsxfeval.c 	status = acpi_ut_execute_STA(node, &flags);
node              833 drivers/acpi/acpica/nsxfeval.c 	struct acpi_namespace_node *node;
node              849 drivers/acpi/acpica/nsxfeval.c 	node = acpi_ns_validate_handle(obj_handle);
node              850 drivers/acpi/acpica/nsxfeval.c 	if (!node) {
node              855 drivers/acpi/acpica/nsxfeval.c 	status = acpi_ns_attach_data(node, handler, data);
node              879 drivers/acpi/acpica/nsxfeval.c 	struct acpi_namespace_node *node;
node              895 drivers/acpi/acpica/nsxfeval.c 	node = acpi_ns_validate_handle(obj_handle);
node              896 drivers/acpi/acpica/nsxfeval.c 	if (!node) {
node              901 drivers/acpi/acpica/nsxfeval.c 	status = acpi_ns_detach_data(node, handler);
node              929 drivers/acpi/acpica/nsxfeval.c 	struct acpi_namespace_node *node;
node              945 drivers/acpi/acpica/nsxfeval.c 	node = acpi_ns_validate_handle(obj_handle);
node              946 drivers/acpi/acpica/nsxfeval.c 	if (!node) {
node              951 drivers/acpi/acpica/nsxfeval.c 	status = acpi_ns_get_attached_data(node, handler, data);
node               50 drivers/acpi/acpica/nsxfname.c 	struct acpi_namespace_node *node = NULL;
node               98 drivers/acpi/acpica/nsxfname.c 	    acpi_ns_get_node(prefix_node, pathname, ACPI_NS_NO_UPSEARCH, &node);
node              100 drivers/acpi/acpica/nsxfname.c 		*ret_handle = ACPI_CAST_PTR(acpi_handle, node);
node              229 drivers/acpi/acpica/nsxfname.c 	struct acpi_namespace_node *node;
node              255 drivers/acpi/acpica/nsxfname.c 	node = acpi_ns_validate_handle(handle);
node              256 drivers/acpi/acpica/nsxfname.c 	if (!node) {
node              264 drivers/acpi/acpica/nsxfname.c 	type = node->type;
node              265 drivers/acpi/acpica/nsxfname.c 	name = node->name.integer;
node              267 drivers/acpi/acpica/nsxfname.c 	if (node->type == ACPI_TYPE_METHOD) {
node              268 drivers/acpi/acpica/nsxfname.c 		param_count = node->object->method.param_count;
node              288 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_HID(node, &hid);
node              296 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_UID(node, &uid);
node              304 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_CID(node, &cid_list);
node              317 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_CLS(node, &cls);
node              348 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_evaluate_numeric_object(METHOD_NAME__ADR, node,
node              356 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_power_methods(node,
node              366 drivers/acpi/acpica/nsxfname.c 		status = acpi_ut_execute_power_methods(node,
node              482 drivers/acpi/acpica/nsxfname.c 	struct acpi_namespace_node *node;
node              548 drivers/acpi/acpica/nsxfname.c 			   NULL, &node);
node              559 drivers/acpi/acpica/nsxfname.c 		if (node->type != ACPI_TYPE_METHOD) {
node              588 drivers/acpi/acpica/nsxfname.c 	status = acpi_ns_attach_object(node, method_obj, ACPI_TYPE_METHOD);
node              594 drivers/acpi/acpica/nsxfname.c 	node->flags |= ANOBJ_ALLOCATED_BUFFER;
node               32 drivers/acpi/acpica/nsxfobj.c 	struct acpi_namespace_node *node;
node               55 drivers/acpi/acpica/nsxfobj.c 	node = acpi_ns_validate_handle(handle);
node               56 drivers/acpi/acpica/nsxfobj.c 	if (!node) {
node               61 drivers/acpi/acpica/nsxfobj.c 	*ret_type = node->type;
node               84 drivers/acpi/acpica/nsxfobj.c 	struct acpi_namespace_node *node;
node              105 drivers/acpi/acpica/nsxfobj.c 	node = acpi_ns_validate_handle(handle);
node              106 drivers/acpi/acpica/nsxfobj.c 	if (!node) {
node              113 drivers/acpi/acpica/nsxfobj.c 	parent_node = node->parent;
node              153 drivers/acpi/acpica/nsxfobj.c 	struct acpi_namespace_node *node;
node              192 drivers/acpi/acpica/nsxfobj.c 	node = acpi_ns_get_next_node_typed(type, parent_node, child_node);
node              193 drivers/acpi/acpica/nsxfobj.c 	if (!node) {
node              199 drivers/acpi/acpica/nsxfobj.c 		*ret_handle = ACPI_CAST_PTR(acpi_handle, node);
node              205 drivers/acpi/acpica/psargs.c 	struct acpi_namespace_node *node;
node              231 drivers/acpi/acpica/psargs.c 				NULL, &node);
node              238 drivers/acpi/acpica/psargs.c 	    possible_method_call && (node->type == ACPI_TYPE_METHOD)) {
node              256 drivers/acpi/acpica/psargs.c 		method_desc = acpi_ns_get_attached_object(node);
node              259 drivers/acpi/acpica/psargs.c 				  node->name.ascii, node, method_desc, path));
node              273 drivers/acpi/acpica/psargs.c 		name_op->common.node = node;
node              279 drivers/acpi/acpica/psargs.c 				    node));
node              285 drivers/acpi/acpica/psargs.c 				  node, method_desc->method.param_count));
node              562 drivers/acpi/acpica/psobject.c 								node);
node              563 drivers/acpi/acpica/psobject.c 					acpi_ns_remove_node((*op)->common.node);
node              564 drivers/acpi/acpica/psobject.c 					(*op)->common.node = NULL;
node              232 drivers/acpi/acpica/psparse.c 				replacement_op->common.node = op->common.node;
node              256 drivers/acpi/acpica/psparse.c 						replacement_op->common.node =
node              257 drivers/acpi/acpica/psparse.c 						    op->common.node;
node              375 drivers/acpi/acpica/psparse.c 		    (op->common.value.arg)->common.node;
node               98 drivers/acpi/acpica/psxface.c 	if (!info || !info->node) {
node              105 drivers/acpi/acpica/psxface.c 	    acpi_ds_begin_method_execution(info->node, info->obj_desc, NULL);
node              120 drivers/acpi/acpica/psxface.c 			  info->node->name.ascii, info->node, info->obj_desc));
node              141 drivers/acpi/acpica/psxface.c 	status = acpi_ds_init_aml_walk(walk_state, op, info->node,
node              265 drivers/acpi/acpica/psxface.c 	status = acpi_ds_init_aml_walk(walk_state, op, info->node,
node              282 drivers/acpi/acpica/psxface.c 	if (info->node && info->node != acpi_gbl_root_node) {
node              284 drivers/acpi/acpica/psxface.c 		    acpi_ds_scope_stack_push(info->node, ACPI_TYPE_METHOD,
node              798 drivers/acpi/acpica/rscalc.c 				temp_size_needed += acpi_ns_get_pathname_length((*sub_object_list)->reference.node);
node              190 drivers/acpi/acpica/rscreate.c 	struct acpi_namespace_node *node;
node              305 drivers/acpi/acpica/rscreate.c 				node = obj_desc->reference.node;
node              314 drivers/acpi/acpica/rscreate.c 				status = acpi_ns_handle_to_pathname((acpi_handle)node, &path_buffer, FALSE);
node              433 drivers/acpi/acpica/rsutils.c acpi_rs_get_prt_method_data(struct acpi_namespace_node *node,
node              446 drivers/acpi/acpica/rsutils.c 	    acpi_ut_evaluate_object(node, METHOD_NAME__PRT, ACPI_BTYPE_PACKAGE,
node              483 drivers/acpi/acpica/rsutils.c acpi_rs_get_crs_method_data(struct acpi_namespace_node *node,
node              496 drivers/acpi/acpica/rsutils.c 	    acpi_ut_evaluate_object(node, METHOD_NAME__CRS, ACPI_BTYPE_BUFFER,
node              534 drivers/acpi/acpica/rsutils.c acpi_rs_get_prs_method_data(struct acpi_namespace_node *node,
node              547 drivers/acpi/acpica/rsutils.c 	    acpi_ut_evaluate_object(node, METHOD_NAME__PRS, ACPI_BTYPE_BUFFER,
node              585 drivers/acpi/acpica/rsutils.c acpi_rs_get_aei_method_data(struct acpi_namespace_node *node,
node              598 drivers/acpi/acpica/rsutils.c 	    acpi_ut_evaluate_object(node, METHOD_NAME__AEI, ACPI_BTYPE_BUFFER,
node              691 drivers/acpi/acpica/rsutils.c acpi_rs_set_srs_method_data(struct acpi_namespace_node *node,
node              708 drivers/acpi/acpica/rsutils.c 	info->prefix_node = node;
node               62 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node               73 drivers/acpi/acpica/rsxface.c 	node = acpi_ns_validate_handle(device_handle);
node               74 drivers/acpi/acpica/rsxface.c 	if (!node) {
node               78 drivers/acpi/acpica/rsxface.c 	if (node->type != ACPI_TYPE_DEVICE) {
node               94 drivers/acpi/acpica/rsxface.c 	*return_node = node;
node              126 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node              132 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node);
node              137 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_get_prt_method_data(node, ret_buffer);
node              171 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node              177 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node);
node              182 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_get_crs_method_data(node, ret_buffer);
node              213 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node              219 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node);
node              224 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_get_prs_method_data(node, ret_buffer);
node              252 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node              264 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_validate_parameters(device_handle, in_buffer, &node);
node              269 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_set_srs_method_data(node, in_buffer);
node              298 drivers/acpi/acpica/rsxface.c 	struct acpi_namespace_node *node;
node              304 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_validate_parameters(device_handle, ret_buffer, &node);
node              309 drivers/acpi/acpica/rsxface.c 	status = acpi_rs_get_aei_method_data(node, ret_buffer);
node              329 drivers/acpi/acpica/tbxfload.c 	struct acpi_namespace_node *node =
node              347 drivers/acpi/acpica/tbxfload.c 	owner_id = node->owner_id;
node              152 drivers/acpi/acpica/utcopy.c 			    internal_object->reference.node;
node              154 drivers/acpi/acpica/utcopy.c 			    acpi_ns_get_type(internal_object->reference.node);
node              240 drivers/acpi/acpica/utdecode.c 	struct acpi_namespace_node *node = (struct acpi_namespace_node *)object;
node              256 drivers/acpi/acpica/utdecode.c 	if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) {
node              264 drivers/acpi/acpica/utdecode.c 	acpi_ut_repair_name(node->name.ascii);
node              268 drivers/acpi/acpica/utdecode.c 	return (node->name.ascii);
node              177 drivers/acpi/acpica/utdelete.c 		if (object->method.node) {
node              178 drivers/acpi/acpica/utdelete.c 			object->method.node = NULL;
node              191 drivers/acpi/acpica/utdelete.c 		if (!(object->region.node->flags & ANOBJ_TEMPORARY)) {
node              193 drivers/acpi/acpica/utdelete.c 						     object->region.node);
node              306 drivers/acpi/acpica/uterror.c 	struct acpi_namespace_node *node = prefix_node;
node              313 drivers/acpi/acpica/uterror.c 					  ACPI_NS_NO_UPSEARCH, &node);
node              319 drivers/acpi/acpica/uterror.c 	acpi_ns_print_node_pathname(node, message);
node              497 drivers/acpi/acpica/utobject.c 							reference.node);
node              673 drivers/acpi/acpica/uttrack.c 								node));
node               80 drivers/acpi/arm64/iort.c 			struct acpi_iort_node *node)
node               87 drivers/acpi/arm64/iort.c 		if (curr->iort_node == node) {
node              102 drivers/acpi/arm64/iort.c static inline void iort_delete_fwnode(struct acpi_iort_node *node)
node              108 drivers/acpi/arm64/iort.c 		if (curr->iort_node == node) {
node              143 drivers/acpi/arm64/iort.c 	(struct acpi_iort_node *node, void *context);
node              258 drivers/acpi/arm64/iort.c static acpi_status iort_match_node_callback(struct acpi_iort_node *node,
node              264 drivers/acpi/arm64/iort.c 	if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT) {
node              278 drivers/acpi/arm64/iort.c 		ncomp = (struct acpi_iort_named_component *)node->node_data;
node              282 drivers/acpi/arm64/iort.c 	} else if (node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX) {
node              287 drivers/acpi/arm64/iort.c 		pci_rc = (struct acpi_iort_root_complex *)node->node_data;
node              325 drivers/acpi/arm64/iort.c static struct acpi_iort_node *iort_node_get_id(struct acpi_iort_node *node,
node              331 drivers/acpi/arm64/iort.c 	if (!node->mapping_offset || !node->mapping_count ||
node              332 drivers/acpi/arm64/iort.c 				     index >= node->mapping_count)
node              335 drivers/acpi/arm64/iort.c 	map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node,
node              336 drivers/acpi/arm64/iort.c 			   node->mapping_offset + index * sizeof(*map));
node              341 drivers/acpi/arm64/iort.c 		       node, node->type);
node              349 drivers/acpi/arm64/iort.c 		if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT ||
node              350 drivers/acpi/arm64/iort.c 		    node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX ||
node              351 drivers/acpi/arm64/iort.c 		    node->type == ACPI_IORT_NODE_SMMU_V3 ||
node              352 drivers/acpi/arm64/iort.c 		    node->type == ACPI_IORT_NODE_PMCG) {
node              361 drivers/acpi/arm64/iort.c static int iort_get_id_mapping_index(struct acpi_iort_node *node)
node              365 drivers/acpi/arm64/iort.c 	switch (node->type) {
node              371 drivers/acpi/arm64/iort.c 		if (node->revision < 1)
node              374 drivers/acpi/arm64/iort.c 		smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node              383 drivers/acpi/arm64/iort.c 		if (smmu->id_mapping_index >= node->mapping_count) {
node              385 drivers/acpi/arm64/iort.c 			       node, node->type);
node              397 drivers/acpi/arm64/iort.c static struct acpi_iort_node *iort_node_map_id(struct acpi_iort_node *node,
node              404 drivers/acpi/arm64/iort.c 	while (node) {
node              408 drivers/acpi/arm64/iort.c 		if (IORT_TYPE_MASK(node->type) & type_mask) {
node              411 drivers/acpi/arm64/iort.c 			return node;
node              414 drivers/acpi/arm64/iort.c 		if (!node->mapping_offset || !node->mapping_count)
node              417 drivers/acpi/arm64/iort.c 		map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node,
node              418 drivers/acpi/arm64/iort.c 				   node->mapping_offset);
node              423 drivers/acpi/arm64/iort.c 			       node, node->type);
node              432 drivers/acpi/arm64/iort.c 		index = iort_get_id_mapping_index(node);
node              435 drivers/acpi/arm64/iort.c 		for (i = 0; i < node->mapping_count; i++, map++) {
node              440 drivers/acpi/arm64/iort.c 			if (!iort_id_map(map, node->type, id, &id))
node              444 drivers/acpi/arm64/iort.c 		if (i == node->mapping_count)
node              447 drivers/acpi/arm64/iort.c 		node = ACPI_ADD_PTR(struct acpi_iort_node, iort_table,
node              460 drivers/acpi/arm64/iort.c 		struct acpi_iort_node *node, u32 *id_out, u8 type_mask,
node              467 drivers/acpi/arm64/iort.c 	parent = iort_node_get_id(node, &id, index);
node              491 drivers/acpi/arm64/iort.c 		struct acpi_iort_node *node;
node              498 drivers/acpi/arm64/iort.c 		node = iort_get_iort_node(dev->fwnode);
node              499 drivers/acpi/arm64/iort.c 		if (node)
node              500 drivers/acpi/arm64/iort.c 			return node;
node              528 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node              531 drivers/acpi/arm64/iort.c 	node = iort_find_dev_node(dev);
node              532 drivers/acpi/arm64/iort.c 	if (!node)
node              535 drivers/acpi/arm64/iort.c 	iort_node_map_id(node, req_id, &dev_id, IORT_MSI_TYPE);
node              549 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node              551 drivers/acpi/arm64/iort.c 	node = iort_find_dev_node(dev);
node              552 drivers/acpi/arm64/iort.c 	if (!node)
node              555 drivers/acpi/arm64/iort.c 	index = iort_get_id_mapping_index(node);
node              558 drivers/acpi/arm64/iort.c 		if (iort_node_get_id(node, dev_id, index))
node              561 drivers/acpi/arm64/iort.c 		for (i = 0; i < node->mapping_count; i++) {
node              562 drivers/acpi/arm64/iort.c 			if (iort_node_map_platform_id(node, dev_id,
node              602 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node              604 drivers/acpi/arm64/iort.c 	node = iort_find_dev_node(dev);
node              605 drivers/acpi/arm64/iort.c 	if (!node)
node              608 drivers/acpi/arm64/iort.c 	node = iort_node_map_id(node, req_id, NULL, IORT_MSI_TYPE);
node              609 drivers/acpi/arm64/iort.c 	if (!node)
node              613 drivers/acpi/arm64/iort.c 	its = (struct acpi_iort_its_group *)node->node_data;
node              647 drivers/acpi/arm64/iort.c 				   struct acpi_iort_node *node)
node              656 drivers/acpi/arm64/iort.c 	index = iort_get_id_mapping_index(node);
node              660 drivers/acpi/arm64/iort.c 	map = ACPI_ADD_PTR(struct acpi_iort_id_mapping, node,
node              661 drivers/acpi/arm64/iort.c 			   node->mapping_offset + index * sizeof(*map));
node              667 drivers/acpi/arm64/iort.c 		       node, node->type);
node              698 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node, *msi_parent = NULL;
node              704 drivers/acpi/arm64/iort.c 	node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT,
node              706 drivers/acpi/arm64/iort.c 	if (!node)
node              710 drivers/acpi/arm64/iort.c 	for (i = 0; i < node->mapping_count; i++) {
node              711 drivers/acpi/arm64/iort.c 		msi_parent = iort_node_map_platform_id(node, NULL,
node              874 drivers/acpi/arm64/iort.c static bool iort_pci_rc_supports_ats(struct acpi_iort_node *node)
node              878 drivers/acpi/arm64/iort.c 	pci_rc = (struct acpi_iort_root_complex *)node->node_data;
node              882 drivers/acpi/arm64/iort.c static int iort_iommu_xlate(struct device *dev, struct acpi_iort_node *node,
node              888 drivers/acpi/arm64/iort.c 	if (!node)
node              891 drivers/acpi/arm64/iort.c 	iort_fwnode = iort_get_fwnode(node);
node              905 drivers/acpi/arm64/iort.c 		return iort_iommu_driver_enabled(node->type) ?
node              913 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node              922 drivers/acpi/arm64/iort.c 	parent = iort_node_map_id(info->node, alias, &streamid,
node              937 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node, *parent;
node              954 drivers/acpi/arm64/iort.c 		node = iort_scan_node(ACPI_IORT_NODE_PCI_ROOT_COMPLEX,
node              956 drivers/acpi/arm64/iort.c 		if (!node)
node              959 drivers/acpi/arm64/iort.c 		info.node = node;
node              963 drivers/acpi/arm64/iort.c 		if (!err && iort_pci_rc_supports_ats(node))
node              968 drivers/acpi/arm64/iort.c 		node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT,
node              970 drivers/acpi/arm64/iort.c 		if (!node)
node              974 drivers/acpi/arm64/iort.c 			parent = iort_node_map_platform_id(node, &streamid,
node             1016 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node             1019 drivers/acpi/arm64/iort.c 	node = iort_scan_node(ACPI_IORT_NODE_NAMED_COMPONENT,
node             1021 drivers/acpi/arm64/iort.c 	if (!node)
node             1024 drivers/acpi/arm64/iort.c 	ncomp = (struct acpi_iort_named_component *)node->node_data;
node             1034 drivers/acpi/arm64/iort.c 	struct acpi_iort_node *node;
node             1038 drivers/acpi/arm64/iort.c 	node = iort_scan_node(ACPI_IORT_NODE_PCI_ROOT_COMPLEX,
node             1040 drivers/acpi/arm64/iort.c 	if (!node || node->revision < 1)
node             1043 drivers/acpi/arm64/iort.c 	rc = (struct acpi_iort_root_complex *)node->node_data;
node             1131 drivers/acpi/arm64/iort.c static int __init arm_smmu_v3_count_resources(struct acpi_iort_node *node)
node             1138 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node             1186 drivers/acpi/arm64/iort.c 					      struct acpi_iort_node *node)
node             1192 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node             1230 drivers/acpi/arm64/iort.c 					     struct acpi_iort_node *node)
node             1236 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node             1253 drivers/acpi/arm64/iort.c 					      struct acpi_iort_node *node)
node             1257 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node             1275 drivers/acpi/arm64/iort.c static int __init arm_smmu_count_resources(struct acpi_iort_node *node)
node             1280 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu *)node->node_data;
node             1294 drivers/acpi/arm64/iort.c 					   struct acpi_iort_node *node)
node             1301 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu *)node->node_data;
node             1308 drivers/acpi/arm64/iort.c 	glb_irq = ACPI_ADD_PTR(u64, node, smmu->global_interrupt_offset);
node             1317 drivers/acpi/arm64/iort.c 	ctx_irq = ACPI_ADD_PTR(u64, node, smmu->context_interrupt_offset);
node             1328 drivers/acpi/arm64/iort.c 					  struct acpi_iort_node *node)
node             1334 drivers/acpi/arm64/iort.c 	smmu = (struct acpi_iort_smmu *)node->node_data;
node             1346 drivers/acpi/arm64/iort.c static int __init arm_smmu_v3_pmcg_count_resources(struct acpi_iort_node *node)
node             1351 drivers/acpi/arm64/iort.c 	pmcg = (struct acpi_iort_pmcg *)node->node_data;
node             1361 drivers/acpi/arm64/iort.c 						   struct acpi_iort_node *node)
node             1366 drivers/acpi/arm64/iort.c 	pmcg = (struct acpi_iort_pmcg *)node->node_data;
node             1403 drivers/acpi/arm64/iort.c 	int (*dev_init)(struct acpi_iort_node *node);
node             1405 drivers/acpi/arm64/iort.c 				  struct acpi_iort_node *node);
node             1406 drivers/acpi/arm64/iort.c 	int (*dev_count_resources)(struct acpi_iort_node *node);
node             1408 drivers/acpi/arm64/iort.c 				     struct acpi_iort_node *node);
node             1410 drivers/acpi/arm64/iort.c 				    struct acpi_iort_node *node);
node             1437 drivers/acpi/arm64/iort.c 			struct acpi_iort_node *node)
node             1439 drivers/acpi/arm64/iort.c 	switch (node->type) {
node             1457 drivers/acpi/arm64/iort.c static int __init iort_add_platform_device(struct acpi_iort_node *node,
node             1470 drivers/acpi/arm64/iort.c 		ret = ops->dev_set_proximity(&pdev->dev, node);
node             1475 drivers/acpi/arm64/iort.c 	count = ops->dev_count_resources(node);
node             1483 drivers/acpi/arm64/iort.c 	ops->dev_init_resources(r, node);
node             1504 drivers/acpi/arm64/iort.c 		ret = platform_device_add_data(pdev, &node, sizeof(node));
node             1509 drivers/acpi/arm64/iort.c 	fwnode = iort_get_fwnode(node);
node             1519 drivers/acpi/arm64/iort.c 		ops->dev_dma_configure(&pdev->dev, node);
node             1521 drivers/acpi/arm64/iort.c 	iort_set_device_domain(&pdev->dev, node);
node              544 drivers/acpi/bus.c 		const struct acpi_device_physical_node *node;
node              546 drivers/acpi/bus.c 		node = list_first_entry(&adev->physical_node_list,
node              547 drivers/acpi/bus.c 					struct acpi_device_physical_node, node);
node              549 drivers/acpi/bus.c 		phys_dev = node->dev;
node               35 drivers/acpi/container.c 	list_for_each_entry(child, &adev->children, node)
node              146 drivers/acpi/ec.c 	struct list_head node;
node             1061 drivers/acpi/ec.c 	list_for_each_entry(handler, &ec->list, node) {
node             1100 drivers/acpi/ec.c 	list_add(&handler->node, &ec->list);
node             1113 drivers/acpi/ec.c 	list_for_each_entry_safe(handler, tmp, &ec->list, node) {
node             1115 drivers/acpi/ec.c 			list_del_init(&handler->node);
node             1116 drivers/acpi/ec.c 			list_add(&handler->node, &free_list);
node             1120 drivers/acpi/ec.c 	list_for_each_entry_safe(handler, tmp, &free_list, node)
node               49 drivers/acpi/evged.c 	struct list_head node;
node              137 drivers/acpi/evged.c 	list_add_tail(&event->node, &geddev->event_list);
node              168 drivers/acpi/evged.c 	list_for_each_entry_safe(event, next, &geddev->event_list, node) {
node              170 drivers/acpi/evged.c 		list_del(&event->node);
node              123 drivers/acpi/glue.c 	list_for_each_entry(adev, &parent->children, node) {
node              209 drivers/acpi/glue.c 	list_for_each_entry(pn, &acpi_dev->physical_node_list, node) {
node              224 drivers/acpi/glue.c 			physnode_list = &pn->node;
node              231 drivers/acpi/glue.c 	list_add(&physical_node->node, physnode_list);
node              275 drivers/acpi/glue.c 	list_for_each_entry(entry, &acpi_dev->physical_node_list, node)
node              279 drivers/acpi/glue.c 			list_del(&entry->node);
node               44 drivers/acpi/hmat/hmat.c 	struct list_head node;
node               49 drivers/acpi/hmat/hmat.c 	struct list_head node;
node               59 drivers/acpi/hmat/hmat.c 	struct list_head node;
node               64 drivers/acpi/hmat/hmat.c 	struct list_head node;
node               72 drivers/acpi/hmat/hmat.c 	list_for_each_entry(initiator, &initiators, node)
node               82 drivers/acpi/hmat/hmat.c 	list_for_each_entry(target, &targets, node)
node              104 drivers/acpi/hmat/hmat.c 	list_add_tail(&initiator->node, &initiators);
node              121 drivers/acpi/hmat/hmat.c 	list_add_tail(&target->node, &targets);
node              236 drivers/acpi/hmat/hmat.c 	list_add_tail(&loc->node, &localities);
node              381 drivers/acpi/hmat/hmat.c 	list_add_tail(&tcache->node, &target->caches);
node              532 drivers/acpi/hmat/hmat.c 	ia = list_entry(a, struct memory_initiator, node);
node              533 drivers/acpi/hmat/hmat.c 	ib = list_entry(b, struct memory_initiator, node);
node              579 drivers/acpi/hmat/hmat.c 		list_for_each_entry(initiator, &initiators, node) {
node              606 drivers/acpi/hmat/hmat.c 	list_for_each_entry(tcache, &target->caches, node)
node              644 drivers/acpi/hmat/hmat.c 	list_for_each_entry(target, &targets, node)
node              679 drivers/acpi/hmat/hmat.c 	list_for_each_entry_safe(target, tnext, &targets, node) {
node              680 drivers/acpi/hmat/hmat.c 		list_for_each_entry_safe(tcache, cnext, &target->caches, node) {
node              681 drivers/acpi/hmat/hmat.c 			list_del(&tcache->node);
node              684 drivers/acpi/hmat/hmat.c 		list_del(&target->node);
node              688 drivers/acpi/hmat/hmat.c 	list_for_each_entry_safe(initiator, inext, &initiators, node) {
node              689 drivers/acpi/hmat/hmat.c 		list_del(&initiator->node);
node              693 drivers/acpi/hmat/hmat.c 	list_for_each_entry_safe(loc, lnext, &localities, node) {
node              694 drivers/acpi/hmat/hmat.c 		list_del(&loc->node);
node              102 drivers/acpi/internal.h 	struct list_head node;
node               39 drivers/acpi/numa.c int node_to_pxm(int node)
node               41 drivers/acpi/numa.c 	if (node < 0)
node               43 drivers/acpi/numa.c 	return node_to_pxm_map[node];
node               46 drivers/acpi/numa.c static void __acpi_map_pxm_to_node(int pxm, int node)
node               48 drivers/acpi/numa.c 	if (pxm_to_node_map[pxm] == NUMA_NO_NODE || node < pxm_to_node_map[pxm])
node               49 drivers/acpi/numa.c 		pxm_to_node_map[pxm] = node;
node               50 drivers/acpi/numa.c 	if (node_to_pxm_map[node] == PXM_INVAL || pxm < node_to_pxm_map[node])
node               51 drivers/acpi/numa.c 		node_to_pxm_map[node] = pxm;
node               56 drivers/acpi/numa.c 	int node;
node               61 drivers/acpi/numa.c 	node = pxm_to_node_map[pxm];
node               63 drivers/acpi/numa.c 	if (node == NUMA_NO_NODE) {
node               66 drivers/acpi/numa.c 		node = first_unset_node(nodes_found_map);
node               67 drivers/acpi/numa.c 		__acpi_map_pxm_to_node(pxm, node);
node               68 drivers/acpi/numa.c 		node_set(node, nodes_found_map);
node               71 drivers/acpi/numa.c 	return node;
node               92 drivers/acpi/numa.c 	int node, min_node;
node               94 drivers/acpi/numa.c 	node = acpi_map_pxm_to_node(pxm);
node               96 drivers/acpi/numa.c 	if (node == NUMA_NO_NODE)
node               97 drivers/acpi/numa.c 		node = 0;
node               99 drivers/acpi/numa.c 	min_node = node;
node              100 drivers/acpi/numa.c 	if (!node_online(node)) {
node              104 drivers/acpi/numa.c 			dist = node_distance(node, n);
node              251 drivers/acpi/numa.c 	int node, pxm;
node              272 drivers/acpi/numa.c 	node = acpi_map_pxm_to_node(pxm);
node              273 drivers/acpi/numa.c 	if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) {
node              278 drivers/acpi/numa.c 	if (numa_add_memblk(node, start, end) < 0) {
node              280 drivers/acpi/numa.c 		       node, (unsigned long long) start,
node              285 drivers/acpi/numa.c 	node_set(node, numa_nodes_parsed);
node              288 drivers/acpi/numa.c 		node, pxm,
node               22 drivers/acpi/nvs.c 	struct list_head node;
node               45 drivers/acpi/nvs.c 	list_add_tail(&region->node, &nvs_region_list);
node               56 drivers/acpi/nvs.c 	list_for_each_entry(region, &nvs_region_list, node) {
node               79 drivers/acpi/nvs.c 	struct list_head node;
node              107 drivers/acpi/nvs.c 		list_add_tail(&entry->node, &nvs_list);
node              118 drivers/acpi/nvs.c 	list_for_each_entry_safe(entry, next, &nvs_list, node) {
node              119 drivers/acpi/nvs.c 		list_del(&entry->node);
node              132 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
node              156 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node) {
node              175 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
node              207 drivers/acpi/nvs.c 	list_for_each_entry(entry, &nvs_list, node)
node               81 drivers/acpi/pci_mcfg.c #define THUNDER_PEM_RES(addr, node) \
node               82 drivers/acpi/pci_mcfg.c 	DEFINE_RES_MEM((addr) + ((u64) (node) << 44), 0x39 * SZ_16M)
node               84 drivers/acpi/pci_mcfg.c #define THUNDER_PEM_QUIRK(rev, node) \
node               85 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 4 + (10 * (node)), MCFG_BUS_ANY,	    \
node               86 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x88001f000000UL, node) },  \
node               87 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 5 + (10 * (node)), MCFG_BUS_ANY,	    \
node               88 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x884057000000UL, node) },  \
node               89 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 6 + (10 * (node)), MCFG_BUS_ANY,	    \
node               90 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x88808f000000UL, node) },  \
node               91 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 7 + (10 * (node)), MCFG_BUS_ANY,	    \
node               92 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x89001f000000UL, node) },  \
node               93 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 8 + (10 * (node)), MCFG_BUS_ANY,	    \
node               94 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x894057000000UL, node) },  \
node               95 drivers/acpi/pci_mcfg.c 	{ "CAVIUM", "THUNDERX", rev, 9 + (10 * (node)), MCFG_BUS_ANY,	    \
node               96 drivers/acpi/pci_mcfg.c 	  &thunder_pem_ecam_ops, THUNDER_PEM_RES(0x89808f000000UL, node) }
node              250 drivers/acpi/pci_root.c 	struct list_head node;
node              274 drivers/acpi/pci_root.c 	struct acpi_handle_node *node, *tmp;
node              283 drivers/acpi/pci_root.c 		node = kzalloc(sizeof(struct acpi_handle_node), GFP_KERNEL);
node              284 drivers/acpi/pci_root.c 		if (!node)
node              287 drivers/acpi/pci_root.c 		INIT_LIST_HEAD(&node->node);
node              288 drivers/acpi/pci_root.c 		node->handle = phandle;
node              289 drivers/acpi/pci_root.c 		list_add(&node->node, &device_list);
node              307 drivers/acpi/pci_root.c 	list_for_each_entry(node, &device_list, node) {
node              308 drivers/acpi/pci_root.c 		acpi_handle hnd = node->handle;
node              334 drivers/acpi/pci_root.c 	list_for_each_entry_safe(node, tmp, &device_list, node)
node              335 drivers/acpi/pci_root.c 		kfree(node);
node              880 drivers/acpi/pci_root.c 	int node = acpi_get_node(device->handle);
node              934 drivers/acpi/pci_root.c 	if (node != NUMA_NO_NODE)
node              935 drivers/acpi/pci_root.c 		dev_printk(KERN_DEBUG, &bus->dev, "on NUMA node %d\n", node);
node               47 drivers/acpi/power.c 	struct list_head node;
node               63 drivers/acpi/power.c 	struct list_head node;
node              107 drivers/acpi/power.c 		list_for_each_entry(e, list, node)
node              109 drivers/acpi/power.c 				list_add_tail(&entry->node, &e->node);
node              113 drivers/acpi/power.c 	list_add_tail(&entry->node, list);
node              121 drivers/acpi/power.c 	list_for_each_entry_safe(entry, e, list, node) {
node              122 drivers/acpi/power.c 		list_del(&entry->node);
node              219 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
node              249 drivers/acpi/power.c 	list_for_each_entry(dep, &resource->dependents, node) {
node              262 drivers/acpi/power.c 	list_add_tail(&dep->node, &resource->dependents);
node              277 drivers/acpi/power.c 	list_for_each_entry(dep, &resource->dependents, node) {
node              279 drivers/acpi/power.c 			list_del(&dep->node);
node              315 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node) {
node              324 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node)
node              349 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, resources, node)
node              374 drivers/acpi/power.c 	list_for_each_entry(dep, &resource->dependents, node) {
node              461 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, list, node) {
node              469 drivers/acpi/power.c 	list_for_each_entry_continue(entry, list, node)
node              480 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
node              488 drivers/acpi/power.c 	list_for_each_entry_continue_reverse(entry, list, node)
node              531 drivers/acpi/power.c 	list_for_each_entry_reverse(entry, resources, node) {
node              555 drivers/acpi/power.c 	list_for_each_entry(entry, resources, node) {
node              602 drivers/acpi/power.c 	list_for_each_entry(entry, list, node) {
node              712 drivers/acpi/power.c 	list_for_each_entry(entry, &dev->wakeup.resources, node) {
node              775 drivers/acpi/power.c 	list_for_each_entry(entry, &dev->wakeup.resources, node) {
node               60 drivers/acpi/pptt.c 							   struct acpi_pptt_processor *node,
node               65 drivers/acpi/pptt.c 	if (resource >= node->number_of_priv_resources)
node               68 drivers/acpi/pptt.c 	ref = ACPI_ADD_PTR(u32, node, sizeof(struct acpi_pptt_processor));
node              204 drivers/acpi/pptt.c 			       struct acpi_pptt_processor *node)
node              213 drivers/acpi/pptt.c 		return (node->flags & ACPI_PPTT_ACPI_LEAF_NODE);
node              216 drivers/acpi/pptt.c 	node_entry = ACPI_PTR_DIFF(node, table_hdr);
node              322 drivers/acpi/pptt.c 						    struct acpi_pptt_processor **node)
node              337 drivers/acpi/pptt.c 		*node = cpu_node;
node               25 drivers/acpi/proc.c 	struct list_head *node, *next;
node               30 drivers/acpi/proc.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
node               32 drivers/acpi/proc.c 		    container_of(node, struct acpi_device, wakeup_list);
node               52 drivers/acpi/proc.c 					node) {
node               57 drivers/acpi/proc.c 				if (&entry->node !=
node               85 drivers/acpi/proc.c 		&adev->physical_node_list, node)
node               99 drivers/acpi/proc.c 	struct list_head *node, *next;
node              112 drivers/acpi/proc.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
node              114 drivers/acpi/proc.c 		    container_of(node, struct acpi_device, wakeup_list);
node             1034 drivers/acpi/property.c 			next = adev->node.next;
node             1039 drivers/acpi/property.c 			child_adev = list_entry(next, struct acpi_device, node);
node             1042 drivers/acpi/property.c 						      node);
node              732 drivers/acpi/resource.c 	list_for_each_entry(rentry, &resource_list, node) {
node               54 drivers/acpi/scan.c 	struct list_head node;
node              128 drivers/acpi/scan.c 	list_for_each_entry(pn, &adev->physical_node_list, node)
node              159 drivers/acpi/scan.c 	list_for_each_entry(pn, &device->physical_node_list, node) {
node              197 drivers/acpi/scan.c 	list_for_each_entry(pn, &device->physical_node_list, node)
node              362 drivers/acpi/scan.c 	list_for_each_entry(child, &adev->children, node) {
node              481 drivers/acpi/scan.c 		list_del(&device->node);
node              483 drivers/acpi/scan.c 	list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node)
node              489 drivers/acpi/scan.c 				list_del(&acpi_device_bus_id->node);
node              649 drivers/acpi/scan.c 	INIT_LIST_HEAD(&device->node);
node              667 drivers/acpi/scan.c 	list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node) {
node              680 drivers/acpi/scan.c 		list_add_tail(&acpi_device_bus_id->node, &acpi_bus_id_list);
node              685 drivers/acpi/scan.c 		list_add_tail(&device->node, &device->parent->children);
node              711 drivers/acpi/scan.c 		list_del(&device->node);
node             1425 drivers/acpi/scan.c 		list_for_each_entry(rentry, &list, node) {
node             1857 drivers/acpi/scan.c 		list_add_tail(&dep->node , &acpi_dep_list);
node             2015 drivers/acpi/scan.c 	list_for_each_entry(child, &device->children, node)
node             2028 drivers/acpi/scan.c 	list_for_each_entry_safe(dep, tmp, &acpi_dep_list, node) {
node             2037 drivers/acpi/scan.c 			list_del(&dep->node);
node             2086 drivers/acpi/scan.c 	list_for_each_entry_reverse(child, &adev->children, node)
node              333 drivers/acpi/sysfs.c 	struct list_head node;
node              373 drivers/acpi/sysfs.c 	list_for_each_entry(attr, &acpi_table_attr_list, node) {
node              419 drivers/acpi/sysfs.c 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
node              539 drivers/acpi/sysfs.c 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
node              716 drivers/acpi/utils.c 	list_for_each_entry(acpi_device_bus_id, &acpi_bus_id_list, node)
node               42 drivers/acpi/wakeup.c 	struct list_head *node, *next;
node               44 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
node               46 drivers/acpi/wakeup.c 			container_of(node, struct acpi_device, wakeup_list);
node               69 drivers/acpi/wakeup.c 	struct list_head *node, *next;
node               71 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
node               73 drivers/acpi/wakeup.c 			container_of(node, struct acpi_device, wakeup_list);
node               91 drivers/acpi/wakeup.c 	struct list_head *node, *next;
node               94 drivers/acpi/wakeup.c 	list_for_each_safe(node, next, &acpi_wakeup_device_list) {
node               95 drivers/acpi/wakeup.c 		struct acpi_device *dev = container_of(node,
node              497 drivers/amba/bus.c 	struct list_head node;
node              514 drivers/amba/bus.c 	list_for_each_entry_safe(ddev, tmp, &deferred_devices, node) {
node              520 drivers/amba/bus.c 		list_del_init(&ddev->node);
node              560 drivers/amba/bus.c 		list_add_tail(&ddev->node, &deferred_devices);
node              400 drivers/android/binder.c 	struct binder_node *node;
node              691 drivers/android/binder.c #define binder_node_lock(node) _binder_node_lock(node, __LINE__)
node              693 drivers/android/binder.c _binder_node_lock(struct binder_node *node, int line)
node              694 drivers/android/binder.c 	__acquires(&node->lock)
node              698 drivers/android/binder.c 	spin_lock(&node->lock);
node              707 drivers/android/binder.c #define binder_node_unlock(node) _binder_node_unlock(node, __LINE__)
node              709 drivers/android/binder.c _binder_node_unlock(struct binder_node *node, int line)
node              710 drivers/android/binder.c 	__releases(&node->lock)
node              714 drivers/android/binder.c 	spin_unlock(&node->lock);
node              724 drivers/android/binder.c #define binder_node_inner_lock(node) _binder_node_inner_lock(node, __LINE__)
node              726 drivers/android/binder.c _binder_node_inner_lock(struct binder_node *node, int line)
node              727 drivers/android/binder.c 	__acquires(&node->lock) __acquires(&node->proc->inner_lock)
node              731 drivers/android/binder.c 	spin_lock(&node->lock);
node              732 drivers/android/binder.c 	if (node->proc)
node              733 drivers/android/binder.c 		binder_inner_proc_lock(node->proc);
node              736 drivers/android/binder.c 		__acquire(&node->proc->inner_lock);
node              745 drivers/android/binder.c #define binder_node_inner_unlock(node) _binder_node_inner_unlock(node, __LINE__)
node              747 drivers/android/binder.c _binder_node_inner_unlock(struct binder_node *node, int line)
node              748 drivers/android/binder.c 	__releases(&node->lock) __releases(&node->proc->inner_lock)
node              750 drivers/android/binder.c 	struct binder_proc *proc = node->proc;
node              758 drivers/android/binder.c 		__release(&node->proc->inner_lock);
node              759 drivers/android/binder.c 	spin_unlock(&node->lock);
node              917 drivers/android/binder.c static void binder_inc_node_tmpref_ilocked(struct binder_node *node);
node             1068 drivers/android/binder.c 	struct binder_node *node;
node             1073 drivers/android/binder.c 		node = rb_entry(n, struct binder_node, rb_node);
node             1075 drivers/android/binder.c 		if (ptr < node->ptr)
node             1077 drivers/android/binder.c 		else if (ptr > node->ptr)
node             1085 drivers/android/binder.c 			binder_inc_node_tmpref_ilocked(node);
node             1086 drivers/android/binder.c 			return node;
node             1095 drivers/android/binder.c 	struct binder_node *node;
node             1098 drivers/android/binder.c 	node = binder_get_node_ilocked(proc, ptr);
node             1100 drivers/android/binder.c 	return node;
node             1110 drivers/android/binder.c 	struct binder_node *node;
node             1120 drivers/android/binder.c 		node = rb_entry(parent, struct binder_node, rb_node);
node             1122 drivers/android/binder.c 		if (ptr < node->ptr)
node             1124 drivers/android/binder.c 		else if (ptr > node->ptr)
node             1132 drivers/android/binder.c 			binder_inc_node_tmpref_ilocked(node);
node             1133 drivers/android/binder.c 			return node;
node             1136 drivers/android/binder.c 	node = new_node;
node             1138 drivers/android/binder.c 	node->tmp_refs++;
node             1139 drivers/android/binder.c 	rb_link_node(&node->rb_node, parent, p);
node             1140 drivers/android/binder.c 	rb_insert_color(&node->rb_node, &proc->nodes);
node             1141 drivers/android/binder.c 	node->debug_id = atomic_inc_return(&binder_last_id);
node             1142 drivers/android/binder.c 	node->proc = proc;
node             1143 drivers/android/binder.c 	node->ptr = ptr;
node             1144 drivers/android/binder.c 	node->cookie = cookie;
node             1145 drivers/android/binder.c 	node->work.type = BINDER_WORK_NODE;
node             1146 drivers/android/binder.c 	node->min_priority = flags & FLAT_BINDER_FLAG_PRIORITY_MASK;
node             1147 drivers/android/binder.c 	node->accept_fds = !!(flags & FLAT_BINDER_FLAG_ACCEPTS_FDS);
node             1148 drivers/android/binder.c 	node->txn_security_ctx = !!(flags & FLAT_BINDER_FLAG_TXN_SECURITY_CTX);
node             1149 drivers/android/binder.c 	spin_lock_init(&node->lock);
node             1150 drivers/android/binder.c 	INIT_LIST_HEAD(&node->work.entry);
node             1151 drivers/android/binder.c 	INIT_LIST_HEAD(&node->async_todo);
node             1154 drivers/android/binder.c 		     proc->pid, current->pid, node->debug_id,
node             1155 drivers/android/binder.c 		     (u64)node->ptr, (u64)node->cookie);
node             1157 drivers/android/binder.c 	return node;
node             1163 drivers/android/binder.c 	struct binder_node *node;
node             1164 drivers/android/binder.c 	struct binder_node *new_node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1169 drivers/android/binder.c 	node = binder_init_node_ilocked(proc, new_node, fp);
node             1171 drivers/android/binder.c 	if (node != new_node)
node             1177 drivers/android/binder.c 	return node;
node             1180 drivers/android/binder.c static void binder_free_node(struct binder_node *node)
node             1182 drivers/android/binder.c 	kfree(node);
node             1186 drivers/android/binder.c static int binder_inc_node_nilocked(struct binder_node *node, int strong,
node             1190 drivers/android/binder.c 	struct binder_proc *proc = node->proc;
node             1192 drivers/android/binder.c 	assert_spin_locked(&node->lock);
node             1198 drivers/android/binder.c 			    node->internal_strong_refs == 0 &&
node             1199 drivers/android/binder.c 			    !(node->proc &&
node             1200 drivers/android/binder.c 			      node == node->proc->context->binder_context_mgr_node &&
node             1201 drivers/android/binder.c 			      node->has_strong_ref)) {
node             1203 drivers/android/binder.c 					node->debug_id);
node             1206 drivers/android/binder.c 			node->internal_strong_refs++;
node             1208 drivers/android/binder.c 			node->local_strong_refs++;
node             1209 drivers/android/binder.c 		if (!node->has_strong_ref && target_list) {
node             1212 drivers/android/binder.c 			binder_dequeue_work_ilocked(&node->work);
node             1215 drivers/android/binder.c 								   &node->work);
node             1219 drivers/android/binder.c 			node->local_weak_refs++;
node             1220 drivers/android/binder.c 		if (!node->has_weak_ref && list_empty(&node->work.entry)) {
node             1223 drivers/android/binder.c 					node->debug_id);
node             1229 drivers/android/binder.c 			binder_enqueue_work_ilocked(&node->work, target_list);
node             1235 drivers/android/binder.c static int binder_inc_node(struct binder_node *node, int strong, int internal,
node             1240 drivers/android/binder.c 	binder_node_inner_lock(node);
node             1241 drivers/android/binder.c 	ret = binder_inc_node_nilocked(node, strong, internal, target_list);
node             1242 drivers/android/binder.c 	binder_node_inner_unlock(node);
node             1247 drivers/android/binder.c static bool binder_dec_node_nilocked(struct binder_node *node,
node             1250 drivers/android/binder.c 	struct binder_proc *proc = node->proc;
node             1252 drivers/android/binder.c 	assert_spin_locked(&node->lock);
node             1257 drivers/android/binder.c 			node->internal_strong_refs--;
node             1259 drivers/android/binder.c 			node->local_strong_refs--;
node             1260 drivers/android/binder.c 		if (node->local_strong_refs || node->internal_strong_refs)
node             1264 drivers/android/binder.c 			node->local_weak_refs--;
node             1265 drivers/android/binder.c 		if (node->local_weak_refs || node->tmp_refs ||
node             1266 drivers/android/binder.c 				!hlist_empty(&node->refs))
node             1270 drivers/android/binder.c 	if (proc && (node->has_strong_ref || node->has_weak_ref)) {
node             1271 drivers/android/binder.c 		if (list_empty(&node->work.entry)) {
node             1272 drivers/android/binder.c 			binder_enqueue_work_ilocked(&node->work, &proc->todo);
node             1276 drivers/android/binder.c 		if (hlist_empty(&node->refs) && !node->local_strong_refs &&
node             1277 drivers/android/binder.c 		    !node->local_weak_refs && !node->tmp_refs) {
node             1279 drivers/android/binder.c 				binder_dequeue_work_ilocked(&node->work);
node             1280 drivers/android/binder.c 				rb_erase(&node->rb_node, &proc->nodes);
node             1283 drivers/android/binder.c 					     node->debug_id);
node             1285 drivers/android/binder.c 				BUG_ON(!list_empty(&node->work.entry));
node             1291 drivers/android/binder.c 				if (node->tmp_refs) {
node             1295 drivers/android/binder.c 				hlist_del(&node->dead_node);
node             1299 drivers/android/binder.c 					     node->debug_id);
node             1307 drivers/android/binder.c static void binder_dec_node(struct binder_node *node, int strong, int internal)
node             1311 drivers/android/binder.c 	binder_node_inner_lock(node);
node             1312 drivers/android/binder.c 	free_node = binder_dec_node_nilocked(node, strong, internal);
node             1313 drivers/android/binder.c 	binder_node_inner_unlock(node);
node             1315 drivers/android/binder.c 		binder_free_node(node);
node             1318 drivers/android/binder.c static void binder_inc_node_tmpref_ilocked(struct binder_node *node)
node             1325 drivers/android/binder.c 	node->tmp_refs++;
node             1341 drivers/android/binder.c static void binder_inc_node_tmpref(struct binder_node *node)
node             1343 drivers/android/binder.c 	binder_node_lock(node);
node             1344 drivers/android/binder.c 	if (node->proc)
node             1345 drivers/android/binder.c 		binder_inner_proc_lock(node->proc);
node             1348 drivers/android/binder.c 	binder_inc_node_tmpref_ilocked(node);
node             1349 drivers/android/binder.c 	if (node->proc)
node             1350 drivers/android/binder.c 		binder_inner_proc_unlock(node->proc);
node             1353 drivers/android/binder.c 	binder_node_unlock(node);
node             1362 drivers/android/binder.c static void binder_dec_node_tmpref(struct binder_node *node)
node             1366 drivers/android/binder.c 	binder_node_inner_lock(node);
node             1367 drivers/android/binder.c 	if (!node->proc)
node             1371 drivers/android/binder.c 	node->tmp_refs--;
node             1372 drivers/android/binder.c 	BUG_ON(node->tmp_refs < 0);
node             1373 drivers/android/binder.c 	if (!node->proc)
node             1383 drivers/android/binder.c 	free_node = binder_dec_node_nilocked(node, 0, 1);
node             1384 drivers/android/binder.c 	binder_node_inner_unlock(node);
node             1386 drivers/android/binder.c 		binder_free_node(node);
node             1389 drivers/android/binder.c static void binder_put_node(struct binder_node *node)
node             1391 drivers/android/binder.c 	binder_dec_node_tmpref(node);
node             1437 drivers/android/binder.c 					struct binder_node *node,
node             1450 drivers/android/binder.c 		if (node < ref->node)
node             1452 drivers/android/binder.c 		else if (node > ref->node)
node             1463 drivers/android/binder.c 	new_ref->node = node;
node             1467 drivers/android/binder.c 	new_ref->data.desc = (node == context->binder_context_mgr_node) ? 0 : 1;
node             1490 drivers/android/binder.c 	binder_node_lock(node);
node             1491 drivers/android/binder.c 	hlist_add_head(&new_ref->node_entry, &node->refs);
node             1496 drivers/android/binder.c 		      node->debug_id);
node             1497 drivers/android/binder.c 	binder_node_unlock(node);
node             1508 drivers/android/binder.c 		      ref->node->debug_id);
node             1513 drivers/android/binder.c 	binder_node_inner_lock(ref->node);
node             1515 drivers/android/binder.c 		binder_dec_node_nilocked(ref->node, 1, 1);
node             1518 drivers/android/binder.c 	delete_node = binder_dec_node_nilocked(ref->node, 0, 1);
node             1519 drivers/android/binder.c 	binder_node_inner_unlock(ref->node);
node             1529 drivers/android/binder.c 		ref->node = NULL;
node             1560 drivers/android/binder.c 			ret = binder_inc_node(ref->node, 1, 1, target_list);
node             1567 drivers/android/binder.c 			ret = binder_inc_node(ref->node, 0, 1, target_list);
node             1597 drivers/android/binder.c 			binder_dec_node(ref->node, strong, 1);
node             1631 drivers/android/binder.c 	struct binder_node *node;
node             1638 drivers/android/binder.c 	node = ref->node;
node             1643 drivers/android/binder.c 	binder_inc_node_tmpref(node);
node             1648 drivers/android/binder.c 	return node;
node             1664 drivers/android/binder.c 	if (ref->node)
node             1665 drivers/android/binder.c 		binder_free_node(ref->node);
node             1747 drivers/android/binder.c 			struct binder_node *node,
node             1757 drivers/android/binder.c 	ref = binder_get_ref_for_node_olocked(proc, node, NULL);
node             1764 drivers/android/binder.c 		ref = binder_get_ref_for_node_olocked(proc, node, new_ref);
node             2304 drivers/android/binder.c 			struct binder_node *node;
node             2307 drivers/android/binder.c 			node = binder_get_node(proc, fp->binder);
node             2308 drivers/android/binder.c 			if (node == NULL) {
node             2315 drivers/android/binder.c 				     node->debug_id, (u64)node->ptr);
node             2316 drivers/android/binder.c 			binder_dec_node(node, hdr->type == BINDER_TYPE_BINDER,
node             2318 drivers/android/binder.c 			binder_put_node(node);
node             2440 drivers/android/binder.c 	struct binder_node *node;
node             2446 drivers/android/binder.c 	node = binder_get_node(proc, fp->binder);
node             2447 drivers/android/binder.c 	if (!node) {
node             2448 drivers/android/binder.c 		node = binder_new_node(proc, fp);
node             2449 drivers/android/binder.c 		if (!node)
node             2452 drivers/android/binder.c 	if (fp->cookie != node->cookie) {
node             2455 drivers/android/binder.c 				  node->debug_id, (u64)fp->cookie,
node             2456 drivers/android/binder.c 				  (u64)node->cookie);
node             2465 drivers/android/binder.c 	ret = binder_inc_ref_for_node(target_proc, node,
node             2479 drivers/android/binder.c 	trace_binder_transaction_node_to_ref(t, node, &rdata);
node             2482 drivers/android/binder.c 		     node->debug_id, (u64)node->ptr,
node             2485 drivers/android/binder.c 	binder_put_node(node);
node             2495 drivers/android/binder.c 	struct binder_node *node;
node             2499 drivers/android/binder.c 	node = binder_get_node_from_ref(proc, fp->handle,
node             2501 drivers/android/binder.c 	if (!node) {
node             2511 drivers/android/binder.c 	binder_node_lock(node);
node             2512 drivers/android/binder.c 	if (node->proc == target_proc) {
node             2517 drivers/android/binder.c 		fp->binder = node->ptr;
node             2518 drivers/android/binder.c 		fp->cookie = node->cookie;
node             2519 drivers/android/binder.c 		if (node->proc)
node             2520 drivers/android/binder.c 			binder_inner_proc_lock(node->proc);
node             2522 drivers/android/binder.c 			__acquire(&node->proc->inner_lock);
node             2523 drivers/android/binder.c 		binder_inc_node_nilocked(node,
node             2526 drivers/android/binder.c 		if (node->proc)
node             2527 drivers/android/binder.c 			binder_inner_proc_unlock(node->proc);
node             2529 drivers/android/binder.c 			__release(&node->proc->inner_lock);
node             2530 drivers/android/binder.c 		trace_binder_transaction_ref_to_node(t, node, &src_rdata);
node             2533 drivers/android/binder.c 			     src_rdata.debug_id, src_rdata.desc, node->debug_id,
node             2534 drivers/android/binder.c 			     (u64)node->ptr);
node             2535 drivers/android/binder.c 		binder_node_unlock(node);
node             2539 drivers/android/binder.c 		binder_node_unlock(node);
node             2540 drivers/android/binder.c 		ret = binder_inc_ref_for_node(target_proc, node,
node             2549 drivers/android/binder.c 		trace_binder_transaction_ref_to_ref(t, node, &src_rdata,
node             2555 drivers/android/binder.c 			     node->debug_id);
node             2558 drivers/android/binder.c 	binder_put_node(node);
node             2757 drivers/android/binder.c 	struct binder_node *node = t->buffer->target_node;
node             2761 drivers/android/binder.c 	BUG_ON(!node);
node             2762 drivers/android/binder.c 	binder_node_lock(node);
node             2765 drivers/android/binder.c 		if (node->has_async_transaction) {
node             2768 drivers/android/binder.c 			node->has_async_transaction = true;
node             2776 drivers/android/binder.c 		binder_node_unlock(node);
node             2788 drivers/android/binder.c 		binder_enqueue_work_ilocked(&t->work, &node->async_todo);
node             2794 drivers/android/binder.c 	binder_node_unlock(node);
node             2821 drivers/android/binder.c 		struct binder_node *node,
node             2827 drivers/android/binder.c 	binder_node_inner_lock(node);
node             2828 drivers/android/binder.c 	if (node->proc) {
node             2829 drivers/android/binder.c 		target_node = node;
node             2830 drivers/android/binder.c 		binder_inc_node_nilocked(node, 1, 0, NULL);
node             2831 drivers/android/binder.c 		binder_inc_node_tmpref_ilocked(node);
node             2832 drivers/android/binder.c 		node->proc->tmp_ref++;
node             2833 drivers/android/binder.c 		*procp = node->proc;
node             2836 drivers/android/binder.c 	binder_node_inner_unlock(node);
node             2951 drivers/android/binder.c 						ref->node, &target_proc,
node             3687 drivers/android/binder.c 			struct binder_node *node;
node             3696 drivers/android/binder.c 			node = binder_get_node(proc, node_ptr);
node             3697 drivers/android/binder.c 			if (node == NULL) {
node             3706 drivers/android/binder.c 			if (cookie != node->cookie) {
node             3711 drivers/android/binder.c 					(u64)node_ptr, node->debug_id,
node             3712 drivers/android/binder.c 					(u64)cookie, (u64)node->cookie);
node             3713 drivers/android/binder.c 				binder_put_node(node);
node             3716 drivers/android/binder.c 			binder_node_inner_lock(node);
node             3718 drivers/android/binder.c 				if (node->pending_strong_ref == 0) {
node             3721 drivers/android/binder.c 						node->debug_id);
node             3722 drivers/android/binder.c 					binder_node_inner_unlock(node);
node             3723 drivers/android/binder.c 					binder_put_node(node);
node             3726 drivers/android/binder.c 				node->pending_strong_ref = 0;
node             3728 drivers/android/binder.c 				if (node->pending_weak_ref == 0) {
node             3731 drivers/android/binder.c 						node->debug_id);
node             3732 drivers/android/binder.c 					binder_node_inner_unlock(node);
node             3733 drivers/android/binder.c 					binder_put_node(node);
node             3736 drivers/android/binder.c 				node->pending_weak_ref = 0;
node             3738 drivers/android/binder.c 			free_node = binder_dec_node_nilocked(node,
node             3745 drivers/android/binder.c 				     node->debug_id, node->local_strong_refs,
node             3746 drivers/android/binder.c 				     node->local_weak_refs, node->tmp_refs);
node             3747 drivers/android/binder.c 			binder_node_inner_unlock(node);
node             3748 drivers/android/binder.c 			binder_put_node(node);
node             3907 drivers/android/binder.c 				     ref->data.weak, ref->node->debug_id);
node             3909 drivers/android/binder.c 			binder_node_lock(ref->node);
node             3914 drivers/android/binder.c 					binder_node_unlock(ref->node);
node             3923 drivers/android/binder.c 				if (ref->node->proc == NULL) {
node             3936 drivers/android/binder.c 					binder_node_unlock(ref->node);
node             3946 drivers/android/binder.c 					binder_node_unlock(ref->node);
node             3973 drivers/android/binder.c 			binder_node_unlock(ref->node);
node             4293 drivers/android/binder.c 			struct binder_node *node = container_of(w, struct binder_node, work);
node             4295 drivers/android/binder.c 			binder_uintptr_t node_ptr = node->ptr;
node             4296 drivers/android/binder.c 			binder_uintptr_t node_cookie = node->cookie;
node             4297 drivers/android/binder.c 			int node_debug_id = node->debug_id;
node             4302 drivers/android/binder.c 			BUG_ON(proc != node->proc);
node             4303 drivers/android/binder.c 			strong = node->internal_strong_refs ||
node             4304 drivers/android/binder.c 					node->local_strong_refs;
node             4305 drivers/android/binder.c 			weak = !hlist_empty(&node->refs) ||
node             4306 drivers/android/binder.c 					node->local_weak_refs ||
node             4307 drivers/android/binder.c 					node->tmp_refs || strong;
node             4308 drivers/android/binder.c 			has_strong_ref = node->has_strong_ref;
node             4309 drivers/android/binder.c 			has_weak_ref = node->has_weak_ref;
node             4312 drivers/android/binder.c 				node->has_weak_ref = 1;
node             4313 drivers/android/binder.c 				node->pending_weak_ref = 1;
node             4314 drivers/android/binder.c 				node->local_weak_refs++;
node             4317 drivers/android/binder.c 				node->has_strong_ref = 1;
node             4318 drivers/android/binder.c 				node->pending_strong_ref = 1;
node             4319 drivers/android/binder.c 				node->local_strong_refs++;
node             4322 drivers/android/binder.c 				node->has_strong_ref = 0;
node             4324 drivers/android/binder.c 				node->has_weak_ref = 0;
node             4332 drivers/android/binder.c 				rb_erase(&node->rb_node, &proc->nodes);
node             4334 drivers/android/binder.c 				binder_node_lock(node);
node             4344 drivers/android/binder.c 				binder_node_unlock(node);
node             4345 drivers/android/binder.c 				binder_free_node(node);
node             4941 drivers/android/binder.c 	struct binder_node *node;
node             4961 drivers/android/binder.c 	node = binder_get_node_from_ref(proc, handle, true, NULL);
node             4962 drivers/android/binder.c 	if (!node)
node             4965 drivers/android/binder.c 	info->strong_count = node->local_strong_refs +
node             4966 drivers/android/binder.c 		node->internal_strong_refs;
node             4967 drivers/android/binder.c 	info->weak_count = node->local_weak_refs;
node             4969 drivers/android/binder.c 	binder_put_node(node);
node             4984 drivers/android/binder.c 		struct binder_node *node = rb_entry(n, struct binder_node,
node             4986 drivers/android/binder.c 		if (node->ptr > ptr) {
node             4987 drivers/android/binder.c 			info->ptr = node->ptr;
node             4988 drivers/android/binder.c 			info->cookie = node->cookie;
node             4989 drivers/android/binder.c 			info->has_strong_ref = node->has_strong_ref;
node             4990 drivers/android/binder.c 			info->has_weak_ref = node->has_weak_ref;
node             5343 drivers/android/binder.c static int binder_node_release(struct binder_node *node, int refs)
node             5347 drivers/android/binder.c 	struct binder_proc *proc = node->proc;
node             5349 drivers/android/binder.c 	binder_release_work(proc, &node->async_todo);
node             5351 drivers/android/binder.c 	binder_node_lock(node);
node             5353 drivers/android/binder.c 	binder_dequeue_work_ilocked(&node->work);
node             5357 drivers/android/binder.c 	BUG_ON(!node->tmp_refs);
node             5358 drivers/android/binder.c 	if (hlist_empty(&node->refs) && node->tmp_refs == 1) {
node             5360 drivers/android/binder.c 		binder_node_unlock(node);
node             5361 drivers/android/binder.c 		binder_free_node(node);
node             5366 drivers/android/binder.c 	node->proc = NULL;
node             5367 drivers/android/binder.c 	node->local_strong_refs = 0;
node             5368 drivers/android/binder.c 	node->local_weak_refs = 0;
node             5372 drivers/android/binder.c 	hlist_add_head(&node->dead_node, &binder_dead_nodes);
node             5375 drivers/android/binder.c 	hlist_for_each_entry(ref, &node->refs, node_entry) {
node             5401 drivers/android/binder.c 		     node->debug_id, refs, death);
node             5402 drivers/android/binder.c 	binder_node_unlock(node);
node             5403 drivers/android/binder.c 	binder_put_node(node);
node             5457 drivers/android/binder.c 		struct binder_node *node;
node             5459 drivers/android/binder.c 		node = rb_entry(n, struct binder_node, rb_node);
node             5466 drivers/android/binder.c 		binder_inc_node_tmpref_ilocked(node);
node             5467 drivers/android/binder.c 		rb_erase(&node->rb_node, &proc->nodes);
node             5469 drivers/android/binder.c 		incoming_refs = binder_node_release(node, incoming_refs);
node             5587 drivers/android/binder.c 	struct binder_node *node;
node             5607 drivers/android/binder.c 		node = container_of(w, struct binder_node, work);
node             5609 drivers/android/binder.c 			   prefix, node->debug_id,
node             5610 drivers/android/binder.c 			   (u64)node->ptr, (u64)node->cookie);
node             5666 drivers/android/binder.c 				       struct binder_node *node)
node             5673 drivers/android/binder.c 	hlist_for_each_entry(ref, &node->refs, node_entry)
node             5677 drivers/android/binder.c 		   node->debug_id, (u64)node->ptr, (u64)node->cookie,
node             5678 drivers/android/binder.c 		   node->has_strong_ref, node->has_weak_ref,
node             5679 drivers/android/binder.c 		   node->local_strong_refs, node->local_weak_refs,
node             5680 drivers/android/binder.c 		   node->internal_strong_refs, count, node->tmp_refs);
node             5683 drivers/android/binder.c 		hlist_for_each_entry(ref, &node->refs, node_entry)
node             5687 drivers/android/binder.c 	if (node->proc) {
node             5688 drivers/android/binder.c 		list_for_each_entry(w, &node->async_todo, entry)
node             5689 drivers/android/binder.c 			print_binder_work_ilocked(m, node->proc, "    ",
node             5697 drivers/android/binder.c 	binder_node_lock(ref->node);
node             5700 drivers/android/binder.c 		   ref->node->proc ? "" : "dead ",
node             5701 drivers/android/binder.c 		   ref->node->debug_id, ref->data.strong,
node             5703 drivers/android/binder.c 	binder_node_unlock(ref->node);
node             5725 drivers/android/binder.c 		struct binder_node *node = rb_entry(n, struct binder_node,
node             5727 drivers/android/binder.c 		if (!print_all && !node->has_async_transaction)
node             5735 drivers/android/binder.c 		binder_inc_node_tmpref_ilocked(node);
node             5740 drivers/android/binder.c 		binder_node_inner_lock(node);
node             5741 drivers/android/binder.c 		print_binder_node_nilocked(m, node);
node             5742 drivers/android/binder.c 		binder_node_inner_unlock(node);
node             5743 drivers/android/binder.c 		last_node = node;
node             5937 drivers/android/binder.c 	struct binder_node *node;
node             5945 drivers/android/binder.c 	hlist_for_each_entry(node, &binder_dead_nodes, dead_node) {
node             5951 drivers/android/binder.c 		node->tmp_refs++;
node             5955 drivers/android/binder.c 		binder_node_lock(node);
node             5956 drivers/android/binder.c 		print_binder_node_nilocked(m, node);
node             5957 drivers/android/binder.c 		binder_node_unlock(node);
node             5958 drivers/android/binder.c 		last_node = node;
node              140 drivers/android/binder_trace.h 	TP_PROTO(struct binder_transaction *t, struct binder_node *node,
node              142 drivers/android/binder_trace.h 	TP_ARGS(t, node, rdata),
node              153 drivers/android/binder_trace.h 		__entry->node_debug_id = node->debug_id;
node              154 drivers/android/binder_trace.h 		__entry->node_ptr = node->ptr;
node              165 drivers/android/binder_trace.h 	TP_PROTO(struct binder_transaction *t, struct binder_node *node,
node              167 drivers/android/binder_trace.h 	TP_ARGS(t, node, rdata),
node              180 drivers/android/binder_trace.h 		__entry->node_debug_id = node->debug_id;
node              181 drivers/android/binder_trace.h 		__entry->node_ptr = node->ptr;
node              190 drivers/android/binder_trace.h 	TP_PROTO(struct binder_transaction *t, struct binder_node *node,
node              193 drivers/android/binder_trace.h 	TP_ARGS(t, node, src_ref, dest_ref),
node              205 drivers/android/binder_trace.h 		__entry->node_debug_id = node->debug_id;
node               34 drivers/ata/ahci_octeon.c 	struct device_node *node = dev->of_node;
node               62 drivers/ata/ahci_octeon.c 	if (!node) {
node               67 drivers/ata/ahci_octeon.c 	ret = of_platform_populate(node, NULL, NULL, dev);
node              310 drivers/ata/libahci_platform.c 				struct device *dev, struct device_node *node)
node              314 drivers/ata/libahci_platform.c 	hpriv->phys[port] = devm_of_phy_get(dev, node, NULL);
node              323 drivers/ata/libahci_platform.c 		if (of_find_property(node, "phys", NULL)) {
node              326 drivers/ata/libahci_platform.c 				node);
node              342 drivers/ata/libahci_platform.c 			node, rc);
node              231 drivers/ata/pata_macio.c 	struct device_node		*node;
node              482 drivers/ata/pata_macio.c 		const char* cable = of_get_property(priv->node, "cable-type",
node              505 drivers/ata/pata_macio.c 	if (of_device_is_compatible(priv->node, "K2-UATA") ||
node              506 drivers/ata/pata_macio.c 	    of_device_is_compatible(priv->node, "shasta-ata"))
node              753 drivers/ata/pata_macio.c 		ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, priv->node, 0, 1);
node              759 drivers/ata/pata_macio.c 					 priv->node, priv->aapl_bus_id, 1);
node              761 drivers/ata/pata_macio.c 				    priv->node, priv->aapl_bus_id, 1);
node              766 drivers/ata/pata_macio.c 					    priv->node, priv->aapl_bus_id, 0);
node              890 drivers/ata/pata_macio.c 	ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, priv->node,
node              948 drivers/ata/pata_macio.c 	if (of_device_is_compatible(priv->node, "shasta-ata")) {
node              951 drivers/ata/pata_macio.c 	} else if (of_device_is_compatible(priv->node, "kauai-ata")) {
node              954 drivers/ata/pata_macio.c 	} else if (of_device_is_compatible(priv->node, "K2-UATA")) {
node              957 drivers/ata/pata_macio.c 	} else if (of_device_is_compatible(priv->node, "keylargo-ata")) {
node              958 drivers/ata/pata_macio.c 		if (of_node_name_eq(priv->node, "ata-4")) {
node              965 drivers/ata/pata_macio.c 	} else if (of_device_is_compatible(priv->node, "heathrow-ata")) {
node              976 drivers/ata/pata_macio.c 	bidp = of_get_property(priv->node, "AAPL,bus-id", NULL);
node             1139 drivers/ata/pata_macio.c 	priv->node = of_node_get(mdev->ofdev.dev.of_node);
node             1283 drivers/ata/pata_macio.c 	priv->node = of_node_get(np);
node              842 drivers/ata/pata_octeon_cf.c 	struct device_node *node;
node              854 drivers/ata/pata_octeon_cf.c 	node = pdev->dev.of_node;
node              855 drivers/ata/pata_octeon_cf.c 	if (node == NULL)
node              862 drivers/ata/pata_octeon_cf.c 	cf_port->is_true_ide = of_property_read_bool(node, "cavium,true-ide");
node              864 drivers/ata/pata_octeon_cf.c 	if (of_property_read_u32(node, "cavium,bus-width", &bus_width) == 0)
node              869 drivers/ata/pata_octeon_cf.c 	n_addr = of_n_addr_cells(node);
node              870 drivers/ata/pata_octeon_cf.c 	n_size = of_n_size_cells(node);
node              872 drivers/ata/pata_octeon_cf.c 	reg_prop = of_find_property(node, "reg", &reg_len);
node              881 drivers/ata/pata_octeon_cf.c 		dma_node = of_parse_phandle(node,
node              320 drivers/auxdisplay/ht16k33.c 	struct device_node *node = client->dev.of_node;
node              339 drivers/auxdisplay/ht16k33.c 	if (!of_get_property(node, "linux,no-autorepeat", NULL))
node              342 drivers/auxdisplay/ht16k33.c 	err = of_property_read_u32(node, "debounce-delay-ms",
node              398 drivers/auxdisplay/ht16k33.c 	struct device_node *node = client->dev.of_node;
node              440 drivers/auxdisplay/ht16k33.c 	err = of_property_read_u32(node, "refresh-rate-hz",
node              480 drivers/auxdisplay/ht16k33.c 	err = of_property_read_u32(node, "default-brightness-level",
node              251 drivers/base/arch_topology.c static int __init get_cpu_for_node(struct device_node *node)
node              256 drivers/base/arch_topology.c 	cpu_node = of_parse_phandle(node, "cpu", 0);
node               26 drivers/base/attribute_container.c 	struct klist_node node;
node               34 drivers/base/attribute_container.c 		container_of(n, struct internal_container, node);
node               41 drivers/base/attribute_container.c 		container_of(n, struct internal_container, node);
node               75 drivers/base/attribute_container.c 	INIT_LIST_HEAD(&cont->node);
node               80 drivers/base/attribute_container.c 	list_add_tail(&cont->node, &attribute_container_list);
node              102 drivers/base/attribute_container.c 	list_del(&cont->node);
node              149 drivers/base/attribute_container.c 	list_for_each_entry(cont, &attribute_container_list, node) {
node              174 drivers/base/attribute_container.c 		klist_add_tail(&ic->node, &cont->containers);
node              214 drivers/base/attribute_container.c 	list_for_each_entry(cont, &attribute_container_list, node) {
node              224 drivers/base/attribute_container.c 		klist_for_each_entry(ic, &cont->containers, node, &iter) {
node              227 drivers/base/attribute_container.c 			klist_del(&ic->node);
node              258 drivers/base/attribute_container.c 	list_for_each_entry(cont, &attribute_container_list, node) {
node              270 drivers/base/attribute_container.c 		klist_for_each_entry(ic, &cont->containers, node, &iter) {
node              298 drivers/base/attribute_container.c 	list_for_each_entry(cont, &attribute_container_list, node) {
node              430 drivers/base/attribute_container.c 	klist_for_each_entry(ic, &cont->containers, node, &iter) {
node              493 drivers/base/bus.c 	list_for_each_entry(sif, &bus->p->interfaces, node)
node              518 drivers/base/bus.c 	list_for_each_entry(sif, &bus->p->interfaces, node)
node             1053 drivers/base/bus.c 	list_add_tail(&sif->node, &subsys->p->interfaces);
node             1078 drivers/base/bus.c 	list_del_init(&sif->node);
node              440 drivers/base/class.c 	list_add_tail(&class_intf->node, &parent->p->interfaces);
node              462 drivers/base/class.c 	list_del_init(&class_intf->node);
node               63 drivers/base/component.c 	struct list_head node;
node               73 drivers/base/component.c 	struct list_head node;
node              155 drivers/base/component.c 	list_for_each_entry(m, &masters, node)
node              167 drivers/base/component.c 	list_for_each_entry(c, &component_list, node) {
node              274 drivers/base/component.c 	list_for_each_entry(m, &masters, node) {
node              443 drivers/base/component.c 	list_del(&master->node);
node              491 drivers/base/component.c 	list_add(&master->node, &masters);
node              683 drivers/base/component.c 	list_add_tail(&component->node, &component_list);
node              689 drivers/base/component.c 		list_del(&component->node);
node              762 drivers/base/component.c 	list_for_each_entry(c, &component_list, node)
node              764 drivers/base/component.c 			list_del(&c->node);
node             2215 drivers/base/core.c 				    &dev->class->p->interfaces, node)
node             2367 drivers/base/core.c 				    &dev->class->p->interfaces, node)
node               28 drivers/base/devres.c 	struct devres_node		node;
node               40 drivers/base/devres.c 	struct devres_node		node[2];
node               50 drivers/base/devres.c static void set_node_dbginfo(struct devres_node *node, const char *name,
node               53 drivers/base/devres.c 	node->name = name;
node               54 drivers/base/devres.c 	node->size = size;
node               57 drivers/base/devres.c static void devres_log(struct device *dev, struct devres_node *node,
node               62 drivers/base/devres.c 			op, node, node->name, (unsigned long)node->size);
node               65 drivers/base/devres.c #define set_node_dbginfo(node, n, s)	do {} while (0)
node               66 drivers/base/devres.c #define devres_log(dev, node, op)	do {} while (0)
node               83 drivers/base/devres.c static struct devres_group * node_to_group(struct devres_node *node)
node               85 drivers/base/devres.c 	if (node->release == &group_open_release)
node               86 drivers/base/devres.c 		return container_of(node, struct devres_group, node[0]);
node               87 drivers/base/devres.c 	if (node->release == &group_close_release)
node               88 drivers/base/devres.c 		return container_of(node, struct devres_group, node[1]);
node              109 drivers/base/devres.c 	INIT_LIST_HEAD(&dr->node.entry);
node              110 drivers/base/devres.c 	dr->node.release = release;
node              114 drivers/base/devres.c static void add_dr(struct device *dev, struct devres_node *node)
node              116 drivers/base/devres.c 	devres_log(dev, node, "ADD");
node              117 drivers/base/devres.c 	BUG_ON(!list_empty(&node->entry));
node              118 drivers/base/devres.c 	list_add_tail(&node->entry, &dev->devres_head);
node              130 drivers/base/devres.c 	set_node_dbginfo(&dr->node, name, size);
node              181 drivers/base/devres.c 	struct devres_node *node;
node              189 drivers/base/devres.c 	list_for_each_entry_safe_reverse(node, tmp,
node              191 drivers/base/devres.c 		struct devres *dr = container_of(node, struct devres, node);
node              193 drivers/base/devres.c 		if (node->release != release)
node              214 drivers/base/devres.c 		BUG_ON(!list_empty(&dr->node.entry));
node              235 drivers/base/devres.c 	add_dr(dev, &dr->node);
node              243 drivers/base/devres.c 	struct devres_node *node;
node              245 drivers/base/devres.c 	list_for_each_entry_reverse(node, &dev->devres_head, entry) {
node              246 drivers/base/devres.c 		struct devres *dr = container_of(node, struct devres, node);
node              248 drivers/base/devres.c 		if (node->release != release)
node              310 drivers/base/devres.c 	dr = find_dr(dev, new_dr->node.release, match, match_data);
node              312 drivers/base/devres.c 		add_dr(dev, &new_dr->node);
node              347 drivers/base/devres.c 		list_del_init(&dr->node.entry);
node              348 drivers/base/devres.c 		devres_log(dev, &dr->node, "REM");
node              433 drivers/base/devres.c 		struct devres_node *node;
node              436 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
node              439 drivers/base/devres.c 		grp = node_to_group(node);
node              446 drivers/base/devres.c 			if (&node->entry == first)
node              448 drivers/base/devres.c 			list_move_tail(&node->entry, todo);
node              464 drivers/base/devres.c 		struct devres_node *node;
node              467 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
node              470 drivers/base/devres.c 		grp = node_to_group(node);
node              471 drivers/base/devres.c 		BUG_ON(!grp || list_empty(&grp->node[0].entry));
node              474 drivers/base/devres.c 		if (list_empty(&grp->node[1].entry))
node              482 drivers/base/devres.c 			list_move_tail(&grp->node[0].entry, todo);
node              483 drivers/base/devres.c 			list_del_init(&grp->node[1].entry);
node              505 drivers/base/devres.c 	list_for_each_entry_safe_reverse(dr, tmp, &todo, node.entry) {
node              506 drivers/base/devres.c 		devres_log(dev, &dr->node, "REL");
node              507 drivers/base/devres.c 		dr->node.release(dev, dr->data);
node              555 drivers/base/devres.c 	grp->node[0].release = &group_open_release;
node              556 drivers/base/devres.c 	grp->node[1].release = &group_close_release;
node              557 drivers/base/devres.c 	INIT_LIST_HEAD(&grp->node[0].entry);
node              558 drivers/base/devres.c 	INIT_LIST_HEAD(&grp->node[1].entry);
node              559 drivers/base/devres.c 	set_node_dbginfo(&grp->node[0], "grp<", 0);
node              560 drivers/base/devres.c 	set_node_dbginfo(&grp->node[1], "grp>", 0);
node              566 drivers/base/devres.c 	add_dr(dev, &grp->node[0]);
node              575 drivers/base/devres.c 	struct devres_node *node;
node              577 drivers/base/devres.c 	list_for_each_entry_reverse(node, &dev->devres_head, entry) {
node              580 drivers/base/devres.c 		if (node->release != &group_open_release)
node              583 drivers/base/devres.c 		grp = container_of(node, struct devres_group, node[0]);
node              588 drivers/base/devres.c 		} else if (list_empty(&grp->node[1].entry))
node              612 drivers/base/devres.c 		add_dr(dev, &grp->node[1]);
node              638 drivers/base/devres.c 		list_del_init(&grp->node[0].entry);
node              639 drivers/base/devres.c 		list_del_init(&grp->node[1].entry);
node              640 drivers/base/devres.c 		devres_log(dev, &grp->node[0], "REM");
node              672 drivers/base/devres.c 		struct list_head *first = &grp->node[0].entry;
node              675 drivers/base/devres.c 		if (!list_empty(&grp->node[1].entry))
node              676 drivers/base/devres.c 			end = grp->node[1].entry.next;
node              823 drivers/base/devres.c 	set_node_dbginfo(&dr->node, "devm_kzalloc_release", size);
node               34 drivers/base/node.c 	struct node *node_dev = to_node(dev);
node              105 drivers/base/node.c static void node_remove_accesses(struct node *node)
node              109 drivers/base/node.c 	list_for_each_entry_safe(c, cnext, &node->access_list, list_node) {
node              120 drivers/base/node.c static struct node_access_nodes *node_init_node_access(struct node *node,
node              126 drivers/base/node.c 	list_for_each_entry(access_node, &node->access_list, list_node)
node              136 drivers/base/node.c 	dev->parent = &node->dev;
node              146 drivers/base/node.c 	list_add_tail(&access_node->list_node, &node->access_list);
node              188 drivers/base/node.c 	struct node *node;
node              194 drivers/base/node.c 	node = node_devices[nid];
node              195 drivers/base/node.c 	c = node_init_node_access(node, access);
node              218 drivers/base/node.c 	struct list_head node;
node              257 drivers/base/node.c static void node_init_cache_dev(struct node *node)
node              265 drivers/base/node.c 	dev->parent = &node->dev;
node              274 drivers/base/node.c 	node->cache_dev = dev;
node              291 drivers/base/node.c 	struct node *node;
node              296 drivers/base/node.c 	node = node_devices[nid];
node              297 drivers/base/node.c 	list_for_each_entry(info, &node->cache_attrs, node) {
node              299 drivers/base/node.c 			dev_warn(&node->dev,
node              306 drivers/base/node.c 	if (!node->cache_dev)
node              307 drivers/base/node.c 		node_init_cache_dev(node);
node              308 drivers/base/node.c 	if (!node->cache_dev)
node              316 drivers/base/node.c 	dev->parent = node->cache_dev;
node              324 drivers/base/node.c 		dev_warn(&node->dev, "failed to add cache level:%d\n",
node              329 drivers/base/node.c 	list_add_tail(&info->node, &node->cache_attrs);
node              337 drivers/base/node.c static void node_remove_caches(struct node *node)
node              341 drivers/base/node.c 	if (!node->cache_dev)
node              344 drivers/base/node.c 	list_for_each_entry_safe(info, next, &node->cache_attrs, node) {
node              345 drivers/base/node.c 		list_del(&info->node);
node              348 drivers/base/node.c 	device_unregister(node->cache_dev);
node              357 drivers/base/node.c static void node_remove_caches(struct node *node) { }
node              564 drivers/base/node.c static inline bool hugetlb_register_node(struct node *node)
node              567 drivers/base/node.c 			node_state(node->dev.id, N_MEMORY)) {
node              568 drivers/base/node.c 		__hugetlb_register_node(node);
node              574 drivers/base/node.c static inline void hugetlb_unregister_node(struct node *node)
node              577 drivers/base/node.c 		__hugetlb_unregister_node(node);
node              587 drivers/base/node.c static inline void hugetlb_register_node(struct node *node) {}
node              589 drivers/base/node.c static inline void hugetlb_unregister_node(struct node *node) {}
node              594 drivers/base/node.c 	struct node *node = to_node(dev);
node              606 drivers/base/node.c 	flush_work(&node->node_work);
node              608 drivers/base/node.c 	kfree(node);
node              617 drivers/base/node.c static int register_node(struct node *node, int num)
node              621 drivers/base/node.c 	node->dev.id = num;
node              622 drivers/base/node.c 	node->dev.bus = &node_subsys;
node              623 drivers/base/node.c 	node->dev.release = node_device_release;
node              624 drivers/base/node.c 	node->dev.groups = node_dev_groups;
node              625 drivers/base/node.c 	error = device_register(&node->dev);
node              628 drivers/base/node.c 		put_device(&node->dev);
node              630 drivers/base/node.c 		hugetlb_register_node(node);
node              632 drivers/base/node.c 		compaction_register_node(node);
node              644 drivers/base/node.c void unregister_node(struct node *node)
node              646 drivers/base/node.c 	hugetlb_unregister_node(node);		/* no-op, if memoryless node */
node              647 drivers/base/node.c 	node_remove_accesses(node);
node              648 drivers/base/node.c 	node_remove_caches(node);
node              649 drivers/base/node.c 	device_unregister(&node->dev);
node              652 drivers/base/node.c struct node *node_devices[MAX_NUMNODES];
node              697 drivers/base/node.c 	struct node *init_node, *targ_node;
node              846 drivers/base/node.c 	struct node *node = container_of(work, struct node, node_work);
node              856 drivers/base/node.c 	if (!hugetlb_register_node(node))
node              857 drivers/base/node.c 		hugetlb_unregister_node(node);
node              912 drivers/base/node.c 	node_devices[nid] = kzalloc(sizeof(struct node), GFP_KERNEL);
node               31 drivers/base/power/clock_ops.c 	struct list_head node;
node              110 drivers/base/power/clock_ops.c 	list_add_tail(&ce->node, &psd->clock_list);
node              274 drivers/base/power/clock_ops.c 	list_for_each_entry(ce, &psd->clock_list, node) {
node              287 drivers/base/power/clock_ops.c 	list_del(&ce->node);
node              312 drivers/base/power/clock_ops.c 	list_for_each_entry(ce, &psd->clock_list, node) {
node              321 drivers/base/power/clock_ops.c 	list_del(&ce->node);
node              377 drivers/base/power/clock_ops.c 	list_for_each_entry_safe_reverse(ce, c, &psd->clock_list, node)
node              378 drivers/base/power/clock_ops.c 		list_move(&ce->node, &list);
node              384 drivers/base/power/clock_ops.c 	list_for_each_entry_safe_reverse(ce, c, &list, node) {
node              385 drivers/base/power/clock_ops.c 		list_del(&ce->node);
node              408 drivers/base/power/clock_ops.c 	list_for_each_entry_reverse(ce, &psd->clock_list, node) {
node              439 drivers/base/power/clock_ops.c 	list_for_each_entry(ce, &psd->clock_list, node)
node             1939 drivers/base/power/domain.c 	struct device_node *node;
node             2011 drivers/base/power/domain.c 	cp->node = of_node_get(np);
node             2173 drivers/base/power/domain.c 		if (cp->node == np) {
node             2192 drivers/base/power/domain.c 			of_node_put(cp->node);
node             2226 drivers/base/power/domain.c 		if (provider->node == genpdspec->np)
node             2622 drivers/base/power/domain.c 		np = it.node;
node              273 drivers/base/power/qos.c 	list_for_each_entry_safe(req, tmp, &f->list, data.flr.node) {
node              192 drivers/base/regmap/internal.h 	struct rb_node node;
node               31 drivers/base/regmap/regcache-rbtree.c 	struct rb_node node;
node               66 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *node;
node               78 drivers/base/regmap/regcache-rbtree.c 	node = rbtree_ctx->root.rb_node;
node               79 drivers/base/regmap/regcache-rbtree.c 	while (node) {
node               80 drivers/base/regmap/regcache-rbtree.c 		rbnode = rb_entry(node, struct regcache_rbtree_node, node);
node               87 drivers/base/regmap/regcache-rbtree.c 			node = node->rb_right;
node               89 drivers/base/regmap/regcache-rbtree.c 			node = node->rb_left;
node              107 drivers/base/regmap/regcache-rbtree.c 		rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node);
node              125 drivers/base/regmap/regcache-rbtree.c 	rb_link_node(&rbnode->node, parent, new);
node              126 drivers/base/regmap/regcache-rbtree.c 	rb_insert_color(&rbnode->node, root);
node              137 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *node;
node              148 drivers/base/regmap/regcache-rbtree.c 	for (node = rb_first(&rbtree_ctx->root); node != NULL;
node              149 drivers/base/regmap/regcache-rbtree.c 	     node = rb_next(node)) {
node              150 drivers/base/regmap/regcache-rbtree.c 		n = rb_entry(node, struct regcache_rbtree_node, node);
node              227 drivers/base/regmap/regcache-rbtree.c 		rbtree_node = rb_entry(next, struct regcache_rbtree_node, node);
node              228 drivers/base/regmap/regcache-rbtree.c 		next = rb_next(&rbtree_node->node);
node              229 drivers/base/regmap/regcache-rbtree.c 		rb_erase(&rbtree_node->node, &rbtree_ctx->root);
node              374 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *node;
node              403 drivers/base/regmap/regcache-rbtree.c 		node = rbtree_ctx->root.rb_node;
node              404 drivers/base/regmap/regcache-rbtree.c 		while (node) {
node              405 drivers/base/regmap/regcache-rbtree.c 			rbnode_tmp = rb_entry(node, struct regcache_rbtree_node,
node              406 drivers/base/regmap/regcache-rbtree.c 					      node);
node              432 drivers/base/regmap/regcache-rbtree.c 				node = node->rb_left;
node              434 drivers/base/regmap/regcache-rbtree.c 				node = node->rb_right;
node              469 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *node;
node              476 drivers/base/regmap/regcache-rbtree.c 	for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
node              477 drivers/base/regmap/regcache-rbtree.c 		rbnode = rb_entry(node, struct regcache_rbtree_node, node);
node              511 drivers/base/regmap/regcache-rbtree.c 	struct rb_node *node;
node              516 drivers/base/regmap/regcache-rbtree.c 	for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
node              517 drivers/base/regmap/regcache-rbtree.c 		rbnode = rb_entry(node, struct regcache_rbtree_node, node);
node              555 drivers/base/regmap/regmap-debugfs.c 		struct regmap_debugfs_node *node;
node              556 drivers/base/regmap/regmap-debugfs.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node              557 drivers/base/regmap/regmap-debugfs.c 		if (!node)
node              559 drivers/base/regmap/regmap-debugfs.c 		node->map = map;
node              560 drivers/base/regmap/regmap-debugfs.c 		node->name = name;
node              562 drivers/base/regmap/regmap-debugfs.c 		list_add(&node->link, &regmap_debugfs_early_list);
node              625 drivers/base/regmap/regmap-debugfs.c 		range_node = rb_entry(next, struct regmap_range_node, node);
node              632 drivers/base/regmap/regmap-debugfs.c 		next = rb_next(&range_node->node);
node              648 drivers/base/regmap/regmap-debugfs.c 		struct regmap_debugfs_node *node, *tmp;
node              651 drivers/base/regmap/regmap-debugfs.c 		list_for_each_entry_safe(node, tmp, &regmap_debugfs_early_list,
node              653 drivers/base/regmap/regmap-debugfs.c 			if (node->map == map) {
node              654 drivers/base/regmap/regmap-debugfs.c 				list_del(&node->link);
node              655 drivers/base/regmap/regmap-debugfs.c 				kfree(node);
node              664 drivers/base/regmap/regmap-debugfs.c 	struct regmap_debugfs_node *node, *tmp;
node              669 drivers/base/regmap/regmap-debugfs.c 	list_for_each_entry_safe(node, tmp, &regmap_debugfs_early_list, link) {
node              670 drivers/base/regmap/regmap-debugfs.c 		regmap_debugfs_init(node->map, node->name);
node              671 drivers/base/regmap/regmap-debugfs.c 		list_del(&node->link);
node              672 drivers/base/regmap/regmap-debugfs.c 		kfree(node);
node              539 drivers/base/regmap/regmap.c 			rb_entry(*new, struct regmap_range_node, node);
node              550 drivers/base/regmap/regmap.c 	rb_link_node(&data->node, parent, new);
node              551 drivers/base/regmap/regmap.c 	rb_insert_color(&data->node, root);
node              559 drivers/base/regmap/regmap.c 	struct rb_node *node = map->range_tree.rb_node;
node              561 drivers/base/regmap/regmap.c 	while (node) {
node              563 drivers/base/regmap/regmap.c 			rb_entry(node, struct regmap_range_node, node);
node              566 drivers/base/regmap/regmap.c 			node = node->rb_left;
node              568 drivers/base/regmap/regmap.c 			node = node->rb_right;
node              583 drivers/base/regmap/regmap.c 		range_node = rb_entry(next, struct regmap_range_node, node);
node              584 drivers/base/regmap/regmap.c 		next = rb_next(&range_node->node);
node              585 drivers/base/regmap/regmap.c 		rb_erase(&range_node->node, &map->range_tree);
node               18 drivers/base/swnode.c 	const struct software_node *node;
node               52 drivers/base/swnode.c software_node_to_swnode(const struct software_node *node)
node               57 drivers/base/swnode.c 	if (!node)
node               64 drivers/base/swnode.c 		if (swnode->node == node)
node               78 drivers/base/swnode.c 	return swnode ? swnode->node : NULL;
node               82 drivers/base/swnode.c struct fwnode_handle *software_node_fwnode(const struct software_node *node)
node               84 drivers/base/swnode.c 	struct swnode *swnode = software_node_to_swnode(node);
node              494 drivers/base/swnode.c 	return !!property_entry_get(swnode->node->properties, propname);
node              504 drivers/base/swnode.c 	return property_entry_read_int_array(swnode->node->properties, propname,
node              514 drivers/base/swnode.c 	return property_entry_read_string_array(swnode->node->properties,
node              578 drivers/base/swnode.c 	if (!swnode || !swnode->node->references)
node              581 drivers/base/swnode.c 	for (ref = swnode->node->references; ref->name; ref++)
node              588 drivers/base/swnode.c 	refnode = software_node_fwnode(ref->refs[index].node);
node              593 drivers/base/swnode.c 		prop = property_entry_get(swnode->node->properties, nargs_prop);
node              649 drivers/base/swnode.c 		if (parent == swnode->node->parent && swnode->node->name &&
node              650 drivers/base/swnode.c 		    !strcmp(name, swnode->node->name)) {
node              659 drivers/base/swnode.c 	return swnode ? swnode->node : NULL;
node              664 drivers/base/swnode.c software_node_register_properties(struct software_node *node,
node              673 drivers/base/swnode.c 	node->properties = props;
node              690 drivers/base/swnode.c 		property_entries_free(swnode->node->properties);
node              691 drivers/base/swnode.c 		kfree(swnode->node);
node              703 drivers/base/swnode.c swnode_register(const struct software_node *node, struct swnode *parent,
node              723 drivers/base/swnode.c 	swnode->node = node;
node              733 drivers/base/swnode.c 	if (node->name)
node              736 drivers/base/swnode.c 					   "%s", node->name);
node              754 drivers/base/swnode.c 		property_entries_free(node->properties);
node              804 drivers/base/swnode.c int software_node_register(const struct software_node *node)
node              806 drivers/base/swnode.c 	struct swnode *parent = software_node_to_swnode(node->parent);
node              808 drivers/base/swnode.c 	if (software_node_to_swnode(node))
node              811 drivers/base/swnode.c 	return PTR_ERR_OR_ZERO(swnode_register(node, parent, 0));
node              819 drivers/base/swnode.c 	struct software_node *node;
node              831 drivers/base/swnode.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              832 drivers/base/swnode.c 	if (!node)
node              835 drivers/base/swnode.c 	ret = software_node_register_properties(node, properties);
node              837 drivers/base/swnode.c 		kfree(node);
node              841 drivers/base/swnode.c 	node->parent = p ? p->node : NULL;
node              843 drivers/base/swnode.c 	return swnode_register(node, p, 1);
node               24 drivers/base/syscore.c 	list_add_tail(&ops->node, &syscore_ops_list);
node               36 drivers/base/syscore.c 	list_del(&ops->node);
node               62 drivers/base/syscore.c 	list_for_each_entry_reverse(ops, &syscore_ops_list, node)
node               79 drivers/base/syscore.c 	list_for_each_entry_continue(ops, &syscore_ops_list, node)
node              100 drivers/base/syscore.c 	list_for_each_entry(ops, &syscore_ops_list, node)
node              122 drivers/base/syscore.c 	list_for_each_entry_reverse(ops, &syscore_ops_list, node)
node              142 drivers/bcma/main.c 	struct device_node *node;
node              149 drivers/bcma/main.c 	for_each_child_of_node(parent->of_node, node) {
node              150 drivers/bcma/main.c 		reg = of_get_address(node, 0, &size, NULL);
node              153 drivers/bcma/main.c 		if (of_translate_address(node, reg) == core->addr)
node              154 drivers/bcma/main.c 			return node;
node              202 drivers/bcma/main.c 	struct device_node *node;
node              204 drivers/bcma/main.c 	node = bcma_of_find_child_device(parent, core);
node              205 drivers/bcma/main.c 	if (node)
node              206 drivers/bcma/main.c 		core->dev.of_node = node;
node              210 drivers/bcma/main.c 	of_dma_configure(&core->dev, node, false);
node               10 drivers/block/drbd/drbd_interval.c sector_t interval_end(struct rb_node *node)
node               12 drivers/block/drbd/drbd_interval.c 	struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb);
node               16 drivers/block/drbd/drbd_interval.c #define NODE_END(node) ((node)->sector + ((node)->size >> 9))
node               71 drivers/block/drbd/drbd_interval.c 	struct rb_node *node = root->rb_node;
node               73 drivers/block/drbd/drbd_interval.c 	while (node) {
node               75 drivers/block/drbd/drbd_interval.c 			rb_entry(node, struct drbd_interval, rb);
node               78 drivers/block/drbd/drbd_interval.c 			node = node->rb_left;
node               80 drivers/block/drbd/drbd_interval.c 			node = node->rb_right;
node               82 drivers/block/drbd/drbd_interval.c 			node = node->rb_left;
node               84 drivers/block/drbd/drbd_interval.c 			node = node->rb_right;
node              114 drivers/block/drbd/drbd_interval.c 	struct rb_node *node = root->rb_node;
node              120 drivers/block/drbd/drbd_interval.c 	while (node) {
node              122 drivers/block/drbd/drbd_interval.c 			rb_entry(node, struct drbd_interval, rb);
node              124 drivers/block/drbd/drbd_interval.c 		if (node->rb_left &&
node              125 drivers/block/drbd/drbd_interval.c 		    sector < interval_end(node->rb_left)) {
node              127 drivers/block/drbd/drbd_interval.c 			node = node->rb_left;
node              134 drivers/block/drbd/drbd_interval.c 			node = node->rb_right;
node              145 drivers/block/drbd/drbd_interval.c 	struct rb_node *node;
node              148 drivers/block/drbd/drbd_interval.c 		node = rb_next(&i->rb);
node              149 drivers/block/drbd/drbd_interval.c 		if (!node)
node              151 drivers/block/drbd/drbd_interval.c 		i = rb_entry(node, struct drbd_interval, rb);
node             3891 drivers/block/mtip32xx/mtip32xx.c static int get_least_used_cpu_on_node(int node)
node             3896 drivers/block/mtip32xx/mtip32xx.c 	node_mask = cpumask_of_node(node);
node              623 drivers/block/pktcdvd.c static inline struct pkt_rb_node *pkt_rbtree_next(struct pkt_rb_node *node)
node              625 drivers/block/pktcdvd.c 	struct rb_node *n = rb_next(&node->rb_node);
node              631 drivers/block/pktcdvd.c static void pkt_rbtree_erase(struct pktcdvd_device *pd, struct pkt_rb_node *node)
node              633 drivers/block/pktcdvd.c 	rb_erase(&node->rb_node, &pd->bio_queue);
node              634 drivers/block/pktcdvd.c 	mempool_free(node, &pd->rb_pool);
node              676 drivers/block/pktcdvd.c static void pkt_rbtree_insert(struct pktcdvd_device *pd, struct pkt_rb_node *node)
node              680 drivers/block/pktcdvd.c 	sector_t s = node->bio->bi_iter.bi_sector;
node              691 drivers/block/pktcdvd.c 	rb_link_node(&node->rb_node, parent, p);
node              692 drivers/block/pktcdvd.c 	rb_insert_color(&node->rb_node, &pd->bio_queue);
node             1167 drivers/block/pktcdvd.c 	struct pkt_rb_node *node, *first_node;
node             1188 drivers/block/pktcdvd.c 	node = first_node;
node             1189 drivers/block/pktcdvd.c 	while (node) {
node             1190 drivers/block/pktcdvd.c 		bio = node->bio;
node             1200 drivers/block/pktcdvd.c 		node = pkt_rbtree_next(node);
node             1201 drivers/block/pktcdvd.c 		if (!node) {
node             1204 drivers/block/pktcdvd.c 				node = rb_entry(n, struct pkt_rb_node, rb_node);
node             1206 drivers/block/pktcdvd.c 		if (node == first_node)
node             1207 drivers/block/pktcdvd.c 			node = NULL;
node             1228 drivers/block/pktcdvd.c 	while ((node = pkt_rbtree_find(pd, zone)) != NULL) {
node             1229 drivers/block/pktcdvd.c 		bio = node->bio;
node             1234 drivers/block/pktcdvd.c 		pkt_rbtree_erase(pd, node);
node             2353 drivers/block/pktcdvd.c 	struct pkt_rb_node *node;
node             2407 drivers/block/pktcdvd.c 	node = mempool_alloc(&pd->rb_pool, GFP_NOIO);
node             2408 drivers/block/pktcdvd.c 	node->bio = bio;
node             2412 drivers/block/pktcdvd.c 	pkt_rbtree_insert(pd, node);
node              207 drivers/block/rbd.c 	struct list_head	node;
node              452 drivers/block/rbd.c 	struct list_head	node;
node              756 drivers/block/rbd.c 	INIT_LIST_HEAD(&rbdc->node);
node              768 drivers/block/rbd.c 	list_add_tail(&rbdc->node, &rbd_client_list);
node              806 drivers/block/rbd.c 	list_for_each_entry(client_node, &rbd_client_list, node) {
node              982 drivers/block/rbd.c 	list_del(&rbdc->node);
node             5544 drivers/block/rbd.c 	INIT_LIST_HEAD(&rbd_dev->node);
node             7133 drivers/block/rbd.c 	list_add_tail(&rbd_dev->node, &rbd_dev_list);
node             7230 drivers/block/rbd.c 		rbd_dev = list_entry(tmp, struct rbd_device, node);
node             7260 drivers/block/rbd.c 	list_del_init(&rbd_dev->node);
node              212 drivers/block/xen-blkback/blkback.c #define foreach_grant_safe(pos, n, rbtree, node) \
node              213 drivers/block/xen-blkback/blkback.c 	for ((pos) = container_of(rb_first((rbtree)), typeof(*(pos)), node), \
node              214 drivers/block/xen-blkback/blkback.c 	     (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL; \
node              215 drivers/block/xen-blkback/blkback.c 	     &(pos)->node != NULL; \
node              216 drivers/block/xen-blkback/blkback.c 	     (pos) = container_of(n, typeof(*(pos)), node), \
node              217 drivers/block/xen-blkback/blkback.c 	     (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL)
node              245 drivers/block/xen-blkback/blkback.c 		this = container_of(*new, struct persistent_gnt, node);
node              260 drivers/block/xen-blkback/blkback.c 	rb_link_node(&(persistent_gnt->node), parent, new);
node              261 drivers/block/xen-blkback/blkback.c 	rb_insert_color(&(persistent_gnt->node), &ring->persistent_gnts);
node              271 drivers/block/xen-blkback/blkback.c 	struct rb_node *node = NULL;
node              273 drivers/block/xen-blkback/blkback.c 	node = ring->persistent_gnts.rb_node;
node              274 drivers/block/xen-blkback/blkback.c 	while (node) {
node              275 drivers/block/xen-blkback/blkback.c 		data = container_of(node, struct persistent_gnt, node);
node              278 drivers/block/xen-blkback/blkback.c 			node = node->rb_left;
node              280 drivers/block/xen-blkback/blkback.c 			node = node->rb_right;
node              318 drivers/block/xen-blkback/blkback.c 	foreach_grant_safe(persistent_gnt, n, root, node) {
node              330 drivers/block/xen-blkback/blkback.c 			!rb_next(&persistent_gnt->node)) {
node              339 drivers/block/xen-blkback/blkback.c 		rb_erase(&persistent_gnt->node, root);
node              427 drivers/block/xen-blkback/blkback.c 	foreach_grant_safe(persistent_gnt, n, root, node) {
node              438 drivers/block/xen-blkback/blkback.c 		rb_erase(&persistent_gnt->node, root);
node              245 drivers/block/xen-blkback/common.h 	struct rb_node node;
node              511 drivers/block/xen-blkback/xenbus.c 	if (be->backend_watch.node) {
node              513 drivers/block/xen-blkback/xenbus.c 		kfree(be->backend_watch.node);
node              514 drivers/block/xen-blkback/xenbus.c 		be->backend_watch.node = NULL;
node               88 drivers/block/xen-blkfront.c 	struct list_head node;
node              311 drivers/block/xen-blkfront.c 		list_add(&gnt_list_entry->node, &rinfo->grants);
node              319 drivers/block/xen-blkfront.c 	                         &rinfo->grants, node) {
node              320 drivers/block/xen-blkfront.c 		list_del(&gnt_list_entry->node);
node              336 drivers/block/xen-blkfront.c 					  node);
node              337 drivers/block/xen-blkfront.c 	list_del(&gnt_list_entry->node);
node             1266 drivers/block/xen-blkfront.c 					 &rinfo->grants, node) {
node             1267 drivers/block/xen-blkfront.c 			list_del(&persistent_gnt->node);
node             1505 drivers/block/xen-blkfront.c 			list_add(&s->grants_used[i]->node, &rinfo->grants);
node             1516 drivers/block/xen-blkfront.c 			list_add_tail(&s->grants_used[i]->node, &rinfo->grants);
node             1525 drivers/block/xen-blkfront.c 				list_add(&s->indirect_grants[i]->node, &rinfo->grants);
node             1540 drivers/block/xen-blkfront.c 				list_add_tail(&s->indirect_grants[i]->node, &rinfo->grants);
node             2676 drivers/block/xen-blkfront.c 					 node) {
node             2681 drivers/block/xen-blkfront.c 			list_del(&gnt_list_entry->node);
node             2685 drivers/block/xen-blkfront.c 			list_add_tail(&gnt_list_entry->node, &rinfo->grants);
node              158 drivers/block/zram/zcomp.c int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node)
node              160 drivers/block/zram/zcomp.c 	struct zcomp *comp = hlist_entry(node, struct zcomp, node);
node              175 drivers/block/zram/zcomp.c int zcomp_cpu_dead(unsigned int cpu, struct hlist_node *node)
node              177 drivers/block/zram/zcomp.c 	struct zcomp *comp = hlist_entry(node, struct zcomp, node);
node              195 drivers/block/zram/zcomp.c 	ret = cpuhp_state_add_instance(CPUHP_ZCOMP_PREPARE, &comp->node);
node              207 drivers/block/zram/zcomp.c 	cpuhp_state_remove_instance(CPUHP_ZCOMP_PREPARE, &comp->node);
node               19 drivers/block/zram/zcomp.h 	struct hlist_node node;
node               22 drivers/block/zram/zcomp.h int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node);
node               23 drivers/block/zram/zcomp.h int zcomp_cpu_dead(unsigned int cpu, struct hlist_node *node);
node              901 drivers/bluetooth/btmtkuart.c 	struct device_node *node = serdev->dev.of_node;
node              906 drivers/bluetooth/btmtkuart.c 		of_property_read_u32(node, "current-speed", &speed);
node               76 drivers/bus/fsl-mc/fsl-mc-allocator.c 	INIT_LIST_HEAD(&resource->node);
node               77 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_add_tail(&resource->node, &res_pool->free_list);
node              131 drivers/bus/fsl-mc/fsl-mc-allocator.c 	if (list_empty(&resource->node)) {
node              139 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_del_init(&resource->node);
node              197 drivers/bus/fsl-mc/fsl-mc-allocator.c 					    struct fsl_mc_resource, node);
node              215 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_del_init(&resource->node);
node              240 drivers/bus/fsl-mc/fsl-mc-allocator.c 	if (!list_empty(&resource->node))
node              243 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_add_tail(&resource->node, &res_pool->free_list);
node              385 drivers/bus/fsl-mc/fsl-mc-allocator.c 		INIT_LIST_HEAD(&mc_dev_irq->resource.node);
node              386 drivers/bus/fsl-mc/fsl-mc-allocator.c 		list_add_tail(&mc_dev_irq->resource.node, &res_pool->free_list);
node              561 drivers/bus/fsl-mc/fsl-mc-allocator.c 	list_for_each_entry_safe(resource, next, &res_pool->free_list, node) {
node              422 drivers/bus/hisi_lpc.c 	list_for_each_entry(rentry, &resource_list, node)
node              466 drivers/bus/hisi_lpc.c 	list_for_each_entry(child, &adev->children, node)
node              487 drivers/bus/hisi_lpc.c 	list_for_each_entry(child, &adev->children, node) {
node             1203 drivers/bus/mvebu-mbus.c mbus_parse_ranges(struct device_node *node,
node             1212 drivers/bus/mvebu-mbus.c 	*ranges_start = of_get_property(node, "ranges", &ranges_len);
node             1220 drivers/bus/mvebu-mbus.c 	*addr_cells = of_n_addr_cells(node);
node             1222 drivers/bus/mvebu-mbus.c 	prop = of_get_property(node, "#address-cells", NULL);
node             1225 drivers/bus/mvebu-mbus.c 	prop = of_get_property(node, "#size-cells", NULL);
node             1232 drivers/bus/mvebu-mbus.c 		pr_warn("malformed ranges entry '%pOFn'\n", node);
node              199 drivers/bus/sunxi-rsb.c 		struct device_node *node, u16 hwaddr, u8 rtaddr)
node              213 drivers/bus/sunxi-rsb.c 	rdev->dev.of_node = node;
node               46 drivers/bus/vexpress-config.c static void vexpress_config_find_prop(struct device_node *node,
node               52 drivers/bus/vexpress-config.c 	of_node_get(node);
node               53 drivers/bus/vexpress-config.c 	while (node) {
node               54 drivers/bus/vexpress-config.c 		if (of_property_read_u32(node, name, val) == 0) {
node               55 drivers/bus/vexpress-config.c 			of_node_put(node);
node               58 drivers/bus/vexpress-config.c 		node = of_get_next_parent(node);
node               62 drivers/bus/vexpress-config.c int vexpress_config_get_topo(struct device_node *node, u32 *site,
node               65 drivers/bus/vexpress-config.c 	vexpress_config_find_prop(node, "arm,vexpress,site", site);
node               70 drivers/bus/vexpress-config.c 	vexpress_config_find_prop(node, "arm,vexpress,position", position);
node               71 drivers/bus/vexpress-config.c 	vexpress_config_find_prop(node, "arm,vexpress,dcc", dcc);
node              155 drivers/bus/vexpress-config.c 	const struct device_node *node = data;
node              158 drivers/bus/vexpress-config.c 			dev->parent->of_node, node);
node              160 drivers/bus/vexpress-config.c 	return dev->parent->of_node == node;
node              163 drivers/bus/vexpress-config.c static int vexpress_config_populate(struct device_node *node)
node              169 drivers/bus/vexpress-config.c 	bridge = of_parse_phandle(node, "arm,vexpress,config-bridge", 0);
node              179 drivers/bus/vexpress-config.c 	ret = of_platform_populate(node, NULL, NULL, parent);
node              189 drivers/bus/vexpress-config.c 	struct device_node *node;
node              192 drivers/bus/vexpress-config.c 	for_each_compatible_node(node, NULL, "arm,vexpress,config-bus") {
node              193 drivers/bus/vexpress-config.c 		err = vexpress_config_populate(node);
node              195 drivers/bus/vexpress-config.c 			of_node_put(node);
node              169 drivers/clk/axs10x/i2s_pll_clock.c 	struct device_node *node = dev->of_node;
node              187 drivers/clk/axs10x/i2s_pll_clock.c 	clk_name = node->name;
node              190 drivers/clk/axs10x/i2s_pll_clock.c 	parent_name = of_clk_get_parent_name(node, 0);
node              203 drivers/clk/axs10x/i2s_pll_clock.c 	return of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              272 drivers/clk/axs10x/pll_clock.c static void __init of_axs10x_pll_clk_setup(struct device_node *node)
node              283 drivers/clk/axs10x/pll_clock.c 	pll_clk->base = of_iomap(node, 0);
node              289 drivers/clk/axs10x/pll_clock.c 	pll_clk->lock = of_iomap(node, 1);
node              295 drivers/clk/axs10x/pll_clock.c 	init.name = node->name;
node              297 drivers/clk/axs10x/pll_clock.c 	parent_name = of_clk_get_parent_name(node, 0);
node              305 drivers/clk/axs10x/pll_clock.c 		pr_err("failed to register %pOFn clock\n", node);
node              309 drivers/clk/axs10x/pll_clock.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clk->hw);
node              311 drivers/clk/axs10x/pll_clock.c 		pr_err("failed to add hw provider for %pOFn clock\n", node);
node              263 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_root_ccu_setup(struct device_node *node)
node              265 drivers/clk/bcm/clk-bcm21664.c 	kona_dt_ccu_setup(&root_ccu_data, node);
node              268 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_aon_ccu_setup(struct device_node *node)
node              270 drivers/clk/bcm/clk-bcm21664.c 	kona_dt_ccu_setup(&aon_ccu_data, node);
node              273 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_master_ccu_setup(struct device_node *node)
node              275 drivers/clk/bcm/clk-bcm21664.c 	kona_dt_ccu_setup(&master_ccu_data, node);
node              278 drivers/clk/bcm/clk-bcm21664.c static void __init kona_dt_slave_ccu_setup(struct device_node *node)
node              280 drivers/clk/bcm/clk-bcm21664.c 	kona_dt_ccu_setup(&slave_ccu_data, node);
node              341 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_root_ccu_setup(struct device_node *node)
node              343 drivers/clk/bcm/clk-bcm281xx.c 	kona_dt_ccu_setup(&root_ccu_data, node);
node              346 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_aon_ccu_setup(struct device_node *node)
node              348 drivers/clk/bcm/clk-bcm281xx.c 	kona_dt_ccu_setup(&aon_ccu_data, node);
node              351 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_hub_ccu_setup(struct device_node *node)
node              353 drivers/clk/bcm/clk-bcm281xx.c 	kona_dt_ccu_setup(&hub_ccu_data, node);
node              356 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_master_ccu_setup(struct device_node *node)
node              358 drivers/clk/bcm/clk-bcm281xx.c 	kona_dt_ccu_setup(&master_ccu_data, node);
node              361 drivers/clk/bcm/clk-bcm281xx.c static void __init kona_dt_slave_ccu_setup(struct device_node *node)
node              363 drivers/clk/bcm/clk-bcm281xx.c 	kona_dt_ccu_setup(&slave_ccu_data, node);
node               18 drivers/clk/bcm/clk-bcm63xx.c static void __init bcm63138_armpll_init(struct device_node *node)
node               20 drivers/clk/bcm/clk-bcm63xx.c 	iproc_armpll_setup(node);
node               51 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_armpll_init(struct device_node *node)
node               53 drivers/clk/bcm/clk-cygnus.c 	iproc_armpll_setup(node);
node              110 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_genpll_clk_init(struct device_node *node)
node              112 drivers/clk/bcm/clk-cygnus.c 	iproc_pll_clk_setup(node, &genpll, NULL, 0, genpll_clk,
node              168 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_lcpll0_clk_init(struct device_node *node)
node              170 drivers/clk/bcm/clk-cygnus.c 	iproc_pll_clk_setup(node, &lcpll0, NULL, 0, lcpll0_clk,
node              246 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_mipipll_clk_init(struct device_node *node)
node              248 drivers/clk/bcm/clk-cygnus.c 	iproc_pll_clk_setup(node, &mipipll, mipipll_vco_params,
node              266 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_asiu_init(struct device_node *node)
node              268 drivers/clk/bcm/clk-cygnus.c 	iproc_asiu_setup(node, asiu_div, asiu_gate, ARRAY_SIZE(asiu_div));
node              308 drivers/clk/bcm/clk-cygnus.c static void __init cygnus_audiopll_clk_init(struct device_node *node)
node              310 drivers/clk/bcm/clk-cygnus.c 	iproc_pll_clk_setup(node, &audiopll, NULL, 0,
node               23 drivers/clk/bcm/clk-hr2.c static void __init hr2_armpll_init(struct device_node *node)
node               25 drivers/clk/bcm/clk-hr2.c 	iproc_armpll_setup(node);
node              244 drivers/clk/bcm/clk-iproc-armpll.c void __init iproc_armpll_setup(struct device_node *node)
node              255 drivers/clk/bcm/clk-iproc-armpll.c 	pll->base = of_iomap(node, 0);
node              259 drivers/clk/bcm/clk-iproc-armpll.c 	init.name = node->name;
node              262 drivers/clk/bcm/clk-iproc-armpll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              271 drivers/clk/bcm/clk-iproc-armpll.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll->hw);
node              185 drivers/clk/bcm/clk-iproc-asiu.c void __init iproc_asiu_setup(struct device_node *node,
node              210 drivers/clk/bcm/clk-iproc-asiu.c 	asiu->div_base = of_iomap(node, 0);
node              214 drivers/clk/bcm/clk-iproc-asiu.c 	asiu->gate_base = of_iomap(node, 1);
node              224 drivers/clk/bcm/clk-iproc-asiu.c 		ret = of_property_read_string_index(node, "clock-output-names",
node              237 drivers/clk/bcm/clk-iproc-asiu.c 		parent_name = of_clk_get_parent_name(node, 0);
node              248 drivers/clk/bcm/clk-iproc-asiu.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get,
node              725 drivers/clk/bcm/clk-iproc-pll.c void iproc_pll_clk_setup(struct device_node *node,
node              756 drivers/clk/bcm/clk-iproc-pll.c 	pll->control_base = of_iomap(node, 0);
node              761 drivers/clk/bcm/clk-iproc-pll.c 	pll->pwr_base = of_iomap(node, 1);
node              765 drivers/clk/bcm/clk-iproc-pll.c 		pll->asiu_base = of_iomap(node, 2);
node              774 drivers/clk/bcm/clk-iproc-pll.c 		pll->status_base = of_iomap(node, 2);
node              786 drivers/clk/bcm/clk-iproc-pll.c 	init.name = node->name;
node              789 drivers/clk/bcm/clk-iproc-pll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              812 drivers/clk/bcm/clk-iproc-pll.c 		parent_name = node->name;
node              814 drivers/clk/bcm/clk-iproc-pll.c 		ret = of_property_read_string_index(node, "clock-output-names",
node              837 drivers/clk/bcm/clk-iproc-pll.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, clk_data);
node              212 drivers/clk/bcm/clk-iproc.h void iproc_armpll_setup(struct device_node *node);
node              213 drivers/clk/bcm/clk-iproc.h void iproc_pll_clk_setup(struct device_node *node,
node              219 drivers/clk/bcm/clk-iproc.h void iproc_asiu_setup(struct device_node *node,
node              760 drivers/clk/bcm/clk-kona-setup.c 	of_clk_del_provider(ccu->node);	/* safe if never added */
node              762 drivers/clk/bcm/clk-kona-setup.c 	of_node_put(ccu->node);
node              763 drivers/clk/bcm/clk-kona-setup.c 	ccu->node = NULL;
node              802 drivers/clk/bcm/clk-kona-setup.c 			struct device_node *node)
node              809 drivers/clk/bcm/clk-kona-setup.c 	ret = of_address_to_resource(node, 0, &res);
node              812 drivers/clk/bcm/clk-kona-setup.c 			node);
node              819 drivers/clk/bcm/clk-kona-setup.c 			node);
node              826 drivers/clk/bcm/clk-kona-setup.c 		pr_err("%s: ccu data not valid for %pOFn\n", __func__, node);
node              833 drivers/clk/bcm/clk-kona-setup.c 			node);
node              836 drivers/clk/bcm/clk-kona-setup.c 	ccu->node = of_node_get(node);
node              849 drivers/clk/bcm/clk-kona-setup.c 	ret = of_clk_add_hw_provider(node, of_clk_kona_onecell_get, ccu);
node              852 drivers/clk/bcm/clk-kona-setup.c 				node, ret);
node              857 drivers/clk/bcm/clk-kona-setup.c 		pr_err("Broadcom %pOFn initialization had errors\n", node);
node              862 drivers/clk/bcm/clk-kona-setup.c 	pr_err("Broadcom %pOFn setup aborted\n", node);
node              483 drivers/clk/bcm/clk-kona.h 	struct device_node *node;
node              507 drivers/clk/bcm/clk-kona.h 				struct device_node *node);
node               96 drivers/clk/bcm/clk-ns2.c static void __init ns2_genpll_scr_clk_init(struct device_node *node)
node               98 drivers/clk/bcm/clk-ns2.c 	iproc_pll_clk_setup(node, &genpll_scr, NULL, 0, genpll_scr_clk,
node              158 drivers/clk/bcm/clk-ns2.c static void __init ns2_genpll_sw_clk_init(struct device_node *node)
node              160 drivers/clk/bcm/clk-ns2.c 	iproc_pll_clk_setup(node, &genpll_sw, NULL, 0, genpll_sw_clk,
node              220 drivers/clk/bcm/clk-ns2.c static void __init ns2_lcpll_ddr_clk_init(struct device_node *node)
node              222 drivers/clk/bcm/clk-ns2.c 	iproc_pll_clk_setup(node, &lcpll_ddr, NULL, 0, lcpll_ddr_clk,
node              282 drivers/clk/bcm/clk-ns2.c static void __init ns2_lcpll_ports_clk_init(struct device_node *node)
node              284 drivers/clk/bcm/clk-ns2.c 	iproc_pll_clk_setup(node, &lcpll_ports, NULL, 0, lcpll_ports_clk,
node               39 drivers/clk/bcm/clk-nsp.c static void __init nsp_armpll_init(struct device_node *node)
node               41 drivers/clk/bcm/clk-nsp.c 	iproc_armpll_setup(node);
node               95 drivers/clk/bcm/clk-nsp.c static void __init nsp_genpll_clk_init(struct device_node *node)
node               97 drivers/clk/bcm/clk-nsp.c 	iproc_pll_clk_setup(node, &genpll, NULL, 0, genpll_clk,
node              134 drivers/clk/bcm/clk-nsp.c static void __init nsp_lcpll0_clk_init(struct device_node *node)
node              136 drivers/clk/bcm/clk-nsp.c 	iproc_pll_clk_setup(node, &lcpll0, NULL, 0, lcpll0_clk,
node              181 drivers/clk/bcm/clk-sr.c static void sr_genpll3_clk_init(struct device_node *node)
node              183 drivers/clk/bcm/clk-sr.c 	iproc_pll_clk_setup(node, &sr_genpll3, NULL, 0, sr_genpll3_clk,
node              641 drivers/clk/clk-cdce925.c 	struct device_node *node = client->dev.of_node;
node              683 drivers/clk/clk-cdce925.c 	parent_name = of_clk_get_parent_name(node, 0);
node              690 drivers/clk/clk-cdce925.c 	if (of_property_read_u32(node, "xtal-load-pf", &value) == 0)
node              718 drivers/clk/clk-cdce925.c 		np_output = of_get_child_by_name(node, child_name);
node               14 drivers/clk/clk-conf.c static int __set_clk_parents(struct device_node *node, bool clk_supplier)
node               20 drivers/clk/clk-conf.c 	num_parents = of_count_phandle_with_args(node, "assigned-clock-parents",
node               24 drivers/clk/clk-conf.c 		       node);
node               27 drivers/clk/clk-conf.c 		rc = of_parse_phandle_with_args(node, "assigned-clock-parents",
node               36 drivers/clk/clk-conf.c 		if (clkspec.np == node && !clk_supplier)
node               42 drivers/clk/clk-conf.c 					index, node);
node               46 drivers/clk/clk-conf.c 		rc = of_parse_phandle_with_args(node, "assigned-clocks",
node               50 drivers/clk/clk-conf.c 		if (clkspec.np == node && !clk_supplier) {
node               58 drivers/clk/clk-conf.c 					index, node);
node               76 drivers/clk/clk-conf.c static int __set_clk_rates(struct device_node *node, bool clk_supplier)
node               85 drivers/clk/clk-conf.c 	of_property_for_each_u32(node, "assigned-clock-rates", prop, cur, rate) {
node               87 drivers/clk/clk-conf.c 			rc = of_parse_phandle_with_args(node, "assigned-clocks",
node               96 drivers/clk/clk-conf.c 			if (clkspec.np == node && !clk_supplier)
node              103 drivers/clk/clk-conf.c 						index, node);
node              131 drivers/clk/clk-conf.c int of_clk_set_defaults(struct device_node *node, bool clk_supplier)
node              135 drivers/clk/clk-conf.c 	if (!node)
node              138 drivers/clk/clk-conf.c 	rc = __set_clk_parents(node, clk_supplier);
node              142 drivers/clk/clk-conf.c 	return __set_clk_rates(node, clk_supplier);
node              162 drivers/clk/clk-fixed-factor.c static struct clk_hw *_of_fixed_factor_clk_setup(struct device_node *node)
node              165 drivers/clk/clk-fixed-factor.c 	const char *clk_name = node->name;
node              170 drivers/clk/clk-fixed-factor.c 	if (of_property_read_u32(node, "clock-div", &div)) {
node              172 drivers/clk/clk-fixed-factor.c 			__func__, node);
node              176 drivers/clk/clk-fixed-factor.c 	if (of_property_read_u32(node, "clock-mult", &mult)) {
node              178 drivers/clk/clk-fixed-factor.c 			__func__, node);
node              182 drivers/clk/clk-fixed-factor.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              184 drivers/clk/clk-fixed-factor.c 	if (of_match_node(set_rate_parent_matches, node))
node              187 drivers/clk/clk-fixed-factor.c 	hw = __clk_hw_register_fixed_factor(NULL, node, clk_name, NULL, 0,
node              194 drivers/clk/clk-fixed-factor.c 		of_node_clear_flag(node, OF_POPULATED);
node              198 drivers/clk/clk-fixed-factor.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node              210 drivers/clk/clk-fixed-factor.c void __init of_fixed_factor_clk_setup(struct device_node *node)
node              212 drivers/clk/clk-fixed-factor.c 	_of_fixed_factor_clk_setup(node);
node               18 drivers/clk/clk-fixed-mmio.c static struct clk_hw *fixed_mmio_clk_setup(struct device_node *node)
node               21 drivers/clk/clk-fixed-mmio.c 	const char *clk_name = node->name;
node               26 drivers/clk/clk-fixed-mmio.c 	base = of_iomap(node, 0);
node               28 drivers/clk/clk-fixed-mmio.c 		pr_err("%pOFn: failed to map address\n", node);
node               34 drivers/clk/clk-fixed-mmio.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               38 drivers/clk/clk-fixed-mmio.c 		pr_err("%pOFn: failed to register fixed rate clock\n", node);
node               42 drivers/clk/clk-fixed-mmio.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, clk);
node               44 drivers/clk/clk-fixed-mmio.c 		pr_err("%pOFn: failed to add clock provider\n", node);
node               52 drivers/clk/clk-fixed-mmio.c static void __init of_fixed_mmio_clk_setup(struct device_node *node)
node               54 drivers/clk/clk-fixed-mmio.c 	fixed_mmio_clk_setup(node);
node              158 drivers/clk/clk-fixed-rate.c static struct clk *_of_fixed_clk_setup(struct device_node *node)
node              161 drivers/clk/clk-fixed-rate.c 	const char *clk_name = node->name;
node              166 drivers/clk/clk-fixed-rate.c 	if (of_property_read_u32(node, "clock-frequency", &rate))
node              169 drivers/clk/clk-fixed-rate.c 	of_property_read_u32(node, "clock-accuracy", &accuracy);
node              171 drivers/clk/clk-fixed-rate.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              178 drivers/clk/clk-fixed-rate.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              190 drivers/clk/clk-fixed-rate.c void __init of_fixed_clk_setup(struct device_node *node)
node              192 drivers/clk/clk-fixed-rate.c 	_of_fixed_clk_setup(node);
node              241 drivers/clk/clk-gpio.c 	struct device_node *node = pdev->dev.of_node;
node              249 drivers/clk/clk-gpio.c 	num_parents = of_clk_get_parent_count(node);
node              256 drivers/clk/clk-gpio.c 		of_clk_parent_fill(node, parent_names, num_parents);
node              261 drivers/clk/clk-gpio.c 	is_mux = of_device_is_compatible(node, "gpio-mux-clock");
node              269 drivers/clk/clk-gpio.c 					node, __func__);
node              272 drivers/clk/clk-gpio.c 					node, __func__,
node              278 drivers/clk/clk-gpio.c 		clk = clk_register_gpio_mux(&pdev->dev, node->name,
node              281 drivers/clk/clk-gpio.c 		clk = clk_register_gpio_gate(&pdev->dev, node->name,
node              287 drivers/clk/clk-gpio.c 	return of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              263 drivers/clk/clk-highbank.c static void __init hb_clk_init(struct device_node *node, const struct clk_ops *ops, unsigned long clkflags)
node              267 drivers/clk/clk-highbank.c 	const char *clk_name = node->name;
node              273 drivers/clk/clk-highbank.c 	rc = of_property_read_u32(node, "reg", &reg);
node              288 drivers/clk/clk-highbank.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              293 drivers/clk/clk-highbank.c 	parent_name = of_clk_get_parent_name(node, 0);
node              304 drivers/clk/clk-highbank.c 	of_clk_add_hw_provider(node, of_clk_hw_simple_get, &hb_clk->hw);
node              307 drivers/clk/clk-highbank.c static void __init hb_pll_init(struct device_node *node)
node              309 drivers/clk/clk-highbank.c 	hb_clk_init(node, &clk_pll_ops, 0);
node              313 drivers/clk/clk-highbank.c static void __init hb_a9periph_init(struct device_node *node)
node              315 drivers/clk/clk-highbank.c 	hb_clk_init(node, &a9periphclk_ops, 0);
node              319 drivers/clk/clk-highbank.c static void __init hb_a9bus_init(struct device_node *node)
node              321 drivers/clk/clk-highbank.c 	hb_clk_init(node, &a9bclk_ops, CLK_IS_CRITICAL);
node              325 drivers/clk/clk-highbank.c static void __init hb_emmc_init(struct device_node *node)
node              327 drivers/clk/clk-highbank.c 	hb_clk_init(node, &periclk_ops, 0);
node              354 drivers/clk/clk-hsdk-pll.c static void __init of_hsdk_pll_clk_setup(struct device_node *node)
node              366 drivers/clk/clk-hsdk-pll.c 	pll_clk->regs = of_iomap(node, 0);
node              372 drivers/clk/clk-hsdk-pll.c 	pll_clk->spec_regs = of_iomap(node, 1);
node              378 drivers/clk/clk-hsdk-pll.c 	init.name = node->name;
node              380 drivers/clk/clk-hsdk-pll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              382 drivers/clk/clk-hsdk-pll.c 	num_parents = of_clk_get_parent_count(node);
node              394 drivers/clk/clk-hsdk-pll.c 		pr_err("failed to register %pOFn clock\n", node);
node              398 drivers/clk/clk-hsdk-pll.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clk->hw);
node              400 drivers/clk/clk-hsdk-pll.c 		pr_err("failed to add hw provider for %pOFn clock\n", node);
node               19 drivers/clk/clk-moxart.c static void __init moxart_of_pll_clk_init(struct device_node *node)
node               25 drivers/clk/clk-moxart.c 	const char *name = node->name;
node               28 drivers/clk/clk-moxart.c 	of_property_read_string(node, "clock-output-names", &name);
node               29 drivers/clk/clk-moxart.c 	parent_name = of_clk_get_parent_name(node, 0);
node               31 drivers/clk/clk-moxart.c 	base = of_iomap(node, 0);
node               33 drivers/clk/clk-moxart.c 		pr_err("%pOF: of_iomap failed\n", node);
node               40 drivers/clk/clk-moxart.c 	ref_clk = of_clk_get(node, 0);
node               42 drivers/clk/clk-moxart.c 		pr_err("%pOF: of_clk_get failed\n", node);
node               48 drivers/clk/clk-moxart.c 		pr_err("%pOF: failed to register clock\n", node);
node               53 drivers/clk/clk-moxart.c 	of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node               58 drivers/clk/clk-moxart.c static void __init moxart_of_apb_clk_init(struct device_node *node)
node               65 drivers/clk/clk-moxart.c 	const char *name = node->name;
node               68 drivers/clk/clk-moxart.c 	of_property_read_string(node, "clock-output-names", &name);
node               69 drivers/clk/clk-moxart.c 	parent_name = of_clk_get_parent_name(node, 0);
node               71 drivers/clk/clk-moxart.c 	base = of_iomap(node, 0);
node               73 drivers/clk/clk-moxart.c 		pr_err("%pOF: of_iomap failed\n", node);
node               84 drivers/clk/clk-moxart.c 	pll_clk = of_clk_get(node, 0);
node               86 drivers/clk/clk-moxart.c 		pr_err("%pOF: of_clk_get failed\n", node);
node               92 drivers/clk/clk-moxart.c 		pr_err("%pOF: failed to register clock\n", node);
node               97 drivers/clk/clk-moxart.c 	of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node               63 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup(struct device_node *node,
node               69 drivers/clk/clk-nspire.c 	const char *clk_name = node->name;
node               73 drivers/clk/clk-nspire.c 	io = of_iomap(node, 0);
node               81 drivers/clk/clk-nspire.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               82 drivers/clk/clk-nspire.c 	parent_name = of_clk_get_parent_name(node, 0);
node               87 drivers/clk/clk-nspire.c 		of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node               90 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup_cx(struct device_node *node)
node               92 drivers/clk/clk-nspire.c 	nspire_ahbdiv_setup(node, nspire_clkinfo_cx);
node               95 drivers/clk/clk-nspire.c static void __init nspire_ahbdiv_setup_classic(struct device_node *node)
node               97 drivers/clk/clk-nspire.c 	nspire_ahbdiv_setup(node, nspire_clkinfo_classic);
node              105 drivers/clk/clk-nspire.c static void __init nspire_clk_setup(struct device_node *node,
node              111 drivers/clk/clk-nspire.c 	const char *clk_name = node->name;
node              114 drivers/clk/clk-nspire.c 	io = of_iomap(node, 0);
node              122 drivers/clk/clk-nspire.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              127 drivers/clk/clk-nspire.c 		of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node              137 drivers/clk/clk-nspire.c static void __init nspire_clk_setup_cx(struct device_node *node)
node              139 drivers/clk/clk-nspire.c 	nspire_clk_setup(node, nspire_clkinfo_cx);
node              142 drivers/clk/clk-nspire.c static void __init nspire_clk_setup_classic(struct device_node *node)
node              144 drivers/clk/clk-nspire.c 	nspire_clk_setup(node, nspire_clkinfo_classic);
node              178 drivers/clk/clk-palmas.c 	struct device_node *node = pdev->dev.of_node;
node              182 drivers/clk/clk-palmas.c 	ret = of_property_read_u32(node, "ti,external-sleep-control",
node              199 drivers/clk/clk-palmas.c 			 node, prop);
node              242 drivers/clk/clk-palmas.c 	struct device_node *node = pdev->dev.of_node;
node              276 drivers/clk/clk-palmas.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_simple_get, &cinfo->hw);
node               69 drivers/clk/clk-pwm.c 	struct device_node *node = pdev->dev.of_node;
node               91 drivers/clk/clk-pwm.c 	if (of_property_read_u32(node, "clock-frequency", &clk_pwm->fixed_rate))
node              110 drivers/clk/clk-pwm.c 	clk_name = node->name;
node              111 drivers/clk/clk-pwm.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              124 drivers/clk/clk-pwm.c 	return of_clk_add_hw_provider(node, of_clk_hw_simple_get, &clk_pwm->hw);
node               86 drivers/clk/clk-qoriq.c 	struct device_node *node;
node             1003 drivers/clk/clk-qoriq.c 	if (!clockgen.node)
node             1031 drivers/clk/clk-qoriq.c *sysclk_from_fixed(struct device_node *node, const char *name)
node             1035 drivers/clk/clk-qoriq.c 	if (of_property_read_u32(node, "clock-frequency", &rate))
node             1061 drivers/clk/clk-qoriq.c 	clk = of_clk_get_by_name(clockgen.node, dtname);
node             1072 drivers/clk/clk-qoriq.c 	clk = of_clk_get(clockgen.node, 0);
node             1084 drivers/clk/clk-qoriq.c 	clk = sysclk_from_fixed(clockgen.node, name);
node             1096 drivers/clk/clk-qoriq.c 	sysclk = of_get_child_by_name(clockgen.node, "sysclk");
node             1128 drivers/clk/clk-qoriq.c static void __init sysclk_init(struct device_node *node)
node             1132 drivers/clk/clk-qoriq.c 	legacy_init_clockgen(node);
node             1136 drivers/clk/clk-qoriq.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node             1427 drivers/clk/clk-qoriq.c 	if (clockgen.node)
node             1430 drivers/clk/clk-qoriq.c 	clockgen.node = np;
node              157 drivers/clk/clk-rk808.c 	struct device_node *node = client->dev.of_node;
node              176 drivers/clk/clk-rk808.c 	of_property_read_string_index(node, "clock-output-names",
node              188 drivers/clk/clk-rk808.c 	of_property_read_string_index(node, "clock-output-names",
node              223 drivers/clk/clk-vt8500.c static __init void vtwm_device_clk_init(struct device_node *node)
node              228 drivers/clk/clk-vt8500.c 	const char *clk_name = node->name;
node              243 drivers/clk/clk-vt8500.c 	rc = of_property_read_u32(node, "enable-reg", &en_reg);
node              246 drivers/clk/clk-vt8500.c 		rc = of_property_read_u32(node, "enable-bit", &dev_clk->en_bit);
node              255 drivers/clk/clk-vt8500.c 	rc = of_property_read_u32(node, "divisor-reg", &div_reg);
node              264 drivers/clk/clk-vt8500.c 		of_property_read_u32(node, "divisor-mask", &dev_clk->div_mask);
node              268 drivers/clk/clk-vt8500.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              289 drivers/clk/clk-vt8500.c 	parent_name = of_clk_get_parent_name(node, 0);
node              301 drivers/clk/clk-vt8500.c 	rc = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node              673 drivers/clk/clk-vt8500.c static __init void vtwm_pll_clk_init(struct device_node *node, int pll_type)
node              678 drivers/clk/clk-vt8500.c 	const char *clk_name = node->name;
node              686 drivers/clk/clk-vt8500.c 	rc = of_property_read_u32(node, "reg", &reg);
node              698 drivers/clk/clk-vt8500.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              703 drivers/clk/clk-vt8500.c 	parent_name = of_clk_get_parent_name(node, 0);
node              715 drivers/clk/clk-vt8500.c 	rc = of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node              722 drivers/clk/clk-vt8500.c static void __init vt8500_pll_init(struct device_node *node)
node              724 drivers/clk/clk-vt8500.c 	vtwm_pll_clk_init(node, PLL_TYPE_VT8500);
node              728 drivers/clk/clk-vt8500.c static void __init wm8650_pll_init(struct device_node *node)
node              730 drivers/clk/clk-vt8500.c 	vtwm_pll_clk_init(node, PLL_TYPE_WM8650);
node              734 drivers/clk/clk-vt8500.c static void __init wm8750_pll_init(struct device_node *node)
node              736 drivers/clk/clk-vt8500.c 	vtwm_pll_clk_init(node, PLL_TYPE_WM8750);
node              740 drivers/clk/clk-vt8500.c static void __init wm8850_pll_init(struct device_node *node)
node              742 drivers/clk/clk-vt8500.c 	vtwm_pll_clk_init(node, PLL_TYPE_WM8850);
node             1482 drivers/clk/clk.c 	list_for_each_entry(cn, &clk_notifier_list, node) {
node             3813 drivers/clk/clk.c int of_clk_hw_register(struct device_node *node, struct clk_hw *hw)
node             3815 drivers/clk/clk.c 	return PTR_ERR_OR_ZERO(__clk_register(NULL, node, hw));
node             4152 drivers/clk/clk.c 	list_for_each_entry(cn, &clk_notifier_list, node)
node             4165 drivers/clk/clk.c 		list_add(&cn->node, &clk_notifier_list);
node             4200 drivers/clk/clk.c 	list_for_each_entry(cn, &clk_notifier_list, node)
node             4212 drivers/clk/clk.c 			list_del(&cn->node);
node             4245 drivers/clk/clk.c 	struct device_node *node;
node             4320 drivers/clk/clk.c 	cp->node = of_node_get(np);
node             4357 drivers/clk/clk.c 	cp->node = of_node_get(np);
node             4450 drivers/clk/clk.c 		if (cp->node == np) {
node             4452 drivers/clk/clk.c 			of_node_put(cp->node);
node             4587 drivers/clk/clk.c 		if (provider->node == clkspec->np) {
node             4766 drivers/clk/clk.c 	struct list_head node;
node             4869 drivers/clk/clk.c 						 &clk_provider_list, node) {
node             4870 drivers/clk/clk.c 				list_del(&clk_provider->node);
node             4880 drivers/clk/clk.c 		list_add_tail(&parent->node, &clk_provider_list);
node             4886 drivers/clk/clk.c 					&clk_provider_list, node) {
node             4896 drivers/clk/clk.c 				list_del(&clk_provider->node);
node               48 drivers/clk/clkdev.c 	list_for_each_entry(p, &clocks, node) {
node              124 drivers/clk/clkdev.c 	list_add_tail(&cl->node, &clocks);
node              141 drivers/clk/clkdev.c 		list_add_tail(&cl->node, &clocks);
node              290 drivers/clk/clkdev.c 	list_del(&cl->node);
node              142 drivers/clk/davinci/pll-da850.c void of_da850_pll0_init(struct device_node *node)
node              147 drivers/clk/davinci/pll-da850.c 	base = of_iomap(node, 0);
node              155 drivers/clk/davinci/pll-da850.c 	of_davinci_pll_init(NULL, node, &da850_pll0_info,
node              745 drivers/clk/davinci/pll.c int of_davinci_pll_init(struct device *dev, struct device_node *node,
node              758 drivers/clk/davinci/pll.c 		parent_name = of_clk_get_parent_name(node, 0);
node              768 drivers/clk/davinci/pll.c 	child = of_get_child_by_name(node, "pllout");
node              773 drivers/clk/davinci/pll.c 	child = of_get_child_by_name(node, "sysclk");
node              811 drivers/clk/davinci/pll.c 	child = of_get_child_by_name(node, "auxclk");
node              826 drivers/clk/davinci/pll.c 	child = of_get_child_by_name(node, "obsclk");
node              115 drivers/clk/davinci/pll.h int of_davinci_pll_init(struct device *dev, struct device_node *node,
node              127 drivers/clk/davinci/pll.h void of_da850_pll0_init(struct device_node *node);
node              483 drivers/clk/davinci/psc.c 	struct device_node *node = dev->of_node;
node              490 drivers/clk/davinci/psc.c 	of_genpd_add_provider_onecell(node, &psc->pm_data);
node              492 drivers/clk/davinci/psc.c 	of_clk_add_provider(node, of_clk_src_onecell_get, &psc->clk_data);
node               16 drivers/clk/h8300/clk-div.c static void __init h8300_div_clk_setup(struct device_node *node)
node               20 drivers/clk/h8300/clk-div.c 	const char *clk_name = node->name;
node               26 drivers/clk/h8300/clk-div.c 	num_parents = of_clk_get_parent_count(node);
node               32 drivers/clk/h8300/clk-div.c 	divcr = of_iomap(node, 0);
node               41 drivers/clk/h8300/clk-div.c 	parent_name = of_clk_get_parent_name(node, 0);
node               42 drivers/clk/h8300/clk-div.c 	of_property_read_u32(node, "renesas,width", &width);
node               47 drivers/clk/h8300/clk-div.c 		of_clk_add_hw_provider(node, of_clk_hw_simple_get, hw);
node               86 drivers/clk/h8300/clk-h8s2678.c static void __init h8s2678_pll_clk_setup(struct device_node *node)
node               89 drivers/clk/h8300/clk-h8s2678.c 	const char *clk_name = node->name;
node               95 drivers/clk/h8300/clk-h8s2678.c 	num_parents = of_clk_get_parent_count(node);
node              106 drivers/clk/h8300/clk-h8s2678.c 	pll_clock->sckcr = of_iomap(node, 0);
node              112 drivers/clk/h8300/clk-h8s2678.c 	pll_clock->pllcr = of_iomap(node, 1);
node              118 drivers/clk/h8300/clk-h8s2678.c 	parent_name = of_clk_get_parent_name(node, 0);
node              133 drivers/clk/h8300/clk-h8s2678.c 	of_clk_add_hw_provider(node, of_clk_hw_simple_get, &pll_clock->hw);
node              448 drivers/clk/hisilicon/clk-hi3620.c static void __init hi3620_mmc_clk_init(struct device_node *node)
node              454 drivers/clk/hisilicon/clk-hi3620.c 	if (!node) {
node              459 drivers/clk/hisilicon/clk-hi3620.c 	base = of_iomap(node, 0);
node              476 drivers/clk/hisilicon/clk-hi3620.c 			hisi_register_clk_mmc(mmc_clk, base, node);
node              480 drivers/clk/hisilicon/clk-hi3620.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              170 drivers/clk/imx/clk-imx6q.c static void of_assigned_ldb_sels(struct device_node *node,
node              178 drivers/clk/imx/clk-imx6q.c 	num_parents = of_count_phandle_with_args(node, "assigned-clock-parents",
node              181 drivers/clk/imx/clk-imx6q.c 		rc = of_parse_phandle_with_args(node, "assigned-clock-parents",
node              190 drivers/clk/imx/clk-imx6q.c 		if (clkspec.np != node || clkspec.args[0] >= IMX6QDL_CLK_END) {
node              196 drivers/clk/imx/clk-imx6q.c 		rc = of_parse_phandle_with_args(node, "assigned-clocks",
node              200 drivers/clk/imx/clk-imx6q.c 		if (clkspec.np != node || clkspec.args[0] >= IMX6QDL_CLK_END) {
node              224 drivers/clk/imx/clk-imx6q.c static bool pll6_bypassed(struct device_node *node)
node              229 drivers/clk/imx/clk-imx6q.c 	num_clocks = of_count_phandle_with_args(node, "assigned-clocks",
node              235 drivers/clk/imx/clk-imx6q.c 		ret = of_parse_phandle_with_args(node, "assigned-clocks",
node              241 drivers/clk/imx/clk-imx6q.c 		if (clkspec.np == node &&
node              250 drivers/clk/imx/clk-imx6q.c 	ret = of_parse_phandle_with_args(node, "assigned-clock-parents",
node              197 drivers/clk/keystone/gate.c static void __init of_psc_clk_init(struct device_node *node, spinlock_t *lock)
node              199 drivers/clk/keystone/gate.c 	const char *clk_name = node->name;
node              211 drivers/clk/keystone/gate.c 	i = of_property_match_string(node, "reg-names", "control");
node              212 drivers/clk/keystone/gate.c 	data->control_base = of_iomap(node, i);
node              218 drivers/clk/keystone/gate.c 	i = of_property_match_string(node, "reg-names", "domain");
node              219 drivers/clk/keystone/gate.c 	data->domain_base = of_iomap(node, i);
node              225 drivers/clk/keystone/gate.c 	of_property_read_u32(node, "domain-id", &data->domain_id);
node              231 drivers/clk/keystone/gate.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              232 drivers/clk/keystone/gate.c 	parent_name = of_clk_get_parent_name(node, 0);
node              240 drivers/clk/keystone/gate.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              244 drivers/clk/keystone/gate.c 	pr_err("%s: error registering clk %pOFn\n", __func__, node);
node              259 drivers/clk/keystone/gate.c static void __init of_keystone_psc_clk_init(struct device_node *node)
node              261 drivers/clk/keystone/gate.c 	of_psc_clk_init(node, &psc_lock);
node              158 drivers/clk/keystone/pll.c static void __init _of_pll_clk_init(struct device_node *node, bool pllctrl)
node              171 drivers/clk/keystone/pll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              172 drivers/clk/keystone/pll.c 	if (of_property_read_u32(node, "fixed-postdiv",	&pll_data->postdiv)) {
node              181 drivers/clk/keystone/pll.c 		i = of_property_match_string(node, "reg-names",
node              183 drivers/clk/keystone/pll.c 		pll_data->pllod = of_iomap(node, i);
node              186 drivers/clk/keystone/pll.c 	i = of_property_match_string(node, "reg-names", "control");
node              187 drivers/clk/keystone/pll.c 	pll_data->pll_ctl0 = of_iomap(node, i);
node              202 drivers/clk/keystone/pll.c 		i = of_property_match_string(node, "reg-names", "multiplier");
node              203 drivers/clk/keystone/pll.c 		pll_data->pllm = of_iomap(node, i);
node              211 drivers/clk/keystone/pll.c 	clk = clk_register_pll(NULL, node->name, parent_name, pll_data);
node              213 drivers/clk/keystone/pll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              218 drivers/clk/keystone/pll.c 	pr_err("%s: error initializing pll %pOFn\n", __func__, node);
node              226 drivers/clk/keystone/pll.c static void __init of_keystone_pll_clk_init(struct device_node *node)
node              228 drivers/clk/keystone/pll.c 	_of_pll_clk_init(node, false);
node              237 drivers/clk/keystone/pll.c static void __init of_keystone_main_pll_clk_init(struct device_node *node)
node              239 drivers/clk/keystone/pll.c 	_of_pll_clk_init(node, true);
node              248 drivers/clk/keystone/pll.c static void __init of_pll_div_clk_init(struct device_node *node)
node              254 drivers/clk/keystone/pll.c 	const char *clk_name = node->name;
node              256 drivers/clk/keystone/pll.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              257 drivers/clk/keystone/pll.c 	reg = of_iomap(node, 0);
node              263 drivers/clk/keystone/pll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              270 drivers/clk/keystone/pll.c 	if (of_property_read_u32(node, "bit-shift", &shift)) {
node              276 drivers/clk/keystone/pll.c 	if (of_property_read_u32(node, "bit-mask", &mask)) {
node              285 drivers/clk/keystone/pll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              297 drivers/clk/keystone/pll.c static void __init of_pll_mux_clk_init(struct device_node *node)
node              303 drivers/clk/keystone/pll.c 	const char *clk_name = node->name;
node              305 drivers/clk/keystone/pll.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              306 drivers/clk/keystone/pll.c 	reg = of_iomap(node, 0);
node              312 drivers/clk/keystone/pll.c 	of_clk_parent_fill(node, parents, 2);
node              318 drivers/clk/keystone/pll.c 	if (of_property_read_u32(node, "bit-shift", &shift)) {
node              323 drivers/clk/keystone/pll.c 	if (of_property_read_u32(node, "bit-mask", &mask)) {
node              332 drivers/clk/keystone/pll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               65 drivers/clk/keystone/sci-clk.c 	struct list_head node;
node              494 drivers/clk/keystone/sci-clk.c 	struct sci_clk *ca = container_of(a, struct sci_clk, node);
node              495 drivers/clk/keystone/sci-clk.c 	struct sci_clk *cb = container_of(b, struct sci_clk, node);
node              553 drivers/clk/keystone/sci-clk.c 				list_add_tail(&sci_clk->node, &clks);
node              589 drivers/clk/keystone/sci-clk.c 					list_add_tail(&sci_clk->node, &clks);
node              609 drivers/clk/keystone/sci-clk.c 	list_for_each_entry(sci_clk, &clks, node) {
node               79 drivers/clk/mediatek/clk-cpumux.c int mtk_clk_register_cpumuxes(struct device_node *node,
node               87 drivers/clk/mediatek/clk-cpumux.c 	regmap = syscon_node_to_regmap(node);
node               89 drivers/clk/mediatek/clk-cpumux.c 		pr_err("Cannot find regmap for %pOF: %ld\n", node,
node               18 drivers/clk/mediatek/clk-cpumux.h int mtk_clk_register_cpumuxes(struct device_node *node,
node              149 drivers/clk/mediatek/clk-mt2701-aud.c 	struct device_node *node = pdev->dev.of_node;
node              154 drivers/clk/mediatek/clk-mt2701-aud.c 	mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks),
node              157 drivers/clk/mediatek/clk-mt2701-aud.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              173 drivers/clk/mediatek/clk-mt2701-aud.c 	of_clk_del_provider(node);
node              106 drivers/clk/mediatek/clk-mt2701-bdp.c 	struct device_node *node = pdev->dev.of_node;
node              110 drivers/clk/mediatek/clk-mt2701-bdp.c 	mtk_clk_register_gates(node, bdp_clks, ARRAY_SIZE(bdp_clks),
node              113 drivers/clk/mediatek/clk-mt2701-bdp.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               48 drivers/clk/mediatek/clk-mt2701-eth.c 	struct device_node *node = pdev->dev.of_node;
node               52 drivers/clk/mediatek/clk-mt2701-eth.c 	mtk_clk_register_gates(node, eth_clks, ARRAY_SIZE(eth_clks),
node               55 drivers/clk/mediatek/clk-mt2701-eth.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               61 drivers/clk/mediatek/clk-mt2701-eth.c 	mtk_register_reset_controller(node, 1, 0x34);
node               41 drivers/clk/mediatek/clk-mt2701-g3d.c 	struct device_node *node = pdev->dev.of_node;
node               46 drivers/clk/mediatek/clk-mt2701-g3d.c 	mtk_clk_register_gates(node, g3d_clks, ARRAY_SIZE(g3d_clks),
node               49 drivers/clk/mediatek/clk-mt2701-g3d.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               55 drivers/clk/mediatek/clk-mt2701-g3d.c 	mtk_register_reset_controller(node, 1, 0xc);
node               45 drivers/clk/mediatek/clk-mt2701-hif.c 	struct device_node *node = pdev->dev.of_node;
node               49 drivers/clk/mediatek/clk-mt2701-hif.c 	mtk_clk_register_gates(node, hif_clks, ARRAY_SIZE(hif_clks),
node               52 drivers/clk/mediatek/clk-mt2701-hif.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               60 drivers/clk/mediatek/clk-mt2701-hif.c 	mtk_register_reset_controller(node, 1, 0x34);
node               48 drivers/clk/mediatek/clk-mt2701-img.c 	struct device_node *node = pdev->dev.of_node;
node               52 drivers/clk/mediatek/clk-mt2701-img.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node               55 drivers/clk/mediatek/clk-mt2701-img.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               91 drivers/clk/mediatek/clk-mt2701-mm.c 	struct device_node *node = pdev->dev.of_node;
node               95 drivers/clk/mediatek/clk-mt2701-mm.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node               98 drivers/clk/mediatek/clk-mt2701-mm.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               59 drivers/clk/mediatek/clk-mt2701-vdec.c 	struct device_node *node = pdev->dev.of_node;
node               63 drivers/clk/mediatek/clk-mt2701-vdec.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node               66 drivers/clk/mediatek/clk-mt2701-vdec.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              670 drivers/clk/mediatek/clk-mt2701.c 	struct device_node *node = pdev->dev.of_node;
node              691 drivers/clk/mediatek/clk-mt2701.c 	mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks),
node              694 drivers/clk/mediatek/clk-mt2701.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              739 drivers/clk/mediatek/clk-mt2701.c static void __init mtk_infrasys_init_early(struct device_node *node)
node              753 drivers/clk/mediatek/clk-mt2701.c 	mtk_clk_register_cpumuxes(node, cpu_muxes, ARRAY_SIZE(cpu_muxes),
node              756 drivers/clk/mediatek/clk-mt2701.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data);
node              767 drivers/clk/mediatek/clk-mt2701.c 	struct device_node *node = pdev->dev.of_node;
node              778 drivers/clk/mediatek/clk-mt2701.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              783 drivers/clk/mediatek/clk-mt2701.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data);
node              787 drivers/clk/mediatek/clk-mt2701.c 	mtk_register_reset_controller(node, 2, 0x30);
node              891 drivers/clk/mediatek/clk-mt2701.c 	struct device_node *node = pdev->dev.of_node;
node              900 drivers/clk/mediatek/clk-mt2701.c 	mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks),
node              906 drivers/clk/mediatek/clk-mt2701.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              910 drivers/clk/mediatek/clk-mt2701.c 	mtk_register_reset_controller(node, 2, 0x0);
node              972 drivers/clk/mediatek/clk-mt2701.c 	struct device_node *node = pdev->dev.of_node;
node              978 drivers/clk/mediatek/clk-mt2701.c 	mtk_clk_register_plls(node, apmixed_plls, ARRAY_SIZE(apmixed_plls),
node              983 drivers/clk/mediatek/clk-mt2701.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               65 drivers/clk/mediatek/clk-mt2712-bdp.c 	struct device_node *node = pdev->dev.of_node;
node               69 drivers/clk/mediatek/clk-mt2712-bdp.c 	mtk_clk_register_gates(node, bdp_clks, ARRAY_SIZE(bdp_clks),
node               72 drivers/clk/mediatek/clk-mt2712-bdp.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               43 drivers/clk/mediatek/clk-mt2712-img.c 	struct device_node *node = pdev->dev.of_node;
node               47 drivers/clk/mediatek/clk-mt2712-img.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node               50 drivers/clk/mediatek/clk-mt2712-img.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               39 drivers/clk/mediatek/clk-mt2712-jpgdec.c 	struct device_node *node = pdev->dev.of_node;
node               43 drivers/clk/mediatek/clk-mt2712-jpgdec.c 	mtk_clk_register_gates(node, jpgdec_clks, ARRAY_SIZE(jpgdec_clks),
node               46 drivers/clk/mediatek/clk-mt2712-jpgdec.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               38 drivers/clk/mediatek/clk-mt2712-mfg.c 	struct device_node *node = pdev->dev.of_node;
node               42 drivers/clk/mediatek/clk-mt2712-mfg.c 	mtk_clk_register_gates(node, mfg_clks, ARRAY_SIZE(mfg_clks),
node               45 drivers/clk/mediatek/clk-mt2712-mfg.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              133 drivers/clk/mediatek/clk-mt2712-mm.c 	struct device_node *node = pdev->dev.of_node;
node              137 drivers/clk/mediatek/clk-mt2712-mm.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node              140 drivers/clk/mediatek/clk-mt2712-mm.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               57 drivers/clk/mediatek/clk-mt2712-vdec.c 	struct device_node *node = pdev->dev.of_node;
node               61 drivers/clk/mediatek/clk-mt2712-vdec.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node               64 drivers/clk/mediatek/clk-mt2712-vdec.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               40 drivers/clk/mediatek/clk-mt2712-venc.c 	struct device_node *node = pdev->dev.of_node;
node               44 drivers/clk/mediatek/clk-mt2712-venc.c 	mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks),
node               47 drivers/clk/mediatek/clk-mt2712-venc.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1264 drivers/clk/mediatek/clk-mt2712.c 	struct device_node *node = pdev->dev.of_node;
node             1268 drivers/clk/mediatek/clk-mt2712.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node             1270 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1281 drivers/clk/mediatek/clk-mt2712.c static void clk_mt2712_top_init_early(struct device_node *node)
node             1295 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data);
node             1307 drivers/clk/mediatek/clk-mt2712.c 	struct device_node *node = pdev->dev.of_node;
node             1335 drivers/clk/mediatek/clk-mt2712.c 	mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks),
node             1338 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data);
node             1351 drivers/clk/mediatek/clk-mt2712.c 	struct device_node *node = pdev->dev.of_node;
node             1355 drivers/clk/mediatek/clk-mt2712.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node             1358 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1364 drivers/clk/mediatek/clk-mt2712.c 	mtk_register_reset_controller(node, 2, 0x30);
node             1373 drivers/clk/mediatek/clk-mt2712.c 	struct device_node *node = pdev->dev.of_node;
node             1377 drivers/clk/mediatek/clk-mt2712.c 	mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks),
node             1380 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1386 drivers/clk/mediatek/clk-mt2712.c 	mtk_register_reset_controller(node, 2, 0);
node             1395 drivers/clk/mediatek/clk-mt2712.c 	struct device_node *node = pdev->dev.of_node;
node             1410 drivers/clk/mediatek/clk-mt2712.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               99 drivers/clk/mediatek/clk-mt6779-aud.c 	struct device_node *node = pdev->dev.of_node;
node              103 drivers/clk/mediatek/clk-mt6779-aud.c 	mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks),
node              106 drivers/clk/mediatek/clk-mt6779-aud.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               48 drivers/clk/mediatek/clk-mt6779-cam.c 	struct device_node *node = pdev->dev.of_node;
node               52 drivers/clk/mediatek/clk-mt6779-cam.c 	mtk_clk_register_gates(node, cam_clks, ARRAY_SIZE(cam_clks),
node               55 drivers/clk/mediatek/clk-mt6779-cam.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               40 drivers/clk/mediatek/clk-mt6779-img.c 	struct device_node *node = pdev->dev.of_node;
node               44 drivers/clk/mediatek/clk-mt6779-img.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node               47 drivers/clk/mediatek/clk-mt6779-img.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               42 drivers/clk/mediatek/clk-mt6779-ipe.c 	struct device_node *node = pdev->dev.of_node;
node               46 drivers/clk/mediatek/clk-mt6779-ipe.c 	mtk_clk_register_gates(node, ipe_clks, ARRAY_SIZE(ipe_clks),
node               49 drivers/clk/mediatek/clk-mt6779-ipe.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               32 drivers/clk/mediatek/clk-mt6779-mfg.c 	struct device_node *node = pdev->dev.of_node;
node               36 drivers/clk/mediatek/clk-mt6779-mfg.c 	mtk_clk_register_gates(node, mfg_clks, ARRAY_SIZE(mfg_clks),
node               39 drivers/clk/mediatek/clk-mt6779-mfg.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               95 drivers/clk/mediatek/clk-mt6779-mm.c 	struct device_node *node = pdev->dev.of_node;
node               99 drivers/clk/mediatek/clk-mt6779-mm.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node              102 drivers/clk/mediatek/clk-mt6779-mm.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               49 drivers/clk/mediatek/clk-mt6779-vdec.c 	struct device_node *node = pdev->dev.of_node;
node               53 drivers/clk/mediatek/clk-mt6779-vdec.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node               56 drivers/clk/mediatek/clk-mt6779-vdec.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               40 drivers/clk/mediatek/clk-mt6779-venc.c 	struct device_node *node = pdev->dev.of_node;
node               44 drivers/clk/mediatek/clk-mt6779-venc.c 	mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks),
node               47 drivers/clk/mediatek/clk-mt6779-venc.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1214 drivers/clk/mediatek/clk-mt6779.c 	struct device_node *node = pdev->dev.of_node;
node             1218 drivers/clk/mediatek/clk-mt6779.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node             1220 drivers/clk/mediatek/clk-mt6779.c 	mtk_clk_register_gates(node, apmixed_clks,
node             1223 drivers/clk/mediatek/clk-mt6779.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1231 drivers/clk/mediatek/clk-mt6779.c 	struct device_node *node = pdev->dev.of_node;
node             1245 drivers/clk/mediatek/clk-mt6779.c 			       node, &mt6779_clk_lock, clk_data);
node             1253 drivers/clk/mediatek/clk-mt6779.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1259 drivers/clk/mediatek/clk-mt6779.c 	struct device_node *node = pdev->dev.of_node;
node             1263 drivers/clk/mediatek/clk-mt6779.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node             1266 drivers/clk/mediatek/clk-mt6779.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               44 drivers/clk/mediatek/clk-mt6797-img.c 	struct device_node *node = pdev->dev.of_node;
node               48 drivers/clk/mediatek/clk-mt6797-img.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node               51 drivers/clk/mediatek/clk-mt6797-img.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              104 drivers/clk/mediatek/clk-mt6797-mm.c 	struct device_node *node = pdev->dev.of_node;
node              108 drivers/clk/mediatek/clk-mt6797-mm.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node              111 drivers/clk/mediatek/clk-mt6797-mm.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               61 drivers/clk/mediatek/clk-mt6797-vdec.c 	struct device_node *node = pdev->dev.of_node;
node               65 drivers/clk/mediatek/clk-mt6797-vdec.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node               68 drivers/clk/mediatek/clk-mt6797-vdec.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               46 drivers/clk/mediatek/clk-mt6797-venc.c 	struct device_node *node = pdev->dev.of_node;
node               50 drivers/clk/mediatek/clk-mt6797-venc.c 	mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks),
node               53 drivers/clk/mediatek/clk-mt6797-venc.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              387 drivers/clk/mediatek/clk-mt6797.c 	struct device_node *node = pdev->dev.of_node;
node              402 drivers/clk/mediatek/clk-mt6797.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              561 drivers/clk/mediatek/clk-mt6797.c static void mtk_infrasys_init_early(struct device_node *node)
node              575 drivers/clk/mediatek/clk-mt6797.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data);
node              587 drivers/clk/mediatek/clk-mt6797.c 	struct device_node *node = pdev->dev.of_node;
node              598 drivers/clk/mediatek/clk-mt6797.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              603 drivers/clk/mediatek/clk-mt6797.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, infra_clk_data);
node              667 drivers/clk/mediatek/clk-mt6797.c 	struct device_node *node = pdev->dev.of_node;
node              673 drivers/clk/mediatek/clk-mt6797.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node              675 drivers/clk/mediatek/clk-mt6797.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              136 drivers/clk/mediatek/clk-mt7622-aud.c 	struct device_node *node = pdev->dev.of_node;
node              141 drivers/clk/mediatek/clk-mt7622-aud.c 	mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks),
node              144 drivers/clk/mediatek/clk-mt7622-aud.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              160 drivers/clk/mediatek/clk-mt7622-aud.c 	of_clk_del_provider(node);
node               71 drivers/clk/mediatek/clk-mt7622-eth.c 	struct device_node *node = pdev->dev.of_node;
node               76 drivers/clk/mediatek/clk-mt7622-eth.c 	mtk_clk_register_gates(node, eth_clks, ARRAY_SIZE(eth_clks),
node               79 drivers/clk/mediatek/clk-mt7622-eth.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               85 drivers/clk/mediatek/clk-mt7622-eth.c 	mtk_register_reset_controller(node, 1, 0x34);
node               93 drivers/clk/mediatek/clk-mt7622-eth.c 	struct device_node *node = pdev->dev.of_node;
node               98 drivers/clk/mediatek/clk-mt7622-eth.c 	mtk_clk_register_gates(node, sgmii_clks, ARRAY_SIZE(sgmii_clks),
node              101 drivers/clk/mediatek/clk-mt7622-eth.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               82 drivers/clk/mediatek/clk-mt7622-hif.c 	struct device_node *node = pdev->dev.of_node;
node               87 drivers/clk/mediatek/clk-mt7622-hif.c 	mtk_clk_register_gates(node, ssusb_clks, ARRAY_SIZE(ssusb_clks),
node               90 drivers/clk/mediatek/clk-mt7622-hif.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               96 drivers/clk/mediatek/clk-mt7622-hif.c 	mtk_register_reset_controller(node, 1, 0x34);
node              104 drivers/clk/mediatek/clk-mt7622-hif.c 	struct device_node *node = pdev->dev.of_node;
node              109 drivers/clk/mediatek/clk-mt7622-hif.c 	mtk_clk_register_gates(node, pcie_clks, ARRAY_SIZE(pcie_clks),
node              112 drivers/clk/mediatek/clk-mt7622-hif.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              118 drivers/clk/mediatek/clk-mt7622-hif.c 	mtk_register_reset_controller(node, 1, 0x34);
node              616 drivers/clk/mediatek/clk-mt7622.c 	struct device_node *node = pdev->dev.of_node;
node              637 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks),
node              644 drivers/clk/mediatek/clk-mt7622.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              649 drivers/clk/mediatek/clk-mt7622.c 	struct device_node *node = pdev->dev.of_node;
node              655 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              658 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_cpumuxes(node, infra_muxes, ARRAY_SIZE(infra_muxes),
node              661 drivers/clk/mediatek/clk-mt7622.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get,
node              666 drivers/clk/mediatek/clk-mt7622.c 	mtk_register_reset_controller(node, 1, 0x30);
node              674 drivers/clk/mediatek/clk-mt7622.c 	struct device_node *node = pdev->dev.of_node;
node              680 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls),
node              683 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_gates(node, apmixed_clks,
node              689 drivers/clk/mediatek/clk-mt7622.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              697 drivers/clk/mediatek/clk-mt7622.c 	struct device_node *node = pdev->dev.of_node;
node              706 drivers/clk/mediatek/clk-mt7622.c 	mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks),
node              712 drivers/clk/mediatek/clk-mt7622.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              718 drivers/clk/mediatek/clk-mt7622.c 	mtk_register_reset_controller(node, 2, 0x0);
node               82 drivers/clk/mediatek/clk-mt7629-eth.c 	struct device_node *node = pdev->dev.of_node;
node               87 drivers/clk/mediatek/clk-mt7629-eth.c 	mtk_clk_register_gates(node, eth_clks, CLK_ETH_NR_CLK, clk_data);
node               89 drivers/clk/mediatek/clk-mt7629-eth.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               95 drivers/clk/mediatek/clk-mt7629-eth.c 	mtk_register_reset_controller(node, 1, 0x34);
node              103 drivers/clk/mediatek/clk-mt7629-eth.c 	struct device_node *node = pdev->dev.of_node;
node              109 drivers/clk/mediatek/clk-mt7629-eth.c 	mtk_clk_register_gates(node, sgmii_clks[id++], CLK_SGMII_NR_CLK,
node              112 drivers/clk/mediatek/clk-mt7629-eth.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               77 drivers/clk/mediatek/clk-mt7629-hif.c 	struct device_node *node = pdev->dev.of_node;
node               82 drivers/clk/mediatek/clk-mt7629-hif.c 	mtk_clk_register_gates(node, ssusb_clks, ARRAY_SIZE(ssusb_clks),
node               85 drivers/clk/mediatek/clk-mt7629-hif.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               91 drivers/clk/mediatek/clk-mt7629-hif.c 	mtk_register_reset_controller(node, 1, 0x34);
node               99 drivers/clk/mediatek/clk-mt7629-hif.c 	struct device_node *node = pdev->dev.of_node;
node              104 drivers/clk/mediatek/clk-mt7629-hif.c 	mtk_clk_register_gates(node, pcie_clks, ARRAY_SIZE(pcie_clks),
node              107 drivers/clk/mediatek/clk-mt7629-hif.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              113 drivers/clk/mediatek/clk-mt7629-hif.c 	mtk_register_reset_controller(node, 1, 0x34);
node              576 drivers/clk/mediatek/clk-mt7629.c 	struct device_node *node = pdev->dev.of_node;
node              598 drivers/clk/mediatek/clk-mt7629.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              603 drivers/clk/mediatek/clk-mt7629.c 	struct device_node *node = pdev->dev.of_node;
node              609 drivers/clk/mediatek/clk-mt7629.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              612 drivers/clk/mediatek/clk-mt7629.c 	mtk_clk_register_cpumuxes(node, infra_muxes, ARRAY_SIZE(infra_muxes),
node              615 drivers/clk/mediatek/clk-mt7629.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get,
node              628 drivers/clk/mediatek/clk-mt7629.c 	struct device_node *node = pdev->dev.of_node;
node              637 drivers/clk/mediatek/clk-mt7629.c 	mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks),
node              643 drivers/clk/mediatek/clk-mt7629.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              655 drivers/clk/mediatek/clk-mt7629.c 	struct device_node *node = pdev->dev.of_node;
node              661 drivers/clk/mediatek/clk-mt7629.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls),
node              664 drivers/clk/mediatek/clk-mt7629.c 	mtk_clk_register_gates(node, apmixed_clks,
node              670 drivers/clk/mediatek/clk-mt7629.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              516 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_topckgen_init(struct device_node *node)
node              522 drivers/clk/mediatek/clk-mt8135.c 	base = of_iomap(node, 0);
node              537 drivers/clk/mediatek/clk-mt8135.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              544 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_infrasys_init(struct device_node *node)
node              551 drivers/clk/mediatek/clk-mt8135.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              556 drivers/clk/mediatek/clk-mt8135.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              561 drivers/clk/mediatek/clk-mt8135.c 	mtk_register_reset_controller(node, 2, 0x30);
node              565 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_pericfg_init(struct device_node *node)
node              571 drivers/clk/mediatek/clk-mt8135.c 	base = of_iomap(node, 0);
node              579 drivers/clk/mediatek/clk-mt8135.c 	mtk_clk_register_gates(node, peri_gates, ARRAY_SIZE(peri_gates),
node              584 drivers/clk/mediatek/clk-mt8135.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              589 drivers/clk/mediatek/clk-mt8135.c 	mtk_register_reset_controller(node, 2, 0);
node              626 drivers/clk/mediatek/clk-mt8135.c static void __init mtk_apmixedsys_init(struct device_node *node)
node              634 drivers/clk/mediatek/clk-mt8135.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node              924 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_topckgen_init(struct device_node *node)
node              930 drivers/clk/mediatek/clk-mt8173.c 	base = of_iomap(node, 0);
node              943 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              952 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_infrasys_init(struct device_node *node)
node              959 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node              963 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_cpumuxes(node, cpu_muxes, ARRAY_SIZE(cpu_muxes),
node              966 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              971 drivers/clk/mediatek/clk-mt8173.c 	mtk_register_reset_controller(node, 2, 0x30);
node              975 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_pericfg_init(struct device_node *node)
node              981 drivers/clk/mediatek/clk-mt8173.c 	base = of_iomap(node, 0);
node              989 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, peri_gates, ARRAY_SIZE(peri_gates),
node              994 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              999 drivers/clk/mediatek/clk-mt8173.c 	mtk_register_reset_controller(node, 2, 0);
node             1078 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_apmixedsys_init(struct device_node *node)
node             1085 drivers/clk/mediatek/clk-mt8173.c 	base = of_iomap(node, 0);
node             1097 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node             1119 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1129 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_imgsys_init(struct device_node *node)
node             1136 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node             1139 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1147 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_mmsys_init(struct device_node *node)
node             1154 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node             1157 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1164 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vdecsys_init(struct device_node *node)
node             1171 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node             1174 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1181 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vencsys_init(struct device_node *node)
node             1188 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks),
node             1191 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1198 drivers/clk/mediatek/clk-mt8173.c static void __init mtk_vencltsys_init(struct device_node *node)
node             1205 drivers/clk/mediatek/clk-mt8173.c 	mtk_clk_register_gates(node, venclt_clks, ARRAY_SIZE(venclt_clks),
node             1208 drivers/clk/mediatek/clk-mt8173.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               74 drivers/clk/mediatek/clk-mt8183-audio.c 	struct device_node *node = pdev->dev.of_node;
node               78 drivers/clk/mediatek/clk-mt8183-audio.c 	mtk_clk_register_gates(node, audio_clks, ARRAY_SIZE(audio_clks),
node               81 drivers/clk/mediatek/clk-mt8183-audio.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               87 drivers/clk/mediatek/clk-mt8183-audio.c 		of_clk_del_provider(node);
node               40 drivers/clk/mediatek/clk-mt8183-cam.c 	struct device_node *node = pdev->dev.of_node;
node               44 drivers/clk/mediatek/clk-mt8183-cam.c 	mtk_clk_register_gates(node, cam_clks, ARRAY_SIZE(cam_clks),
node               47 drivers/clk/mediatek/clk-mt8183-cam.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               40 drivers/clk/mediatek/clk-mt8183-img.c 	struct device_node *node = pdev->dev.of_node;
node               44 drivers/clk/mediatek/clk-mt8183-img.c 	mtk_clk_register_gates(node, img_clks, ARRAY_SIZE(img_clks),
node               47 drivers/clk/mediatek/clk-mt8183-img.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               33 drivers/clk/mediatek/clk-mt8183-ipu0.c 	struct device_node *node = pdev->dev.of_node;
node               37 drivers/clk/mediatek/clk-mt8183-ipu0.c 	mtk_clk_register_gates(node, ipu_core0_clks, ARRAY_SIZE(ipu_core0_clks),
node               40 drivers/clk/mediatek/clk-mt8183-ipu0.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               33 drivers/clk/mediatek/clk-mt8183-ipu1.c 	struct device_node *node = pdev->dev.of_node;
node               37 drivers/clk/mediatek/clk-mt8183-ipu1.c 	mtk_clk_register_gates(node, ipu_core1_clks, ARRAY_SIZE(ipu_core1_clks),
node               40 drivers/clk/mediatek/clk-mt8183-ipu1.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               31 drivers/clk/mediatek/clk-mt8183-ipu_adl.c 	struct device_node *node = pdev->dev.of_node;
node               35 drivers/clk/mediatek/clk-mt8183-ipu_adl.c 	mtk_clk_register_gates(node, ipu_adl_clks, ARRAY_SIZE(ipu_adl_clks),
node               38 drivers/clk/mediatek/clk-mt8183-ipu_adl.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              100 drivers/clk/mediatek/clk-mt8183-ipu_conn.c 	struct device_node *node = pdev->dev.of_node;
node              104 drivers/clk/mediatek/clk-mt8183-ipu_conn.c 	mtk_clk_register_gates(node, ipu_conn_clks, ARRAY_SIZE(ipu_conn_clks),
node              107 drivers/clk/mediatek/clk-mt8183-ipu_conn.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               32 drivers/clk/mediatek/clk-mt8183-mfgcfg.c 	struct device_node *node = pdev->dev.of_node;
node               38 drivers/clk/mediatek/clk-mt8183-mfgcfg.c 	mtk_clk_register_gates_with_dev(node, mfg_clks, ARRAY_SIZE(mfg_clks),
node               41 drivers/clk/mediatek/clk-mt8183-mfgcfg.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               88 drivers/clk/mediatek/clk-mt8183-mm.c 	struct device_node *node = pdev->dev.of_node;
node               92 drivers/clk/mediatek/clk-mt8183-mm.c 	mtk_clk_register_gates(node, mm_clks, ARRAY_SIZE(mm_clks),
node               95 drivers/clk/mediatek/clk-mt8183-mm.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               44 drivers/clk/mediatek/clk-mt8183-vdec.c 	struct device_node *node = pdev->dev.of_node;
node               48 drivers/clk/mediatek/clk-mt8183-vdec.c 	mtk_clk_register_gates(node, vdec_clks, ARRAY_SIZE(vdec_clks),
node               51 drivers/clk/mediatek/clk-mt8183-vdec.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               36 drivers/clk/mediatek/clk-mt8183-venc.c 	struct device_node *node = pdev->dev.of_node;
node               40 drivers/clk/mediatek/clk-mt8183-venc.c 	mtk_clk_register_gates(node, venc_clks, ARRAY_SIZE(venc_clks),
node               43 drivers/clk/mediatek/clk-mt8183-venc.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1158 drivers/clk/mediatek/clk-mt8183.c 	struct device_node *node = pdev->dev.of_node;
node             1162 drivers/clk/mediatek/clk-mt8183.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node             1164 drivers/clk/mediatek/clk-mt8183.c 	mtk_clk_register_gates(node, apmixed_clks, ARRAY_SIZE(apmixed_clks),
node             1167 drivers/clk/mediatek/clk-mt8183.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1172 drivers/clk/mediatek/clk-mt8183.c static void clk_mt8183_top_init_early(struct device_node *node)
node             1184 drivers/clk/mediatek/clk-mt8183.c 	of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data);
node             1194 drivers/clk/mediatek/clk-mt8183.c 	struct device_node *node = pdev->dev.of_node;
node             1209 drivers/clk/mediatek/clk-mt8183.c 		node, &mt8183_clk_lock, top_clk_data);
node             1217 drivers/clk/mediatek/clk-mt8183.c 	mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks),
node             1220 drivers/clk/mediatek/clk-mt8183.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, top_clk_data);
node             1226 drivers/clk/mediatek/clk-mt8183.c 	struct device_node *node = pdev->dev.of_node;
node             1231 drivers/clk/mediatek/clk-mt8183.c 	mtk_clk_register_gates(node, infra_clks, ARRAY_SIZE(infra_clks),
node             1234 drivers/clk/mediatek/clk-mt8183.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1242 drivers/clk/mediatek/clk-mt8183.c 	mtk_register_reset_controller_set_clr(node, 4, INFRA_RST0_SET_OFFSET);
node             1250 drivers/clk/mediatek/clk-mt8183.c 	struct device_node *node = pdev->dev.of_node;
node             1254 drivers/clk/mediatek/clk-mt8183.c 	mtk_clk_register_gates(node, peri_clks, ARRAY_SIZE(peri_clks),
node             1257 drivers/clk/mediatek/clk-mt8183.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node             1263 drivers/clk/mediatek/clk-mt8183.c 	struct device_node *node = pdev->dev.of_node;
node             1276 drivers/clk/mediatek/clk-mt8183.c 	return of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               50 drivers/clk/mediatek/clk-mt8516-aud.c static void __init mtk_audsys_init(struct device_node *node)
node               57 drivers/clk/mediatek/clk-mt8516-aud.c 	mtk_clk_register_gates(node, aud_clks, ARRAY_SIZE(aud_clks), clk_data);
node               59 drivers/clk/mediatek/clk-mt8516-aud.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              677 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_topckgen_init(struct device_node *node)
node              683 drivers/clk/mediatek/clk-mt8516.c 	base = of_iomap(node, 0);
node              693 drivers/clk/mediatek/clk-mt8516.c 	mtk_clk_register_gates(node, top_clks, ARRAY_SIZE(top_clks), clk_data);
node              701 drivers/clk/mediatek/clk-mt8516.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              708 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_infracfg_init(struct device_node *node)
node              714 drivers/clk/mediatek/clk-mt8516.c 	base = of_iomap(node, 0);
node              725 drivers/clk/mediatek/clk-mt8516.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              787 drivers/clk/mediatek/clk-mt8516.c static void __init mtk_apmixedsys_init(struct device_node *node)
node              793 drivers/clk/mediatek/clk-mt8516.c 	base = of_iomap(node, 0);
node              801 drivers/clk/mediatek/clk-mt8516.c 	mtk_clk_register_plls(node, plls, ARRAY_SIZE(plls), clk_data);
node              803 drivers/clk/mediatek/clk-mt8516.c 	r = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               97 drivers/clk/mediatek/clk-mtk.c int mtk_clk_register_gates_with_dev(struct device_node *node,
node              109 drivers/clk/mediatek/clk-mtk.c 	regmap = syscon_node_to_regmap(node);
node              111 drivers/clk/mediatek/clk-mtk.c 		pr_err("Cannot find regmap for %pOF: %ld\n", node,
node              141 drivers/clk/mediatek/clk-mtk.c int mtk_clk_register_gates(struct device_node *node,
node              145 drivers/clk/mediatek/clk-mtk.c 	return mtk_clk_register_gates_with_dev(node,
node              168 drivers/clk/mediatek/clk-mtk.h int mtk_clk_register_gates(struct device_node *node,
node              172 drivers/clk/mediatek/clk-mtk.h int mtk_clk_register_gates_with_dev(struct device_node *node,
node              238 drivers/clk/mediatek/clk-mtk.h void mtk_clk_register_plls(struct device_node *node,
node              191 drivers/clk/mediatek/clk-mux.c 			   int num, struct device_node *node,
node              199 drivers/clk/mediatek/clk-mux.c 	regmap = syscon_node_to_regmap(node);
node              201 drivers/clk/mediatek/clk-mux.c 		pr_err("Cannot find regmap for %pOF: %ld\n", node,
node               85 drivers/clk/mediatek/clk-mux.h 			   int num, struct device_node *node,
node              344 drivers/clk/mediatek/clk-pll.c void mtk_clk_register_plls(struct device_node *node,
node              351 drivers/clk/mediatek/clk-pll.c 	base = of_iomap(node, 0);
node              251 drivers/clk/mvebu/clk-corediv.c mvebu_corediv_clk_init(struct device_node *node,
node              262 drivers/clk/mvebu/clk-corediv.c 	base = of_iomap(node, 0);
node              266 drivers/clk/mvebu/clk-corediv.c 	parent_name = of_clk_get_parent_name(node, 0);
node              284 drivers/clk/mvebu/clk-corediv.c 		of_property_read_string_index(node, "clock-output-names",
node              302 drivers/clk/mvebu/clk-corediv.c 	of_clk_add_provider(node, of_clk_src_onecell_get, &clk_data);
node              311 drivers/clk/mvebu/clk-corediv.c static void __init armada370_corediv_clk_init(struct device_node *node)
node              313 drivers/clk/mvebu/clk-corediv.c 	return mvebu_corediv_clk_init(node, &armada370_corediv_soc);
node              318 drivers/clk/mvebu/clk-corediv.c static void __init armada375_corediv_clk_init(struct device_node *node)
node              320 drivers/clk/mvebu/clk-corediv.c 	return mvebu_corediv_clk_init(node, &armada375_corediv_soc);
node              325 drivers/clk/mvebu/clk-corediv.c static void __init armada380_corediv_clk_init(struct device_node *node)
node              327 drivers/clk/mvebu/clk-corediv.c 	return mvebu_corediv_clk_init(node, &armada380_corediv_soc);
node              332 drivers/clk/mvebu/clk-corediv.c static void __init mv98dx3236_corediv_clk_init(struct device_node *node)
node              334 drivers/clk/mvebu/clk-corediv.c 	return mvebu_corediv_clk_init(node, &mv98dx3236_corediv_soc);
node              166 drivers/clk/mvebu/clk-cpu.c static void __init of_cpu_clk_setup(struct device_node *node)
node              169 drivers/clk/mvebu/clk-cpu.c 	void __iomem *clock_complex_base = of_iomap(node, 0);
node              170 drivers/clk/mvebu/clk-cpu.c 	void __iomem *pmu_dfs_base = of_iomap(node, 1);
node              210 drivers/clk/mvebu/clk-cpu.c 		cpuclk[cpu].parent_name = of_clk_get_parent_name(node, 0);
node              231 drivers/clk/mvebu/clk-cpu.c 	of_clk_add_provider(node, of_clk_src_onecell_get, &clk_data);
node              247 drivers/clk/mvebu/clk-cpu.c static void __init of_mv98dx3236_cpu_clk_setup(struct device_node *node)
node              249 drivers/clk/mvebu/clk-cpu.c 	of_clk_add_provider(node, of_clk_src_simple_get, NULL);
node               16 drivers/clk/pistachio/clk.c pistachio_clk_alloc_provider(struct device_node *node, unsigned int num_clks)
node               28 drivers/clk/pistachio/clk.c 	p->node = node;
node               29 drivers/clk/pistachio/clk.c 	p->base = of_iomap(node, 0);
node               54 drivers/clk/pistachio/clk.c 	of_clk_add_provider(p->node, of_clk_src_onecell_get, &p->clk_data);
node              142 drivers/clk/pistachio/clk.h 	struct device_node *node;
node              148 drivers/clk/pistachio/clk.h pistachio_clk_alloc_provider(struct device_node *node, unsigned int num_clks);
node              123 drivers/clk/qcom/common.c 	struct device_node *node = NULL;
node              132 drivers/clk/qcom/common.c 		node = of_get_child_by_name(clocks_node, path);
node              136 drivers/clk/qcom/common.c 	if (!node) {
node              151 drivers/clk/qcom/common.c 	of_node_put(node);
node               54 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos_clkout_init(struct device_node *node, u32 mux_mask)
node               74 drivers/clk/samsung/clk-exynos-clkout.c 		parents[i] = of_clk_get_by_name(node, name);
node               87 drivers/clk/samsung/clk-exynos-clkout.c 	clkout->reg = of_iomap(node, 0);
node              110 drivers/clk/samsung/clk-exynos-clkout.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get, &clkout->data);
node              138 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos4_clkout_init(struct device_node *node)
node              140 drivers/clk/samsung/clk-exynos-clkout.c 	exynos_clkout_init(node, EXYNOS4_CLKOUT_MUX_MASK);
node              149 drivers/clk/samsung/clk-exynos-clkout.c static void __init exynos5_clkout_init(struct device_node *node)
node              151 drivers/clk/samsung/clk-exynos-clkout.c 	exynos_clkout_init(node, EXYNOS5_CLKOUT_MUX_MASK);
node              291 drivers/clk/samsung/clk.c 	list_for_each_entry(reg_cache, &clock_reg_cache_list, node) {
node              304 drivers/clk/samsung/clk.c 	list_for_each_entry(reg_cache, &clock_reg_cache_list, node)
node              338 drivers/clk/samsung/clk.c 	list_add_tail(&reg_cache->node, &clock_reg_cache_list);
node              275 drivers/clk/samsung/clk.h 	struct list_head node;
node               94 drivers/clk/socfpga/clk-gate-a10.c static void __init __socfpga_gate_init(struct device_node *node,
node              103 drivers/clk/socfpga/clk-gate-a10.c 	const char *clk_name = node->name;
node              112 drivers/clk/socfpga/clk-gate-a10.c 	rc = of_property_read_u32_array(node, "clk-gate", clk_gate, 2);
node              124 drivers/clk/socfpga/clk-gate-a10.c 	rc = of_property_read_u32(node, "fixed-divider", &fixed_div);
node              130 drivers/clk/socfpga/clk-gate-a10.c 	rc = of_property_read_u32_array(node, "div-reg", div_reg, 3);
node              139 drivers/clk/socfpga/clk-gate-a10.c 	rc = of_property_read_u32_array(node, "clk-phase", clk_phase, 2);
node              153 drivers/clk/socfpga/clk-gate-a10.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              159 drivers/clk/socfpga/clk-gate-a10.c 	init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS);
node              168 drivers/clk/socfpga/clk-gate-a10.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              173 drivers/clk/socfpga/clk-gate-a10.c void __init socfpga_a10_gate_init(struct device_node *node)
node              175 drivers/clk/socfpga/clk-gate-a10.c 	__socfpga_gate_init(node, &gateclk_ops);
node              171 drivers/clk/socfpga/clk-gate.c void __init socfpga_gate_init(struct device_node *node)
node              179 drivers/clk/socfpga/clk-gate.c 	const char *clk_name = node->name;
node              193 drivers/clk/socfpga/clk-gate.c 	rc = of_property_read_u32_array(node, "clk-gate", clk_gate, 2);
node              205 drivers/clk/socfpga/clk-gate.c 	rc = of_property_read_u32(node, "fixed-divider", &fixed_div);
node              211 drivers/clk/socfpga/clk-gate.c 	rc = of_property_read_u32_array(node, "div-reg", div_reg, 3);
node              220 drivers/clk/socfpga/clk-gate.c 	rc = of_property_read_u32_array(node, "clk-phase", clk_phase, 2);
node              226 drivers/clk/socfpga/clk-gate.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              232 drivers/clk/socfpga/clk-gate.c 	init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS);
node              246 drivers/clk/socfpga/clk-gate.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               60 drivers/clk/socfpga/clk-periph-a10.c static __init void __socfpga_periph_init(struct device_node *node,
node               66 drivers/clk/socfpga/clk-periph-a10.c 	const char *clk_name = node->name;
node               73 drivers/clk/socfpga/clk-periph-a10.c 	of_property_read_u32(node, "reg", &reg);
node               81 drivers/clk/socfpga/clk-periph-a10.c 	rc = of_property_read_u32_array(node, "div-reg", div_reg, 3);
node               90 drivers/clk/socfpga/clk-periph-a10.c 	rc = of_property_read_u32(node, "fixed-divider", &fixed_div);
node               96 drivers/clk/socfpga/clk-periph-a10.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              102 drivers/clk/socfpga/clk-periph-a10.c 	init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS);
node              112 drivers/clk/socfpga/clk-periph-a10.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              125 drivers/clk/socfpga/clk-periph-a10.c void __init socfpga_a10_periph_init(struct device_node *node)
node              127 drivers/clk/socfpga/clk-periph-a10.c 	__socfpga_periph_init(node, &periclk_ops);
node               50 drivers/clk/socfpga/clk-periph.c static __init void __socfpga_periph_init(struct device_node *node,
node               56 drivers/clk/socfpga/clk-periph.c 	const char *clk_name = node->name;
node               63 drivers/clk/socfpga/clk-periph.c 	of_property_read_u32(node, "reg", &reg);
node               71 drivers/clk/socfpga/clk-periph.c 	rc = of_property_read_u32_array(node, "div-reg", div_reg, 3);
node               80 drivers/clk/socfpga/clk-periph.c 	rc = of_property_read_u32(node, "fixed-divider", &fixed_div);
node               86 drivers/clk/socfpga/clk-periph.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               92 drivers/clk/socfpga/clk-periph.c 	init.num_parents = of_clk_parent_fill(node, parent_name,
node              103 drivers/clk/socfpga/clk-periph.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              106 drivers/clk/socfpga/clk-periph.c void __init socfpga_periph_init(struct device_node *node)
node              108 drivers/clk/socfpga/clk-periph.c 	__socfpga_periph_init(node, &periclk_ops);
node               66 drivers/clk/socfpga/clk-pll-a10.c static struct clk * __init __socfpga_pll_init(struct device_node *node,
node               72 drivers/clk/socfpga/clk-pll-a10.c 	const char *clk_name = node->name;
node               79 drivers/clk/socfpga/clk-pll-a10.c 	of_property_read_u32(node, "reg", &reg);
node               91 drivers/clk/socfpga/clk-pll-a10.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               98 drivers/clk/socfpga/clk-pll-a10.c 			of_clk_get_parent_name(node, i)) != NULL)
node              113 drivers/clk/socfpga/clk-pll-a10.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              117 drivers/clk/socfpga/clk-pll-a10.c void __init socfpga_a10_pll_init(struct device_node *node)
node              119 drivers/clk/socfpga/clk-pll-a10.c 	__socfpga_pll_init(node, &clk_pll_ops);
node               73 drivers/clk/socfpga/clk-pll.c static __init struct clk *__socfpga_pll_init(struct device_node *node,
node               79 drivers/clk/socfpga/clk-pll.c 	const char *clk_name = node->name;
node               85 drivers/clk/socfpga/clk-pll.c 	of_property_read_u32(node, "reg", &reg);
node               97 drivers/clk/socfpga/clk-pll.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              103 drivers/clk/socfpga/clk-pll.c 	init.num_parents = of_clk_parent_fill(node, parent_name, SOCFPGA_MAX_PARENTS);
node              116 drivers/clk/socfpga/clk-pll.c 	rc = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              120 drivers/clk/socfpga/clk-pll.c void __init socfpga_pll_init(struct device_node *node)
node              122 drivers/clk/socfpga/clk-pll.c 	__socfpga_pll_init(node, &clk_pll_ops);
node               32 drivers/clk/socfpga/clk.h void __init socfpga_pll_init(struct device_node *node);
node               33 drivers/clk/socfpga/clk.h void __init socfpga_periph_init(struct device_node *node);
node               34 drivers/clk/socfpga/clk.h void __init socfpga_gate_init(struct device_node *node);
node               35 drivers/clk/socfpga/clk.h void socfpga_a10_pll_init(struct device_node *node);
node               36 drivers/clk/socfpga/clk.h void socfpga_a10_periph_init(struct device_node *node);
node               37 drivers/clk/socfpga/clk.h void socfpga_a10_gate_init(struct device_node *node);
node               43 drivers/clk/sprd/common.c 	struct device_node *node = pdev->dev.of_node;
node               47 drivers/clk/sprd/common.c 	if (of_find_property(node, "sprd,syscon", NULL)) {
node               48 drivers/clk/sprd/common.c 		regmap = syscon_regmap_lookup_by_phandle(node, "sprd,syscon");
node             1428 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun4i_ccu_init(struct device_node *node,
node             1434 drivers/clk/sunxi-ng/ccu-sun4i-a10.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node             1437 drivers/clk/sunxi-ng/ccu-sun4i-a10.c 		       of_node_full_name(node));
node             1467 drivers/clk/sunxi-ng/ccu-sun4i-a10.c 	sunxi_ccu_probe(node, reg, desc);
node             1470 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun4i_a10_ccu_setup(struct device_node *node)
node             1472 drivers/clk/sunxi-ng/ccu-sun4i-a10.c 	sun4i_ccu_init(node, &sun4i_a10_ccu_desc);
node             1477 drivers/clk/sunxi-ng/ccu-sun4i-a10.c static void __init sun7i_a20_ccu_setup(struct device_node *node)
node             1479 drivers/clk/sunxi-ng/ccu-sun4i-a10.c 	sun4i_ccu_init(node, &sun7i_a20_ccu_desc);
node              188 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c static void __init sunxi_r_ccu_init(struct device_node *node,
node              193 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              195 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              199 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c 	sunxi_ccu_probe(node, reg, desc);
node              202 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c static void __init sun50i_h6_r_ccu_setup(struct device_node *node)
node              204 drivers/clk/sunxi-ng/ccu-sun50i-h6-r.c 	sunxi_r_ccu_init(node, &sun50i_h6_r_ccu_desc);
node              987 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_ccu_init(struct device_node *node,
node              993 drivers/clk/sunxi-ng/ccu-sun5i.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              995 drivers/clk/sunxi-ng/ccu-sun5i.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node             1015 drivers/clk/sunxi-ng/ccu-sun5i.c 	sunxi_ccu_probe(node, reg, desc);
node             1018 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_a10s_ccu_setup(struct device_node *node)
node             1020 drivers/clk/sunxi-ng/ccu-sun5i.c 	sun5i_ccu_init(node, &sun5i_a10s_ccu_desc);
node             1025 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_a13_ccu_setup(struct device_node *node)
node             1027 drivers/clk/sunxi-ng/ccu-sun5i.c 	sun5i_ccu_init(node, &sun5i_a13_ccu_desc);
node             1032 drivers/clk/sunxi-ng/ccu-sun5i.c static void __init sun5i_gr8_ccu_setup(struct device_node *node)
node             1034 drivers/clk/sunxi-ng/ccu-sun5i.c 	sun5i_ccu_init(node, &sun5i_gr8_ccu_desc);
node             1229 drivers/clk/sunxi-ng/ccu-sun6i-a31.c static void __init sun6i_a31_ccu_setup(struct device_node *node)
node             1234 drivers/clk/sunxi-ng/ccu-sun6i-a31.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node             1236 drivers/clk/sunxi-ng/ccu-sun6i-a31.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node             1260 drivers/clk/sunxi-ng/ccu-sun6i-a31.c 	sunxi_ccu_probe(node, reg, &sun6i_a31_ccu_desc);
node              727 drivers/clk/sunxi-ng/ccu-sun8i-a23.c static void __init sun8i_a23_ccu_setup(struct device_node *node)
node              732 drivers/clk/sunxi-ng/ccu-sun8i-a23.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              734 drivers/clk/sunxi-ng/ccu-sun8i-a23.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              748 drivers/clk/sunxi-ng/ccu-sun8i-a23.c 	sunxi_ccu_probe(node, reg, &sun8i_a23_ccu_desc);
node              787 drivers/clk/sunxi-ng/ccu-sun8i-a33.c static void __init sun8i_a33_ccu_setup(struct device_node *node)
node              792 drivers/clk/sunxi-ng/ccu-sun8i-a33.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              794 drivers/clk/sunxi-ng/ccu-sun8i-a33.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              808 drivers/clk/sunxi-ng/ccu-sun8i-a33.c 	sunxi_ccu_probe(node, reg, &sun8i_a33_ccu_desc);
node             1139 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sunxi_h3_h5_ccu_init(struct device_node *node,
node             1145 drivers/clk/sunxi-ng/ccu-sun8i-h3.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node             1147 drivers/clk/sunxi-ng/ccu-sun8i-h3.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node             1156 drivers/clk/sunxi-ng/ccu-sun8i-h3.c 	sunxi_ccu_probe(node, reg, desc);
node             1166 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sun8i_h3_ccu_setup(struct device_node *node)
node             1168 drivers/clk/sunxi-ng/ccu-sun8i-h3.c 	sunxi_h3_h5_ccu_init(node, &sun8i_h3_ccu_desc);
node             1173 drivers/clk/sunxi-ng/ccu-sun8i-h3.c static void __init sun50i_h5_ccu_setup(struct device_node *node)
node             1175 drivers/clk/sunxi-ng/ccu-sun8i-h3.c 	sunxi_h3_h5_ccu_init(node, &sun50i_h5_ccu_desc);
node              257 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sunxi_r_ccu_init(struct device_node *node,
node              262 drivers/clk/sunxi-ng/ccu-sun8i-r.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              264 drivers/clk/sunxi-ng/ccu-sun8i-r.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              268 drivers/clk/sunxi-ng/ccu-sun8i-r.c 	sunxi_ccu_probe(node, reg, desc);
node              271 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun8i_a83t_r_ccu_setup(struct device_node *node)
node              273 drivers/clk/sunxi-ng/ccu-sun8i-r.c 	sunxi_r_ccu_init(node, &sun8i_a83t_r_ccu_desc);
node              278 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun8i_h3_r_ccu_setup(struct device_node *node)
node              280 drivers/clk/sunxi-ng/ccu-sun8i-r.c 	sunxi_r_ccu_init(node, &sun8i_h3_r_ccu_desc);
node              285 drivers/clk/sunxi-ng/ccu-sun8i-r.c static void __init sun50i_a64_r_ccu_setup(struct device_node *node)
node              287 drivers/clk/sunxi-ng/ccu-sun8i-r.c 	sunxi_r_ccu_init(node, &sun50i_a64_r_ccu_desc);
node              797 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3_v3s_ccu_init(struct device_node *node,
node              803 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              805 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              814 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c 	sunxi_ccu_probe(node, reg, ccu_desc);
node              817 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3s_ccu_setup(struct device_node *node)
node              819 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c 	sun8i_v3_v3s_ccu_init(node, &sun8i_v3s_ccu_desc);
node              822 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c static void __init sun8i_v3_ccu_setup(struct device_node *node)
node              824 drivers/clk/sunxi-ng/ccu-sun8i-v3s.c 	sun8i_v3_v3s_ccu_init(node, &sun8i_v3_ccu_desc);
node              525 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c static void __init suniv_f1c100s_ccu_setup(struct device_node *node)
node              530 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              532 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c 		pr_err("%pOF: Could not map the clock registers\n", node);
node              541 drivers/clk/sunxi-ng/ccu-suniv-f1c100s.c 	sunxi_ccu_probe(node, reg, &suniv_ccu_desc);
node               82 drivers/clk/sunxi-ng/ccu_common.c int sunxi_ccu_probe(struct device_node *node, void __iomem *reg,
node              106 drivers/clk/sunxi-ng/ccu_common.c 		ret = of_clk_hw_register(node, hw);
node              113 drivers/clk/sunxi-ng/ccu_common.c 	ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get,
node              124 drivers/clk/sunxi-ng/ccu_common.c 	reset->rcdev.of_node = node;
node              141 drivers/clk/sunxi-ng/ccu_common.c 	of_clk_del_provider(node);
node               66 drivers/clk/sunxi-ng/ccu_common.h int sunxi_ccu_probe(struct device_node *node, void __iomem *reg,
node               14 drivers/clk/sunxi/clk-a10-codec.c static void __init sun4i_codec_clk_setup(struct device_node *node)
node               17 drivers/clk/sunxi/clk-a10-codec.c 	const char *clk_name = node->name, *parent_name;
node               20 drivers/clk/sunxi/clk-a10-codec.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               24 drivers/clk/sunxi/clk-a10-codec.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               25 drivers/clk/sunxi/clk-a10-codec.c 	parent_name = of_clk_get_parent_name(node, 0);
node               32 drivers/clk/sunxi/clk-a10-codec.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               17 drivers/clk/sunxi/clk-a10-hosc.c static void __init sun4i_osc_clk_setup(struct device_node *node)
node               22 drivers/clk/sunxi/clk-a10-hosc.c 	const char *clk_name = node->name;
node               25 drivers/clk/sunxi/clk-a10-hosc.c 	if (of_property_read_u32(node, "clock-frequency", &rate))
node               36 drivers/clk/sunxi/clk-a10-hosc.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               39 drivers/clk/sunxi/clk-a10-hosc.c 	gate->reg = of_iomap(node, 0);
node               53 drivers/clk/sunxi/clk-a10-hosc.c 	of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               21 drivers/clk/sunxi/clk-a10-mod1.c static void __init sun4i_mod1_clk_setup(struct device_node *node)
node               27 drivers/clk/sunxi/clk-a10-mod1.c 	const char *clk_name = node->name;
node               31 drivers/clk/sunxi/clk-a10-mod1.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               43 drivers/clk/sunxi/clk-a10-mod1.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               44 drivers/clk/sunxi/clk-a10-mod1.c 	i = of_clk_parent_fill(node, parents, SUN4I_MOD1_MAX_PARENTS);
node               61 drivers/clk/sunxi/clk-a10-mod1.c 	of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               38 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun4i_pll2_setup(struct device_node *node,
node               41 drivers/clk/sunxi/clk-a10-pll2.c 	const char *clk_name = node->name, *parent;
node               49 drivers/clk/sunxi/clk-a10-pll2.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               61 drivers/clk/sunxi/clk-a10-pll2.c 	parent = of_clk_get_parent_name(node, 0);
node              120 drivers/clk/sunxi/clk-a10-pll2.c 	of_property_read_string_index(node, "clock-output-names",
node              135 drivers/clk/sunxi/clk-a10-pll2.c 	of_property_read_string_index(node, "clock-output-names",
node              144 drivers/clk/sunxi/clk-a10-pll2.c 	of_property_read_string_index(node, "clock-output-names",
node              153 drivers/clk/sunxi/clk-a10-pll2.c 	of_property_read_string_index(node, "clock-output-names",
node              163 drivers/clk/sunxi/clk-a10-pll2.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              181 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun4i_a10_pll2_setup(struct device_node *node)
node              183 drivers/clk/sunxi/clk-a10-pll2.c 	sun4i_pll2_setup(node, 0);
node              189 drivers/clk/sunxi/clk-a10-pll2.c static void __init sun5i_a13_pll2_setup(struct device_node *node)
node              191 drivers/clk/sunxi/clk-a10-pll2.c 	sun4i_pll2_setup(node, 1);
node               85 drivers/clk/sunxi/clk-a10-ve.c static void __init sun4i_ve_clk_setup(struct device_node *node)
node               92 drivers/clk/sunxi/clk-a10-ve.c 	const char *clk_name = node->name;
node               96 drivers/clk/sunxi/clk-a10-ve.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              108 drivers/clk/sunxi/clk-a10-ve.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              109 drivers/clk/sunxi/clk-a10-ve.c 	parent = of_clk_get_parent_name(node, 0);
node              128 drivers/clk/sunxi/clk-a10-ve.c 	err = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              140 drivers/clk/sunxi/clk-a10-ve.c 	reset_data->rcdev.of_node = node;
node              152 drivers/clk/sunxi/clk-a10-ve.c 	of_clk_del_provider(node);
node               52 drivers/clk/sunxi/clk-a20-gmac.c static void __init sun7i_a20_gmac_clk_setup(struct device_node *node)
node               57 drivers/clk/sunxi/clk-a20-gmac.c 	const char *clk_name = node->name;
node               61 drivers/clk/sunxi/clk-a20-gmac.c 	if (of_property_read_string(node, "clock-output-names", &clk_name))
node               74 drivers/clk/sunxi/clk-a20-gmac.c 	if (of_clk_parent_fill(node, parents, 2) != 2)
node               77 drivers/clk/sunxi/clk-a20-gmac.c 	reg = of_iomap(node, 0);
node              100 drivers/clk/sunxi/clk-a20-gmac.c 	of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              176 drivers/clk/sunxi/clk-factors.c static struct clk *__sunxi_factors_register(struct device_node *node,
node              187 drivers/clk/sunxi/clk-factors.c 	const char *clk_name = node->name;
node              192 drivers/clk/sunxi/clk-factors.c 	i = of_clk_parent_fill(node, parents, FACTORS_MAX_PARENTS);
node              201 drivers/clk/sunxi/clk-factors.c 		of_property_read_string(node, "clock-output-names", &clk_name);
node              253 drivers/clk/sunxi/clk-factors.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              272 drivers/clk/sunxi/clk-factors.c struct clk *sunxi_factors_register(struct device_node *node,
node              277 drivers/clk/sunxi/clk-factors.c 	return __sunxi_factors_register(node, data, lock, reg, 0);
node              280 drivers/clk/sunxi/clk-factors.c struct clk *sunxi_factors_register_critical(struct device_node *node,
node              285 drivers/clk/sunxi/clk-factors.c 	return __sunxi_factors_register(node, data, lock, reg, CLK_IS_CRITICAL);
node              288 drivers/clk/sunxi/clk-factors.c void sunxi_factors_unregister(struct device_node *node, struct clk *clk)
node              298 drivers/clk/sunxi/clk-factors.c 	of_clk_del_provider(node);
node               54 drivers/clk/sunxi/clk-factors.h struct clk *sunxi_factors_register(struct device_node *node,
node               58 drivers/clk/sunxi/clk-factors.h struct clk *sunxi_factors_register_critical(struct device_node *node,
node               63 drivers/clk/sunxi/clk-factors.h void sunxi_factors_unregister(struct device_node *node, struct clk *clk);
node               68 drivers/clk/sunxi/clk-mod0.c static void __init sun4i_a10_mod0_setup(struct device_node *node)
node               72 drivers/clk/sunxi/clk-mod0.c 	reg = of_iomap(node, 0);
node               82 drivers/clk/sunxi/clk-mod0.c 	sunxi_factors_register(node, &sun4i_a10_mod0_data,
node              129 drivers/clk/sunxi/clk-mod0.c static void __init sun9i_a80_mod0_setup(struct device_node *node)
node              133 drivers/clk/sunxi/clk-mod0.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              136 drivers/clk/sunxi/clk-mod0.c 		       node);
node              140 drivers/clk/sunxi/clk-mod0.c 	sunxi_factors_register(node, &sun9i_a80_mod0_data,
node              147 drivers/clk/sunxi/clk-mod0.c static void __init sun5i_a13_mbus_setup(struct device_node *node)
node              151 drivers/clk/sunxi/clk-mod0.c 	reg = of_iomap(node, 0);
node              158 drivers/clk/sunxi/clk-mod0.c 	sunxi_factors_register_critical(node, &sun4i_a10_mod0_data,
node              290 drivers/clk/sunxi/clk-mod0.c static void __init sunxi_mmc_setup(struct device_node *node,
node              299 drivers/clk/sunxi/clk-mod0.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              301 drivers/clk/sunxi/clk-mod0.c 		pr_err("Couldn't map the %pOFn clock registers\n", node);
node              314 drivers/clk/sunxi/clk-mod0.c 	clk_data->clks[0] = sunxi_factors_register(node, data, lock, reg);
node              341 drivers/clk/sunxi/clk-mod0.c 		if (of_property_read_string_index(node, "clock-output-names",
node              343 drivers/clk/sunxi/clk-mod0.c 			init.name = node->name;
node              352 drivers/clk/sunxi/clk-mod0.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              364 drivers/clk/sunxi/clk-mod0.c static void __init sun4i_a10_mmc_setup(struct device_node *node)
node              366 drivers/clk/sunxi/clk-mod0.c 	sunxi_mmc_setup(node, &sun4i_a10_mod0_data, &sun4i_a10_mmc_lock);
node              372 drivers/clk/sunxi/clk-mod0.c static void __init sun9i_a80_mmc_setup(struct device_node *node)
node              374 drivers/clk/sunxi/clk-mod0.c 	sunxi_mmc_setup(node, &sun9i_a80_mod0_data, &sun9i_a80_mmc_lock);
node               18 drivers/clk/sunxi/clk-simple-gates.c static void __init sunxi_simple_gates_setup(struct device_node *node,
node               33 drivers/clk/sunxi/clk-simple-gates.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               37 drivers/clk/sunxi/clk-simple-gates.c 	clk_parent = of_clk_get_parent_name(node, 0);
node               43 drivers/clk/sunxi/clk-simple-gates.c 	number = of_property_count_u32_elems(node, "clock-indices");
node               44 drivers/clk/sunxi/clk-simple-gates.c 	of_property_read_u32_index(node, "clock-indices", number - 1, &number);
node               50 drivers/clk/sunxi/clk-simple-gates.c 	of_property_for_each_u32(node, "clock-indices", prop, p, index) {
node               51 drivers/clk/sunxi/clk-simple-gates.c 		of_property_read_string_index(node, "clock-output-names",
node               76 drivers/clk/sunxi/clk-simple-gates.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node               84 drivers/clk/sunxi/clk-simple-gates.c 	of_address_to_resource(node, 0, &res);
node               88 drivers/clk/sunxi/clk-simple-gates.c static void __init sunxi_simple_gates_init(struct device_node *node)
node               90 drivers/clk/sunxi/clk-simple-gates.c 	sunxi_simple_gates_setup(node, NULL, 0);
node              146 drivers/clk/sunxi/clk-simple-gates.c static void __init sun4i_a10_ahb_init(struct device_node *node)
node              148 drivers/clk/sunxi/clk-simple-gates.c 	sunxi_simple_gates_setup(node, sun4i_a10_ahb_critical_clocks,
node              164 drivers/clk/sunxi/clk-simple-gates.c static void __init sun4i_a10_dram_init(struct device_node *node)
node              166 drivers/clk/sunxi/clk-simple-gates.c 	sunxi_simple_gates_setup(node, sun4i_a10_dram_critical_clocks,
node              101 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_display_init(struct device_node *node,
node              105 drivers/clk/sunxi/clk-sun4i-display.c 	const char *clk_name = node->name;
node              115 drivers/clk/sunxi/clk-sun4i-display.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              117 drivers/clk/sunxi/clk-sun4i-display.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              123 drivers/clk/sunxi/clk-sun4i-display.c 	ret = of_clk_parent_fill(node, parents, data->parents);
node              169 drivers/clk/sunxi/clk-sun4i-display.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              187 drivers/clk/sunxi/clk-sun4i-display.c 	reset_data->rcdev.of_node = node;
node              207 drivers/clk/sunxi/clk-sun4i-display.c 	of_clk_del_provider(node);
node              218 drivers/clk/sunxi/clk-sun4i-display.c 	of_address_to_resource(node, 0, &res);
node              232 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_tcon_ch0_setup(struct device_node *node)
node              234 drivers/clk/sunxi/clk-sun4i-display.c 	sun4i_a10_display_init(node, &sun4i_a10_tcon_ch0_data);
node              251 drivers/clk/sunxi/clk-sun4i-display.c static void __init sun4i_a10_display_setup(struct device_node *node)
node              253 drivers/clk/sunxi/clk-sun4i-display.c 	sun4i_a10_display_init(node, &sun4i_a10_display_data);
node               21 drivers/clk/sunxi/clk-sun4i-pll3.c static void __init sun4i_a10_pll3_setup(struct device_node *node)
node               23 drivers/clk/sunxi/clk-sun4i-pll3.c 	const char *clk_name = node->name, *parent;
node               31 drivers/clk/sunxi/clk-sun4i-pll3.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               32 drivers/clk/sunxi/clk-sun4i-pll3.c 	parent = of_clk_get_parent_name(node, 0);
node               34 drivers/clk/sunxi/clk-sun4i-pll3.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               68 drivers/clk/sunxi/clk-sun4i-pll3.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               85 drivers/clk/sunxi/clk-sun4i-pll3.c 	of_address_to_resource(node, 0, &res);
node              224 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c static void __init tcon_ch1_setup(struct device_node *node)
node              227 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	const char *clk_name = node->name;
node              235 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              237 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              243 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	ret = of_clk_parent_fill(node, parents, TCON_CH1_SCLK2_PARENTS);
node              269 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              283 drivers/clk/sunxi/clk-sun4i-tcon-ch1.c 	of_address_to_resource(node, 0, &res);
node               22 drivers/clk/sunxi/clk-sun8i-apb0.c static struct clk *sun8i_a23_apb0_register(struct device_node *node,
node               25 drivers/clk/sunxi/clk-sun8i-apb0.c 	const char *clk_name = node->name;
node               30 drivers/clk/sunxi/clk-sun8i-apb0.c 	clk_parent = of_clk_get_parent_name(node, 0);
node               34 drivers/clk/sunxi/clk-sun8i-apb0.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               42 drivers/clk/sunxi/clk-sun8i-apb0.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               54 drivers/clk/sunxi/clk-sun8i-apb0.c static void sun8i_a23_apb0_setup(struct device_node *node)
node               60 drivers/clk/sunxi/clk-sun8i-apb0.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               73 drivers/clk/sunxi/clk-sun8i-apb0.c 	clk = sun8i_a23_apb0_register(node, reg);
node               81 drivers/clk/sunxi/clk-sun8i-apb0.c 	of_address_to_resource(node, 0, &res);
node               20 drivers/clk/sunxi/clk-sun8i-bus-gates.c static void __init sun8i_h3_bus_gates_init(struct device_node *node)
node               36 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               41 drivers/clk/sunxi/clk-sun8i-bus-gates.c 		int idx = of_property_match_string(node, "clock-names",
node               46 drivers/clk/sunxi/clk-sun8i-bus-gates.c 		parents[i] = of_clk_get_parent_name(node, idx);
node               53 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	number = of_property_count_u32_elems(node, "clock-indices");
node               54 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	of_property_read_u32_index(node, "clock-indices", number - 1, &number);
node               61 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	of_property_for_each_u32(node, "clock-indices", prop, p, index) {
node               62 drivers/clk/sunxi/clk-sun8i-bus-gates.c 		of_property_read_string_index(node, "clock-output-names",
node               94 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              102 drivers/clk/sunxi/clk-sun8i-bus-gates.c 	of_address_to_resource(node, 0, &res);
node               24 drivers/clk/sunxi/clk-sun8i-mbus.c static void __init sun8i_a23_mbus_setup(struct device_node *node)
node               26 drivers/clk/sunxi/clk-sun8i-mbus.c 	int num_parents = of_clk_get_parent_count(node);
node               28 drivers/clk/sunxi/clk-sun8i-mbus.c 	const char *clk_name = node->name;
node               41 drivers/clk/sunxi/clk-sun8i-mbus.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               59 drivers/clk/sunxi/clk-sun8i-mbus.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node               60 drivers/clk/sunxi/clk-sun8i-mbus.c 	of_clk_parent_fill(node, parents, num_parents);
node               85 drivers/clk/sunxi/clk-sun8i-mbus.c 	err = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              104 drivers/clk/sunxi/clk-sun8i-mbus.c 	of_address_to_resource(node, 0, &res);
node               76 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_pll4_setup(struct device_node *node)
node               80 drivers/clk/sunxi/clk-sun9i-core.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node               83 drivers/clk/sunxi/clk-sun9i-core.c 		       node);
node               87 drivers/clk/sunxi/clk-sun9i-core.c 	sunxi_factors_register(node, &sun9i_a80_pll4_data,
node              130 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_gt_setup(struct device_node *node)
node              134 drivers/clk/sunxi/clk-sun9i-core.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              137 drivers/clk/sunxi/clk-sun9i-core.c 		       node);
node              142 drivers/clk/sunxi/clk-sun9i-core.c 	sunxi_factors_register_critical(node, &sun9i_a80_gt_data,
node              185 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_ahb_setup(struct device_node *node)
node              189 drivers/clk/sunxi/clk-sun9i-core.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              192 drivers/clk/sunxi/clk-sun9i-core.c 		       node);
node              196 drivers/clk/sunxi/clk-sun9i-core.c 	sunxi_factors_register(node, &sun9i_a80_ahb_data,
node              211 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_apb0_setup(struct device_node *node)
node              215 drivers/clk/sunxi/clk-sun9i-core.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              218 drivers/clk/sunxi/clk-sun9i-core.c 		       node);
node              222 drivers/clk/sunxi/clk-sun9i-core.c 	sunxi_factors_register(node, &sun9i_a80_apb0_data,
node              268 drivers/clk/sunxi/clk-sun9i-core.c static void __init sun9i_a80_apb1_setup(struct device_node *node)
node              272 drivers/clk/sunxi/clk-sun9i-core.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              275 drivers/clk/sunxi/clk-sun9i-core.c 		       node);
node              279 drivers/clk/sunxi/clk-sun9i-core.c 	sunxi_factors_register(node, &sun9i_a80_apb1_data,
node              183 drivers/clk/sunxi/clk-sun9i-cpus.c static void sun9i_a80_cpus_setup(struct device_node *node)
node              185 drivers/clk/sunxi/clk-sun9i-cpus.c 	const char *clk_name = node->name;
node              197 drivers/clk/sunxi/clk-sun9i-cpus.c 	cpus->reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              201 drivers/clk/sunxi/clk-sun9i-cpus.c 	of_property_read_string(node, "clock-output-names", &clk_name);
node              204 drivers/clk/sunxi/clk-sun9i-cpus.c 	ret = of_clk_parent_fill(node, parents, SUN9I_CPUS_MAX_PARENTS);
node              224 drivers/clk/sunxi/clk-sun9i-cpus.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              236 drivers/clk/sunxi/clk-sun9i-cpus.c 	of_address_to_resource(node, 0, &res);
node              556 drivers/clk/sunxi/clk-sunxi.c static struct clk * __init sunxi_factors_clk_setup(struct device_node *node,
node              561 drivers/clk/sunxi/clk-sunxi.c 	reg = of_iomap(node, 0);
node              564 drivers/clk/sunxi/clk-sunxi.c 		       node);
node              568 drivers/clk/sunxi/clk-sunxi.c 	return sunxi_factors_register(node, data, &clk_lock, reg);
node              571 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll1_clk_setup(struct device_node *node)
node              573 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun4i_pll1_data);
node              578 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_pll1_clk_setup(struct device_node *node)
node              580 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun6i_a31_pll1_data);
node              585 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_pll1_clk_setup(struct device_node *node)
node              587 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun8i_a23_pll1_data);
node              592 drivers/clk/sunxi/clk-sunxi.c static void __init sun7i_pll4_clk_setup(struct device_node *node)
node              594 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun7i_a20_pll4_data);
node              599 drivers/clk/sunxi/clk-sunxi.c static void __init sun5i_ahb_clk_setup(struct device_node *node)
node              601 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun5i_a13_ahb_data);
node              606 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_ahb1_clk_setup(struct device_node *node)
node              608 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun6i_ahb1_data);
node              613 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_apb1_clk_setup(struct device_node *node)
node              615 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun4i_apb1_data);
node              620 drivers/clk/sunxi/clk-sunxi.c static void __init sun7i_out_clk_setup(struct device_node *node)
node              622 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun7i_a20_out_data);
node              650 drivers/clk/sunxi/clk-sunxi.c static struct clk * __init sunxi_mux_clk_setup(struct device_node *node,
node              655 drivers/clk/sunxi/clk-sunxi.c 	const char *clk_name = node->name;
node              660 drivers/clk/sunxi/clk-sunxi.c 	reg = of_iomap(node, 0);
node              662 drivers/clk/sunxi/clk-sunxi.c 		pr_err("Could not map registers for mux-clk: %pOF\n", node);
node              666 drivers/clk/sunxi/clk-sunxi.c 	i = of_clk_parent_fill(node, parents, SUNXI_MAX_PARENTS);
node              667 drivers/clk/sunxi/clk-sunxi.c 	if (of_property_read_string(node, "clock-output-names", &clk_name)) {
node              669 drivers/clk/sunxi/clk-sunxi.c 		       __func__, node);
node              684 drivers/clk/sunxi/clk-sunxi.c 	if (of_clk_add_provider(node, of_clk_src_simple_get, clk)) {
node              697 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_cpu_clk_setup(struct device_node *node)
node              700 drivers/clk/sunxi/clk-sunxi.c 	sunxi_mux_clk_setup(node, &sun4i_cpu_mux_data, CLK_IS_CRITICAL);
node              705 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_ahb1_mux_clk_setup(struct device_node *node)
node              707 drivers/clk/sunxi/clk-sunxi.c 	sunxi_mux_clk_setup(node, &sun6i_a31_ahb1_mux_data, 0);
node              712 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_ahb2_clk_setup(struct device_node *node)
node              714 drivers/clk/sunxi/clk-sunxi.c 	sunxi_mux_clk_setup(node, &sun8i_h3_ahb2_mux_data, 0);
node              775 drivers/clk/sunxi/clk-sunxi.c static void __init sunxi_divider_clk_setup(struct device_node *node,
node              779 drivers/clk/sunxi/clk-sunxi.c 	const char *clk_name = node->name;
node              783 drivers/clk/sunxi/clk-sunxi.c 	reg = of_iomap(node, 0);
node              785 drivers/clk/sunxi/clk-sunxi.c 		pr_err("Could not map registers for mux-clk: %pOF\n", node);
node              789 drivers/clk/sunxi/clk-sunxi.c 	clk_parent = of_clk_get_parent_name(node, 0);
node              791 drivers/clk/sunxi/clk-sunxi.c 	if (of_property_read_string(node, "clock-output-names", &clk_name)) {
node              793 drivers/clk/sunxi/clk-sunxi.c 		       __func__, node);
node              807 drivers/clk/sunxi/clk-sunxi.c 	if (of_clk_add_provider(node, of_clk_src_simple_get, clk)) {
node              814 drivers/clk/sunxi/clk-sunxi.c 		of_clk_del_provider(node);
node              826 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_ahb_clk_setup(struct device_node *node)
node              828 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divider_clk_setup(node, &sun4i_ahb_data);
node              833 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_apb0_clk_setup(struct device_node *node)
node              835 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divider_clk_setup(node, &sun4i_apb0_data);
node              840 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_axi_clk_setup(struct device_node *node)
node              842 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divider_clk_setup(node, &sun4i_axi_data);
node              847 drivers/clk/sunxi/clk-sunxi.c static void __init sun8i_axi_clk_setup(struct device_node *node)
node              849 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divider_clk_setup(node, &sun8i_a23_axi_data);
node              943 drivers/clk/sunxi/clk-sunxi.c static struct clk ** __init sunxi_divs_clk_setup(struct device_node *node,
node              968 drivers/clk/sunxi/clk-sunxi.c 			of_property_read_string_index(node, "clock-output-names",
node              977 drivers/clk/sunxi/clk-sunxi.c 		of_property_read_string_index(node, "clock-output-names",
node              992 drivers/clk/sunxi/clk-sunxi.c 	pclk = sunxi_factors_clk_setup(node, &factors);
node              999 drivers/clk/sunxi/clk-sunxi.c 	reg = of_iomap(node, 0);
node             1001 drivers/clk/sunxi/clk-sunxi.c 		pr_err("Could not map registers for divs-clk: %pOF\n", node);
node             1020 drivers/clk/sunxi/clk-sunxi.c 		if (of_property_read_string_index(node, "clock-output-names",
node             1092 drivers/clk/sunxi/clk-sunxi.c 	if (of_clk_add_provider(node, of_clk_src_onecell_get, clk_data)) {
node             1110 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll5_clk_setup(struct device_node *node)
node             1112 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divs_clk_setup(node, &pll5_divs_data);
node             1117 drivers/clk/sunxi/clk-sunxi.c static void __init sun4i_pll6_clk_setup(struct device_node *node)
node             1119 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divs_clk_setup(node, &pll6_divs_data);
node             1124 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_pll6_clk_setup(struct device_node *node)
node             1126 drivers/clk/sunxi/clk-sunxi.c 	sunxi_divs_clk_setup(node, &sun6i_a31_pll6_divs_data);
node             1162 drivers/clk/sunxi/clk-sunxi.c static void __init sun6i_display_setup(struct device_node *node)
node             1164 drivers/clk/sunxi/clk-sunxi.c 	sunxi_factors_clk_setup(node, &sun6i_display_data);
node               88 drivers/clk/sunxi/clk-usb.c static void __init sunxi_usb_clk_setup(struct device_node *node,
node              101 drivers/clk/sunxi/clk-usb.c 	reg = of_io_request_and_map(node, 0, of_node_full_name(node));
node              105 drivers/clk/sunxi/clk-usb.c 	clk_parent = of_clk_get_parent_name(node, 0);
node              125 drivers/clk/sunxi/clk-usb.c 		of_property_read_string_index(node, "clock-output-names",
node              138 drivers/clk/sunxi/clk-usb.c 	of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              149 drivers/clk/sunxi/clk-usb.c 		reset_data->clk = of_clk_get(node, 0);
node              161 drivers/clk/sunxi/clk-usb.c 	reset_data->rcdev.of_node = node;
node              172 drivers/clk/sunxi/clk-usb.c static void __init sun4i_a10_usb_setup(struct device_node *node)
node              174 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun4i_a10_usb_clk_data, &sun4i_a10_usb_lock);
node              183 drivers/clk/sunxi/clk-usb.c static void __init sun5i_a13_usb_setup(struct device_node *node)
node              185 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun5i_a13_usb_clk_data, &sun4i_a10_usb_lock);
node              194 drivers/clk/sunxi/clk-usb.c static void __init sun6i_a31_usb_setup(struct device_node *node)
node              196 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun6i_a31_usb_clk_data, &sun4i_a10_usb_lock);
node              205 drivers/clk/sunxi/clk-usb.c static void __init sun8i_a23_usb_setup(struct device_node *node)
node              207 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun8i_a23_usb_clk_data, &sun4i_a10_usb_lock);
node              217 drivers/clk/sunxi/clk-usb.c static void __init sun8i_h3_usb_setup(struct device_node *node)
node              219 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun8i_h3_usb_clk_data, &sun4i_a10_usb_lock);
node              231 drivers/clk/sunxi/clk-usb.c static void __init sun9i_a80_usb_mod_setup(struct device_node *node)
node              233 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun9i_a80_usb_mod_data, &a80_usb_mod_lock);
node              245 drivers/clk/sunxi/clk-usb.c static void __init sun9i_a80_usb_phy_setup(struct device_node *node)
node              247 drivers/clk/sunxi/clk-usb.c 	sunxi_usb_clk_setup(node, &sun9i_a80_usb_phy_data, &a80_usb_phy_lock);
node              378 drivers/clk/tegra/clk-emc.c 				   struct device_node *node)
node              383 drivers/clk/tegra/clk-emc.c 	err = of_property_read_u32(node, "clock-frequency", &tmp);
node              385 drivers/clk/tegra/clk-emc.c 		pr_err("timing %pOF: failed to read rate\n", node);
node              391 drivers/clk/tegra/clk-emc.c 	err = of_property_read_u32(node, "nvidia,parent-clock-frequency", &tmp);
node              393 drivers/clk/tegra/clk-emc.c 		pr_err("timing %pOF: failed to read parent rate\n", node);
node              399 drivers/clk/tegra/clk-emc.c 	timing->parent = of_clk_get_by_name(node, "emc-parent");
node              401 drivers/clk/tegra/clk-emc.c 		pr_err("timing %pOF: failed to get parent clock\n", node);
node              415 drivers/clk/tegra/clk-emc.c 		       node, __clk_get_name(timing->parent));
node              437 drivers/clk/tegra/clk-emc.c 				struct device_node *node,
node              442 drivers/clk/tegra/clk-emc.c 	int child_count = of_get_child_count(node);
node              455 drivers/clk/tegra/clk-emc.c 	for_each_child_of_node(node, child) {
node              485 drivers/clk/tegra/clk-emc.c 	struct device_node *node;
node              499 drivers/clk/tegra/clk-emc.c 	for_each_child_of_node(np, node) {
node              500 drivers/clk/tegra/clk-emc.c 		err = of_property_read_u32(node, "nvidia,ram-code",
node              509 drivers/clk/tegra/clk-emc.c 		err = load_timings_from_dt(tegra, node, node_ram_code);
node              511 drivers/clk/tegra/clk-emc.c 			of_node_put(node);
node             1324 drivers/clk/tegra/clk-tegra114.c 	struct device_node *node;
node             1332 drivers/clk/tegra/clk-tegra114.c 	node = of_find_matching_node(NULL, pmc_match);
node             1333 drivers/clk/tegra/clk-tegra114.c 	if (!node) {
node             1339 drivers/clk/tegra/clk-tegra114.c 	pmc_base = of_iomap(node, 0);
node             1415 drivers/clk/tegra/clk-tegra124.c 	struct device_node *node;
node             1424 drivers/clk/tegra/clk-tegra124.c 	node = of_find_matching_node(NULL, pmc_match);
node             1425 drivers/clk/tegra/clk-tegra124.c 	if (!node) {
node             1431 drivers/clk/tegra/clk-tegra124.c 	pmc_base = of_iomap(node, 0);
node             1139 drivers/clk/tegra/clk-tegra20.c 	struct device_node *node;
node             1147 drivers/clk/tegra/clk-tegra20.c 	node = of_find_matching_node(NULL, pmc_match);
node             1148 drivers/clk/tegra/clk-tegra20.c 	if (!node) {
node             1153 drivers/clk/tegra/clk-tegra20.c 	pmc_base = of_iomap(node, 0);
node             3509 drivers/clk/tegra/clk-tegra210.c 	struct device_node *node;
node             3518 drivers/clk/tegra/clk-tegra210.c 	node = of_find_matching_node(NULL, pmc_match);
node             3519 drivers/clk/tegra/clk-tegra210.c 	if (!node) {
node             3525 drivers/clk/tegra/clk-tegra210.c 	pmc_base = of_iomap(node, 0);
node             1307 drivers/clk/tegra/clk-tegra30.c 	struct device_node *node;
node             1315 drivers/clk/tegra/clk-tegra30.c 	node = of_find_matching_node(NULL, pmc_match);
node             1316 drivers/clk/tegra/clk-tegra30.c 	if (!node) {
node             1321 drivers/clk/tegra/clk-tegra30.c 	pmc_base = of_iomap(node, 0);
node              877 drivers/clk/ti/adpll.c 	struct device_node *node = pdev->dev.of_node;
node              895 drivers/clk/ti/adpll.c 	d->np = node;
node              137 drivers/clk/ti/apll.c 					  struct device_node *node)
node              145 drivers/clk/ti/apll.c 	clk = of_clk_get(node, 0);
node              148 drivers/clk/ti/apll.c 			 node);
node              149 drivers/clk/ti/apll.c 		if (!ti_clk_retry_init(node, hw, omap_clk_register_apll))
node              157 drivers/clk/ti/apll.c 	clk = of_clk_get(node, 1);
node              160 drivers/clk/ti/apll.c 			 node);
node              161 drivers/clk/ti/apll.c 		if (!ti_clk_retry_init(node, hw, omap_clk_register_apll))
node              169 drivers/clk/ti/apll.c 	clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name);
node              171 drivers/clk/ti/apll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              184 drivers/clk/ti/apll.c static void __init of_dra7_apll_setup(struct device_node *node)
node              201 drivers/clk/ti/apll.c 	init->name = node->name;
node              204 drivers/clk/ti/apll.c 	init->num_parents = of_clk_get_parent_count(node);
node              206 drivers/clk/ti/apll.c 		pr_err("dra7 apll %pOFn must have parent(s)\n", node);
node              214 drivers/clk/ti/apll.c 	of_clk_parent_fill(node, parent_names, init->num_parents);
node              218 drivers/clk/ti/apll.c 	ret = ti_clk_get_reg_addr(node, 0, &ad->control_reg);
node              219 drivers/clk/ti/apll.c 	ret |= ti_clk_get_reg_addr(node, 1, &ad->idlest_reg);
node              227 drivers/clk/ti/apll.c 	omap_clk_register_apll(&clk_hw->hw, node);
node              345 drivers/clk/ti/apll.c static void __init of_omap2_apll_setup(struct device_node *node)
node              365 drivers/clk/ti/apll.c 	init->name = node->name;
node              368 drivers/clk/ti/apll.c 	init->num_parents = of_clk_get_parent_count(node);
node              370 drivers/clk/ti/apll.c 		pr_err("%pOFn must have one parent\n", node);
node              374 drivers/clk/ti/apll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              377 drivers/clk/ti/apll.c 	if (of_property_read_u32(node, "ti,clock-frequency", &val)) {
node              378 drivers/clk/ti/apll.c 		pr_err("%pOFn missing clock-frequency\n", node);
node              383 drivers/clk/ti/apll.c 	if (of_property_read_u32(node, "ti,bit-shift", &val)) {
node              384 drivers/clk/ti/apll.c 		pr_err("%pOFn missing bit-shift\n", node);
node              392 drivers/clk/ti/apll.c 	if (of_property_read_u32(node, "ti,idlest-shift", &val)) {
node              393 drivers/clk/ti/apll.c 		pr_err("%pOFn missing idlest-shift\n", node);
node              399 drivers/clk/ti/apll.c 	ret = ti_clk_get_reg_addr(node, 0, &ad->control_reg);
node              400 drivers/clk/ti/apll.c 	ret |= ti_clk_get_reg_addr(node, 1, &ad->autoidle_reg);
node              401 drivers/clk/ti/apll.c 	ret |= ti_clk_get_reg_addr(node, 2, &ad->idlest_reg);
node              406 drivers/clk/ti/apll.c 	clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name);
node              408 drivers/clk/ti/apll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               32 drivers/clk/ti/autoidle.c 	struct list_head	node;
node              153 drivers/clk/ti/autoidle.c 	list_for_each_entry(c, &autoidle_clks, node)
node              167 drivers/clk/ti/autoidle.c 	list_for_each_entry(c, &autoidle_clks, node)
node              182 drivers/clk/ti/autoidle.c int __init of_ti_clk_autoidle_setup(struct device_node *node)
node              189 drivers/clk/ti/autoidle.c 	if (of_property_read_u32(node, "ti,autoidle-shift", &shift))
node              198 drivers/clk/ti/autoidle.c 	clk->name = node->name;
node              199 drivers/clk/ti/autoidle.c 	ret = ti_clk_get_reg_addr(node, 0, &clk->reg);
node              205 drivers/clk/ti/autoidle.c 	if (of_property_read_bool(node, "ti,invert-autoidle-bit"))
node              208 drivers/clk/ti/autoidle.c 	list_add(&clk->node, &autoidle_clks);
node              171 drivers/clk/ti/clk-dra7-atl.c static void __init of_dra7_atl_clock_setup(struct device_node *node)
node              186 drivers/clk/ti/clk-dra7-atl.c 	init.name = node->name;
node              189 drivers/clk/ti/clk-dra7-atl.c 	init.num_parents = of_clk_get_parent_count(node);
node              193 drivers/clk/ti/clk-dra7-atl.c 		       node);
node              202 drivers/clk/ti/clk-dra7-atl.c 	parent_names[0] = of_clk_get_parent_name(node, 0);
node              206 drivers/clk/ti/clk-dra7-atl.c 	clk = ti_clk_register(NULL, &clk_hw->hw, node->name);
node              209 drivers/clk/ti/clk-dra7-atl.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              221 drivers/clk/ti/clk-dra7-atl.c 	struct device_node *node = pdev->dev.of_node;
node              226 drivers/clk/ti/clk-dra7-atl.c 	if (!node)
node              233 drivers/clk/ti/clk-dra7-atl.c 	cinfo->iobase = of_iomap(node, 0);
node              249 drivers/clk/ti/clk-dra7-atl.c 		rc = of_parse_phandle_with_args(node, "ti,provided-clocks",
node              271 drivers/clk/ti/clk-dra7-atl.c 		cfg_node = of_get_child_by_name(node, prop);
node              134 drivers/clk/ti/clk.c 	struct device_node *node, *parent;
node              171 drivers/clk/ti/clk.c 		node = of_find_node_by_name(NULL, buf);
node              173 drivers/clk/ti/clk.c 			parent = node;
node              174 drivers/clk/ti/clk.c 			node = of_get_child_by_name(parent, "clk");
node              178 drivers/clk/ti/clk.c 		clkspec.np = node;
node              185 drivers/clk/ti/clk.c 				of_node_put(node);
node              190 drivers/clk/ti/clk.c 		of_node_put(node);
node              218 drivers/clk/ti/clk.c 	struct device_node *node;
node              235 drivers/clk/ti/clk.c int __init ti_clk_retry_init(struct device_node *node, void *user,
node              240 drivers/clk/ti/clk.c 	pr_debug("%pOFn: adding to retry list...\n", node);
node              245 drivers/clk/ti/clk.c 	retry->node = node;
node              263 drivers/clk/ti/clk.c int ti_clk_get_reg_addr(struct device_node *node, int index,
node              270 drivers/clk/ti/clk.c 		if (clocks_node_ptr[i] == node->parent)
node              275 drivers/clk/ti/clk.c 		pr_err("clk-provider not found for %pOFn!\n", node);
node              281 drivers/clk/ti/clk.c 	if (of_property_read_u32_index(node, "reg", index, &val)) {
node              282 drivers/clk/ti/clk.c 		pr_err("%pOFn must have reg[%d]!\n", node, index);
node              385 drivers/clk/ti/clk.c 			pr_debug("retry-init: %pOFn\n", retry->node);
node              386 drivers/clk/ti/clk.c 			retry->func(retry->user, retry->node);
node              553 drivers/clk/ti/clk.c 	list_add(&oclk->node, &clk_hw_omap_clocks);
node              573 drivers/clk/ti/clk.c 	list_for_each_entry(hw, &clk_hw_omap_clocks, node) {
node              593 drivers/clk/ti/clk.c 	list_for_each_entry(oclk, &clk_hw_omap_clocks, node) {
node               57 drivers/clk/ti/clkctrl.c 	struct list_head node;
node              241 drivers/clk/ti/clkctrl.c 	list_for_each_entry(entry, &provider->clocks, node) {
node              257 drivers/clk/ti/clkctrl.c 			 struct device_node *node, struct clk_hw *clk_hw,
node              268 drivers/clk/ti/clkctrl.c 				      node->parent, node, offset,
node              271 drivers/clk/ti/clkctrl.c 		init.name = kasprintf(GFP_KERNEL, "%pOFn:%04x:%d", node,
node              295 drivers/clk/ti/clkctrl.c 	list_add(&clkctrl_clk->node, &provider->clocks);
node              307 drivers/clk/ti/clkctrl.c 		       struct device_node *node, u16 offset,
node              320 drivers/clk/ti/clkctrl.c 	if (_ti_clkctrl_clk_register(provider, node, &clk_hw->hw, offset,
node              328 drivers/clk/ti/clkctrl.c 		      struct device_node *node, u16 offset,
node              355 drivers/clk/ti/clkctrl.c 	if (_ti_clkctrl_clk_register(provider, node, &mux->hw, offset,
node              363 drivers/clk/ti/clkctrl.c 		      struct device_node *node, u16 offset,
node              386 drivers/clk/ti/clkctrl.c 		       node, offset, data->bit);
node              391 drivers/clk/ti/clkctrl.c 	if (_ti_clkctrl_clk_register(provider, node, &div->hw, offset,
node              399 drivers/clk/ti/clkctrl.c 			  struct device_node *node,
node              411 drivers/clk/ti/clkctrl.c 			_ti_clkctrl_setup_gate(provider, node, data->offset,
node              416 drivers/clk/ti/clkctrl.c 			_ti_clkctrl_setup_div(provider, node, data->offset,
node              421 drivers/clk/ti/clkctrl.c 			_ti_clkctrl_setup_mux(provider, node, data->offset,
node              440 drivers/clk/ti/clkctrl.c static void __init _ti_omap4_clkctrl_setup(struct device_node *node)
node              456 drivers/clk/ti/clkctrl.c 	    of_node_name_eq(node, "clk"))
node              459 drivers/clk/ti/clkctrl.c 	addrp = of_get_address(node, 0, NULL, NULL);
node              460 drivers/clk/ti/clkctrl.c 	addr = (u32)of_translate_address(node, addrp);
node              527 drivers/clk/ti/clkctrl.c 		pr_err("%pOF not found from clkctrl data.\n", node);
node              535 drivers/clk/ti/clkctrl.c 	provider->base = of_iomap(node, 0);
node              538 drivers/clk/ti/clkctrl.c 		provider->clkdm_name = kasprintf(GFP_KERNEL, "%pOFnxxx", node->parent);
node              550 drivers/clk/ti/clkctrl.c 		provider->clkdm_name = kasprintf(GFP_KERNEL, "%pOFn", node);
node              592 drivers/clk/ti/clkctrl.c 		_ti_clkctrl_setup_subclks(provider, node, reg_data,
node              614 drivers/clk/ti/clkctrl.c 					      node->parent, node,
node              618 drivers/clk/ti/clkctrl.c 					      node, reg_data->offset, 0);
node              633 drivers/clk/ti/clkctrl.c 		list_add(&clkctrl_clk->node, &provider->clocks);
node              638 drivers/clk/ti/clkctrl.c 	ret = of_clk_add_hw_provider(node, _ti_omap4_clkctrl_xlate, provider);
node              640 drivers/clk/ti/clkctrl.c 		ti_clk_retry_init(node, provider, _clkctrl_add_provider);
node              226 drivers/clk/ti/clock.h int ti_clk_get_reg_addr(struct device_node *node, int index,
node              229 drivers/clk/ti/clock.h int ti_clk_retry_init(struct device_node *node, void *user,
node              231 drivers/clk/ti/clock.h int ti_clk_add_component(struct device_node *node, struct clk_hw *hw, int type);
node              233 drivers/clk/ti/clock.h int of_ti_clk_autoidle_setup(struct device_node *node);
node              128 drivers/clk/ti/clockdomain.c static void __init of_ti_clockdomain_setup(struct device_node *node)
node              132 drivers/clk/ti/clockdomain.c 	const char *clkdm_name = node->name;
node              136 drivers/clk/ti/clockdomain.c 	num_clks = of_clk_get_parent_count(node);
node              139 drivers/clk/ti/clockdomain.c 		clk = of_clk_get(node, i);
node              142 drivers/clk/ti/clockdomain.c 			       __func__, node, i, PTR_ERR(clk));
node               64 drivers/clk/ti/composite.c 	struct device_node *node;
node               76 drivers/clk/ti/composite.c static struct device_node *_get_component_node(struct device_node *node, int i)
node               81 drivers/clk/ti/composite.c 	rc = of_parse_phandle_with_args(node, "clocks", "#clock-cells", i,
node               89 drivers/clk/ti/composite.c static struct component_clk *_lookup_component(struct device_node *node)
node               94 drivers/clk/ti/composite.c 		if (comp->node == node)
node              120 drivers/clk/ti/composite.c 				       struct device_node *node)
node              139 drivers/clk/ti/composite.c 				 cclk->comp_nodes[i]->name, node);
node              140 drivers/clk/ti/composite.c 			if (!ti_clk_retry_init(node, hw,
node              148 drivers/clk/ti/composite.c 			       node, component_clk_types[comp->type]);
node              171 drivers/clk/ti/composite.c 		pr_err("%s: no parents found for %pOFn!\n", __func__, node);
node              175 drivers/clk/ti/composite.c 	clk = clk_register_composite(NULL, node->name,
node              185 drivers/clk/ti/composite.c 		ret = ti_clk_add_alias(NULL, clk, node->name);
node              190 drivers/clk/ti/composite.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              205 drivers/clk/ti/composite.c static void __init of_ti_composite_clk_setup(struct device_node *node)
node              212 drivers/clk/ti/composite.c 	num_clks = of_clk_get_parent_count(node);
node              215 drivers/clk/ti/composite.c 		pr_err("composite clk %pOFn must have component(s)\n", node);
node              225 drivers/clk/ti/composite.c 		cclk->comp_nodes[i] = _get_component_node(node, i);
node              227 drivers/clk/ti/composite.c 	_register_composite(&cclk->hw, node);
node              241 drivers/clk/ti/composite.c int __init ti_clk_add_component(struct device_node *node, struct clk_hw *hw,
node              248 drivers/clk/ti/composite.c 	num_parents = of_clk_get_parent_count(node);
node              251 drivers/clk/ti/composite.c 		pr_err("component-clock %pOFn must have parent(s)\n", node);
node              259 drivers/clk/ti/composite.c 	of_clk_parent_fill(node, parent_names, num_parents);
node              270 drivers/clk/ti/composite.c 	clk->node = node;
node              429 drivers/clk/ti/divider.c __init ti_clk_get_div_table(struct device_node *node)
node              438 drivers/clk/ti/divider.c 	divspec = of_get_property(node, "ti,dividers", &num_div);
node              449 drivers/clk/ti/divider.c 		of_property_read_u32_index(node, "ti,dividers", i, &val);
node              455 drivers/clk/ti/divider.c 		pr_err("no valid dividers for %pOFn table\n", node);
node              467 drivers/clk/ti/divider.c 		of_property_read_u32_index(node, "ti,dividers", i, &val);
node              478 drivers/clk/ti/divider.c static int _get_divider_width(struct device_node *node,
node              489 drivers/clk/ti/divider.c 		if (of_property_read_u32(node, "ti,min-div", &min_div))
node              492 drivers/clk/ti/divider.c 		if (of_property_read_u32(node, "ti,max-div", &max_div)) {
node              493 drivers/clk/ti/divider.c 			pr_err("no max-div for %pOFn!\n", node);
node              522 drivers/clk/ti/divider.c static int __init ti_clk_divider_populate(struct device_node *node,
node              529 drivers/clk/ti/divider.c 	ret = ti_clk_get_reg_addr(node, 0, reg);
node              533 drivers/clk/ti/divider.c 	if (!of_property_read_u32(node, "ti,bit-shift", &val))
node              539 drivers/clk/ti/divider.c 		if (!of_property_read_u32(node, "ti,latch-bit", &val))
node              548 drivers/clk/ti/divider.c 	if (of_property_read_bool(node, "ti,index-starts-at-one"))
node              551 drivers/clk/ti/divider.c 	if (of_property_read_bool(node, "ti,index-power-of-two"))
node              554 drivers/clk/ti/divider.c 	if (of_property_read_bool(node, "ti,set-rate-parent"))
node              557 drivers/clk/ti/divider.c 	*table = ti_clk_get_div_table(node);
node              562 drivers/clk/ti/divider.c 	*width = _get_divider_width(node, *table, *div_flags);
node              573 drivers/clk/ti/divider.c static void __init of_ti_divider_clk_setup(struct device_node *node)
node              585 drivers/clk/ti/divider.c 	parent_name = of_clk_get_parent_name(node, 0);
node              587 drivers/clk/ti/divider.c 	if (ti_clk_divider_populate(node, &reg, &table, &flags,
node              591 drivers/clk/ti/divider.c 	clk = _register_divider(NULL, node->name, parent_name, flags, &reg,
node              595 drivers/clk/ti/divider.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              596 drivers/clk/ti/divider.c 		of_ti_clk_autoidle_setup(node);
node              605 drivers/clk/ti/divider.c static void __init of_ti_composite_divider_clk_setup(struct device_node *node)
node              614 drivers/clk/ti/divider.c 	if (ti_clk_divider_populate(node, &div->reg, &div->table, &val,
node              619 drivers/clk/ti/divider.c 	if (!ti_clk_add_component(node, &div->hw, CLK_COMPONENT_TYPE_DIVIDER))
node              162 drivers/clk/ti/dpll.c 				  struct device_node *node)
node              170 drivers/clk/ti/dpll.c 	clk = of_clk_get(node, 0);
node              173 drivers/clk/ti/dpll.c 			 node);
node              174 drivers/clk/ti/dpll.c 		if (!ti_clk_retry_init(node, hw, _register_dpll))
node              182 drivers/clk/ti/dpll.c 	clk = of_clk_get(node, 1);
node              186 drivers/clk/ti/dpll.c 			 node);
node              187 drivers/clk/ti/dpll.c 		if (!ti_clk_retry_init(node, hw, _register_dpll))
node              196 drivers/clk/ti/dpll.c 	clk = ti_clk_register_omap_hw(NULL, &clk_hw->hw, node->name);
node              199 drivers/clk/ti/dpll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              223 drivers/clk/ti/dpll.c static void _register_dpll_x2(struct device_node *node,
node              230 drivers/clk/ti/dpll.c 	const char *name = node->name;
node              233 drivers/clk/ti/dpll.c 	parent_name = of_clk_get_parent_name(node, 0);
node              235 drivers/clk/ti/dpll.c 		pr_err("%pOFn must have parent\n", node);
node              257 drivers/clk/ti/dpll.c 		ret = of_property_count_elems_of_size(node, "reg", 1);
node              260 drivers/clk/ti/dpll.c 		} else if (ti_clk_get_reg_addr(node, 0, &clk_hw->clksel_reg)) {
node              273 drivers/clk/ti/dpll.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              285 drivers/clk/ti/dpll.c static void __init of_ti_dpll_setup(struct device_node *node,
node              305 drivers/clk/ti/dpll.c 	init->name = node->name;
node              308 drivers/clk/ti/dpll.c 	init->num_parents = of_clk_get_parent_count(node);
node              310 drivers/clk/ti/dpll.c 		pr_err("%pOFn must have parent(s)\n", node);
node              318 drivers/clk/ti/dpll.c 	of_clk_parent_fill(node, parent_names, init->num_parents);
node              322 drivers/clk/ti/dpll.c 	if (ti_clk_get_reg_addr(node, 0, &dd->control_reg))
node              331 drivers/clk/ti/dpll.c 		if (ti_clk_get_reg_addr(node, 1, &dd->mult_div1_reg))
node              338 drivers/clk/ti/dpll.c 		if (ti_clk_get_reg_addr(node, 1, &dd->idlest_reg))
node              341 drivers/clk/ti/dpll.c 		if (ti_clk_get_reg_addr(node, 2, &dd->mult_div1_reg))
node              346 drivers/clk/ti/dpll.c 		if (ti_clk_get_reg_addr(node, 3, &dd->autoidle_reg))
node              350 drivers/clk/ti/dpll.c 	if (of_property_read_bool(node, "ti,low-power-stop"))
node              353 drivers/clk/ti/dpll.c 	if (of_property_read_bool(node, "ti,low-power-bypass"))
node              356 drivers/clk/ti/dpll.c 	if (of_property_read_bool(node, "ti,lock"))
node              362 drivers/clk/ti/dpll.c 	_register_dpll(&clk_hw->hw, node);
node              374 drivers/clk/ti/dpll.c static void __init of_ti_omap4_dpll_x2_setup(struct device_node *node)
node              376 drivers/clk/ti/dpll.c 	_register_dpll_x2(node, &dpll_x2_ck_ops, &clkhwops_omap4_dpllmx);
node              383 drivers/clk/ti/dpll.c static void __init of_ti_am3_dpll_x2_setup(struct device_node *node)
node              385 drivers/clk/ti/dpll.c 	_register_dpll_x2(node, &dpll_x2_ck_ops, NULL);
node              392 drivers/clk/ti/dpll.c static void __init of_ti_omap3_dpll_setup(struct device_node *node)
node              409 drivers/clk/ti/dpll.c 	     of_node_name_eq(node, "dpll5_ck"))
node              410 drivers/clk/ti/dpll.c 		of_ti_dpll_setup(node, &omap3_dpll5_ck_ops, &dd);
node              412 drivers/clk/ti/dpll.c 		of_ti_dpll_setup(node, &omap3_dpll_ck_ops, &dd);
node              417 drivers/clk/ti/dpll.c static void __init of_ti_omap3_core_dpll_setup(struct device_node *node)
node              431 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &omap3_dpll_core_ck_ops, &dd);
node              436 drivers/clk/ti/dpll.c static void __init of_ti_omap3_per_dpll_setup(struct device_node *node)
node              451 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd);
node              456 drivers/clk/ti/dpll.c static void __init of_ti_omap3_per_jtype_dpll_setup(struct device_node *node)
node              473 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd);
node              479 drivers/clk/ti/dpll.c static void __init of_ti_omap4_dpll_setup(struct device_node *node)
node              493 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
node              498 drivers/clk/ti/dpll.c static void __init of_ti_omap5_mpu_dpll_setup(struct device_node *node)
node              514 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
node              519 drivers/clk/ti/dpll.c static void __init of_ti_omap4_core_dpll_setup(struct device_node *node)
node              533 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd);
node              540 drivers/clk/ti/dpll.c static void __init of_ti_omap4_m4xen_dpll_setup(struct device_node *node)
node              556 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd);
node              561 drivers/clk/ti/dpll.c static void __init of_ti_omap4_jtype_dpll_setup(struct device_node *node)
node              577 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd);
node              583 drivers/clk/ti/dpll.c static void __init of_ti_am3_no_gate_dpll_setup(struct device_node *node)
node              597 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd);
node              602 drivers/clk/ti/dpll.c static void __init of_ti_am3_jtype_dpll_setup(struct device_node *node)
node              617 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
node              622 drivers/clk/ti/dpll.c static void __init of_ti_am3_no_gate_jtype_dpll_setup(struct device_node *node)
node              637 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd);
node              643 drivers/clk/ti/dpll.c static void __init of_ti_am3_dpll_setup(struct device_node *node)
node              657 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
node              661 drivers/clk/ti/dpll.c static void __init of_ti_am3_core_dpll_setup(struct device_node *node)
node              675 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd);
node              680 drivers/clk/ti/dpll.c static void __init of_ti_omap2_core_dpll_setup(struct device_node *node)
node              690 drivers/clk/ti/dpll.c 	of_ti_dpll_setup(node, &omap2_dpll_core_ck_ops, &dd);
node              532 drivers/clk/ti/fapll.c static void __init ti_fapll_setup(struct device_node *node)
node              555 drivers/clk/ti/fapll.c 	init->name = node->name;
node              557 drivers/clk/ti/fapll.c 	init->num_parents = of_clk_get_parent_count(node);
node              559 drivers/clk/ti/fapll.c 		pr_err("%pOFn must have two parents\n", node);
node              563 drivers/clk/ti/fapll.c 	of_clk_parent_fill(node, parent_name, 2);
node              566 drivers/clk/ti/fapll.c 	fd->clk_ref = of_clk_get(node, 0);
node              568 drivers/clk/ti/fapll.c 		pr_err("%pOFn could not get clk_ref\n", node);
node              572 drivers/clk/ti/fapll.c 	fd->clk_bypass = of_clk_get(node, 1);
node              574 drivers/clk/ti/fapll.c 		pr_err("%pOFn could not get clk_bypass\n", node);
node              578 drivers/clk/ti/fapll.c 	fd->base = of_iomap(node, 0);
node              580 drivers/clk/ti/fapll.c 		pr_err("%pOFn could not get IO base\n", node);
node              587 drivers/clk/ti/fapll.c 	fd->name = node->name;
node              612 drivers/clk/ti/fapll.c 		if (of_property_read_string_index(node, "clock-output-names",
node              616 drivers/clk/ti/fapll.c 		if (of_property_read_u32_index(node, "clock-indices", i,
node              634 drivers/clk/ti/fapll.c 						 output_name, node->name,
node              646 drivers/clk/ti/fapll.c 	of_clk_add_provider(node, of_clk_src_onecell_get, &fd->outputs);
node               36 drivers/clk/ti/fixed-factor.c static void __init of_ti_fixed_factor_clk_setup(struct device_node *node)
node               39 drivers/clk/ti/fixed-factor.c 	const char *clk_name = node->name;
node               44 drivers/clk/ti/fixed-factor.c 	if (of_property_read_u32(node, "ti,clock-div", &div)) {
node               45 drivers/clk/ti/fixed-factor.c 		pr_err("%pOFn must have a clock-div property\n", node);
node               49 drivers/clk/ti/fixed-factor.c 	if (of_property_read_u32(node, "ti,clock-mult", &mult)) {
node               50 drivers/clk/ti/fixed-factor.c 		pr_err("%pOFn must have a clock-mult property\n", node);
node               54 drivers/clk/ti/fixed-factor.c 	if (of_property_read_bool(node, "ti,set-rate-parent"))
node               57 drivers/clk/ti/fixed-factor.c 	parent_name = of_clk_get_parent_name(node, 0);
node               63 drivers/clk/ti/fixed-factor.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               64 drivers/clk/ti/fixed-factor.c 		of_ti_clk_autoidle_setup(node);
node              134 drivers/clk/ti/gate.c static void __init _of_ti_gate_clk_setup(struct device_node *node,
node              147 drivers/clk/ti/gate.c 		if (ti_clk_get_reg_addr(node, 0, &reg))
node              150 drivers/clk/ti/gate.c 		if (!of_property_read_u32(node, "ti,bit-shift", &val))
node              154 drivers/clk/ti/gate.c 	if (of_clk_get_parent_count(node) != 1) {
node              155 drivers/clk/ti/gate.c 		pr_err("%pOFn must have 1 parent\n", node);
node              159 drivers/clk/ti/gate.c 	parent_name = of_clk_get_parent_name(node, 0);
node              161 drivers/clk/ti/gate.c 	if (of_property_read_bool(node, "ti,set-rate-parent"))
node              164 drivers/clk/ti/gate.c 	if (of_property_read_bool(node, "ti,set-bit-to-disable"))
node              167 drivers/clk/ti/gate.c 	clk = _register_gate(NULL, node->name, parent_name, flags, &reg,
node              171 drivers/clk/ti/gate.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              175 drivers/clk/ti/gate.c _of_ti_composite_gate_clk_setup(struct device_node *node,
node              185 drivers/clk/ti/gate.c 	if (ti_clk_get_reg_addr(node, 0, &gate->enable_reg))
node              188 drivers/clk/ti/gate.c 	of_property_read_u32(node, "ti,bit-shift", &val);
node              193 drivers/clk/ti/gate.c 	if (!ti_clk_add_component(node, &gate->hw, CLK_COMPONENT_TYPE_GATE))
node              201 drivers/clk/ti/gate.c of_ti_composite_no_wait_gate_clk_setup(struct device_node *node)
node              203 drivers/clk/ti/gate.c 	_of_ti_composite_gate_clk_setup(node, NULL);
node              209 drivers/clk/ti/gate.c static void __init of_ti_composite_interface_clk_setup(struct device_node *node)
node              211 drivers/clk/ti/gate.c 	_of_ti_composite_gate_clk_setup(node, &clkhwops_iclk_wait);
node              217 drivers/clk/ti/gate.c static void __init of_ti_composite_gate_clk_setup(struct device_node *node)
node              219 drivers/clk/ti/gate.c 	_of_ti_composite_gate_clk_setup(node, &clkhwops_wait);
node              225 drivers/clk/ti/gate.c static void __init of_ti_clkdm_gate_clk_setup(struct device_node *node)
node              227 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clkdm_clk_ops, NULL);
node              232 drivers/clk/ti/gate.c static void __init of_ti_hsdiv_gate_clk_setup(struct device_node *node)
node              234 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clk_hsdiv_restore_ops,
node              240 drivers/clk/ti/gate.c static void __init of_ti_gate_clk_setup(struct device_node *node)
node              242 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clk_ops, NULL);
node              246 drivers/clk/ti/gate.c static void __init of_ti_wait_gate_clk_setup(struct device_node *node)
node              248 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clk_ops, &clkhwops_wait);
node              254 drivers/clk/ti/gate.c static void __init of_ti_am35xx_gate_clk_setup(struct device_node *node)
node              256 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clk_ops,
node              262 drivers/clk/ti/gate.c static void __init of_ti_dss_gate_clk_setup(struct device_node *node)
node              264 drivers/clk/ti/gate.c 	_of_ti_gate_clk_setup(node, &omap_gate_clk_ops,
node               68 drivers/clk/ti/interface.c static void __init _of_ti_interface_clk_setup(struct device_node *node,
node               77 drivers/clk/ti/interface.c 	if (ti_clk_get_reg_addr(node, 0, &reg))
node               80 drivers/clk/ti/interface.c 	if (!of_property_read_u32(node, "ti,bit-shift", &val))
node               83 drivers/clk/ti/interface.c 	parent_name = of_clk_get_parent_name(node, 0);
node               85 drivers/clk/ti/interface.c 		pr_err("%pOFn must have a parent\n", node);
node               89 drivers/clk/ti/interface.c 	clk = _register_interface(NULL, node->name, parent_name, &reg,
node               93 drivers/clk/ti/interface.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node               96 drivers/clk/ti/interface.c static void __init of_ti_interface_clk_setup(struct device_node *node)
node               98 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node, &clkhwops_iclk_wait);
node              103 drivers/clk/ti/interface.c static void __init of_ti_no_wait_interface_clk_setup(struct device_node *node)
node              105 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node, &clkhwops_iclk);
node              111 drivers/clk/ti/interface.c static void __init of_ti_hsotgusb_interface_clk_setup(struct device_node *node)
node              113 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node,
node              119 drivers/clk/ti/interface.c static void __init of_ti_dss_interface_clk_setup(struct device_node *node)
node              121 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node,
node              127 drivers/clk/ti/interface.c static void __init of_ti_ssi_interface_clk_setup(struct device_node *node)
node              129 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node, &clkhwops_omap3430es2_iclk_ssi_wait);
node              134 drivers/clk/ti/interface.c static void __init of_ti_am35xx_interface_clk_setup(struct device_node *node)
node              136 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node, &clkhwops_am35xx_ipss_wait);
node              143 drivers/clk/ti/interface.c static void __init of_ti_omap2430_interface_clk_setup(struct device_node *node)
node              145 drivers/clk/ti/interface.c 	_of_ti_interface_clk_setup(node, &clkhwops_omap2430_i2chs_wait);
node              173 drivers/clk/ti/mux.c static void of_mux_clk_setup(struct device_node *node)
node              185 drivers/clk/ti/mux.c 	num_parents = of_clk_get_parent_count(node);
node              187 drivers/clk/ti/mux.c 		pr_err("mux-clock %pOFn must have parents\n", node);
node              194 drivers/clk/ti/mux.c 	of_clk_parent_fill(node, parent_names, num_parents);
node              196 drivers/clk/ti/mux.c 	if (ti_clk_get_reg_addr(node, 0, &reg))
node              199 drivers/clk/ti/mux.c 	of_property_read_u32(node, "ti,bit-shift", &shift);
node              201 drivers/clk/ti/mux.c 	of_property_read_u32(node, "ti,latch-bit", &latch);
node              203 drivers/clk/ti/mux.c 	if (of_property_read_bool(node, "ti,index-starts-at-one"))
node              206 drivers/clk/ti/mux.c 	if (of_property_read_bool(node, "ti,set-rate-parent"))
node              216 drivers/clk/ti/mux.c 	clk = _register_mux(NULL, node->name, parent_names, num_parents,
node              221 drivers/clk/ti/mux.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              257 drivers/clk/ti/mux.c static void __init of_ti_composite_mux_clk_setup(struct device_node *node)
node              267 drivers/clk/ti/mux.c 	if (ti_clk_get_reg_addr(node, 0, &mux->reg))
node              270 drivers/clk/ti/mux.c 	if (!of_property_read_u32(node, "ti,bit-shift", &val))
node              273 drivers/clk/ti/mux.c 	if (of_property_read_bool(node, "ti,index-starts-at-one"))
node              276 drivers/clk/ti/mux.c 	num_parents = of_clk_get_parent_count(node);
node              279 drivers/clk/ti/mux.c 		pr_err("%pOFn must have parents\n", node);
node              286 drivers/clk/ti/mux.c 	if (!ti_clk_add_component(node, &mux->hw, CLK_COMPONENT_TYPE_MUX))
node               29 drivers/clk/versatile/clk-sp810.c 	struct device_node *node;
node               82 drivers/clk/versatile/clk-sp810.c static void __init clk_sp810_of_setup(struct device_node *node)
node               96 drivers/clk/versatile/clk-sp810.c 	if (of_clk_parent_fill(node, parent_names, num) != num) {
node              102 drivers/clk/versatile/clk-sp810.c 	sp810->node = node;
node              103 drivers/clk/versatile/clk-sp810.c 	sp810->base = of_iomap(node, 0);
node              112 drivers/clk/versatile/clk-sp810.c 	deprecated = !of_find_property(node, "assigned-clock-parents", NULL);
node              135 drivers/clk/versatile/clk-sp810.c 	of_clk_add_provider(node, clk_sp810_timerclken_of_get, sp810);
node               74 drivers/clk/zynqmp/clkc.c 	struct clock_topology node[MAX_NODES];
node              522 drivers/clk/zynqmp/clkc.c 	clk_nodes = clock[clk_id].node;
node              565 drivers/clk/zynqmp/clkc.c 	nodes = clock[clk_id].node;
node              690 drivers/clk/zynqmp/clkc.c 		ret = zynqmp_clock_get_topology(i, clock[i].node,
node               32 drivers/clocksource/arc_timer.c static int noinline arc_get_timer_clk(struct device_node *node)
node               37 drivers/clocksource/arc_timer.c 	clk = of_clk_get(node, 0);
node              103 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_gfrc(struct device_node *node)
node              114 drivers/clocksource/arc_timer.c 	ret = arc_get_timer_clk(node);
node              161 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_rtc(struct device_node *node)
node              178 drivers/clocksource/arc_timer.c 	ret = arc_get_timer_clk(node);
node              214 drivers/clocksource/arc_timer.c static int __init arc_cs_setup_timer1(struct device_node *node)
node              222 drivers/clocksource/arc_timer.c 	ret = arc_get_timer_clk(node);
node              325 drivers/clocksource/arc_timer.c static int __init arc_clockevent_setup(struct device_node *node)
node              330 drivers/clocksource/arc_timer.c 	arc_timer_irq = irq_of_parse_and_map(node, 0);
node              336 drivers/clocksource/arc_timer.c 	ret = arc_get_timer_clk(node);
node               70 drivers/clocksource/bcm2835_timer.c static int __init bcm2835_timer_init(struct device_node *node)
node               77 drivers/clocksource/bcm2835_timer.c 	base = of_iomap(node, 0);
node               83 drivers/clocksource/bcm2835_timer.c 	ret = of_property_read_u32(node, "clock-frequency", &freq);
node               92 drivers/clocksource/bcm2835_timer.c 	clocksource_mmio_init(base + REG_COUNTER_LO, node->name,
node               95 drivers/clocksource/bcm2835_timer.c 	irq = irq_of_parse_and_map(node, DEFAULT_TIMER);
node              111 drivers/clocksource/bcm2835_timer.c 	timer->evt.name = node->name;
node              116 drivers/clocksource/bcm2835_timer.c 	timer->act.name = node->name;
node              169 drivers/clocksource/bcm_kona_timer.c static int __init kona_timer_init(struct device_node *node)
node              174 drivers/clocksource/bcm_kona_timer.c 	external_clk = of_clk_get_by_name(node, NULL);
node              179 drivers/clocksource/bcm_kona_timer.c 	} else if (!of_property_read_u32(node, "clock-frequency", &freq)) {
node              187 drivers/clocksource/bcm_kona_timer.c 	timers.tmr_irq = irq_of_parse_and_map(node, 0);
node              190 drivers/clocksource/bcm_kona_timer.c 	timers.tmr_regs = of_iomap(node, 0);
node               52 drivers/clocksource/clksrc-dbx500-prcmu.c static int __init clksrc_dbx500_prcmu_init(struct device_node *node)
node               54 drivers/clocksource/clksrc-dbx500-prcmu.c 	clksrc_dbx500_timer_base = of_iomap(node, 0);
node              130 drivers/clocksource/h8300_timer16.c static int __init h8300_16timer_init(struct device_node *node)
node              137 drivers/clocksource/h8300_timer16.c 	clk = of_clk_get(node, 0);
node              144 drivers/clocksource/h8300_timer16.c 	base[REG_CH] = of_iomap(node, 0);
node              150 drivers/clocksource/h8300_timer16.c 	base[REG_COMM] = of_iomap(node, 1);
node              157 drivers/clocksource/h8300_timer16.c 	irq = irq_of_parse_and_map(node, 0);
node              163 drivers/clocksource/h8300_timer16.c 	of_property_read_u32(node, "renesas,channel", &ch);
node              160 drivers/clocksource/h8300_timer8.c static int __init h8300_8timer_init(struct device_node *node)
node              166 drivers/clocksource/h8300_timer8.c 	clk = of_clk_get(node, 0);
node              173 drivers/clocksource/h8300_timer8.c 	base = of_iomap(node, 0);
node              180 drivers/clocksource/h8300_timer8.c 	irq = irq_of_parse_and_map(node, 0);
node              123 drivers/clocksource/h8300_tpu.c static int __init h8300_tpu_init(struct device_node *node)
node              129 drivers/clocksource/h8300_tpu.c 	clk = of_clk_get(node, 0);
node              135 drivers/clocksource/h8300_tpu.c 	base[CH_L] = of_iomap(node, CH_L);
node              140 drivers/clocksource/h8300_tpu.c 	base[CH_H] = of_iomap(node, CH_H);
node              136 drivers/clocksource/jcore-pit.c static int __init jcore_pit_init(struct device_node *node)
node              143 drivers/clocksource/jcore-pit.c 	jcore_pit_base = of_iomap(node, 0);
node              149 drivers/clocksource/jcore-pit.c 	pit_irq = irq_of_parse_and_map(node, 0);
node              221 drivers/clocksource/jcore-pit.c 		pit->base = of_iomap(node, cpu);
node              186 drivers/clocksource/mips-gic-timer.c static int __init gic_clocksource_of_init(struct device_node *node)
node              191 drivers/clocksource/mips-gic-timer.c 	if (!mips_gic_present() || !node->parent ||
node              192 drivers/clocksource/mips-gic-timer.c 	    !of_device_is_compatible(node->parent, "mti,gic")) {
node              197 drivers/clocksource/mips-gic-timer.c 	clk = of_clk_get(node, 0);
node              207 drivers/clocksource/mips-gic-timer.c 	} else if (of_property_read_u32(node, "clock-frequency",
node              212 drivers/clocksource/mips-gic-timer.c 	gic_timer_irq = irq_of_parse_and_map(node, 0);
node              247 drivers/clocksource/nomadik-mtu.c static int __init nmdk_timer_of_init(struct device_node *node)
node              254 drivers/clocksource/nomadik-mtu.c 	base = of_iomap(node, 0);
node              260 drivers/clocksource/nomadik-mtu.c 	pclk = of_clk_get_by_name(node, "apb_pclk");
node              266 drivers/clocksource/nomadik-mtu.c 	clk = of_clk_get_by_name(node, "timclk");
node              272 drivers/clocksource/nomadik-mtu.c 	irq = irq_of_parse_and_map(node, 0);
node              211 drivers/clocksource/timer-atcpit100.c static void fill_vdso_need_info(struct device_node *node)
node              214 drivers/clocksource/timer-atcpit100.c 	of_address_to_resource(node, 0, &timer_res);
node              221 drivers/clocksource/timer-atcpit100.c static int __init atcpit100_timer_init(struct device_node *node)
node              227 drivers/clocksource/timer-atcpit100.c 	ret = timer_of_init(node, &to);
node              237 drivers/clocksource/timer-atcpit100.c 		node->name, timer_of_rate(&to), 300, 32,
node              260 drivers/clocksource/timer-atcpit100.c 	fill_vdso_need_info(node);
node              166 drivers/clocksource/timer-atmel-pit.c static int __init at91sam926x_pit_dt_init(struct device_node *node)
node              177 drivers/clocksource/timer-atmel-pit.c 	data->base = of_iomap(node, 0);
node              184 drivers/clocksource/timer-atmel-pit.c 	data->mck = of_clk_get(node, 0);
node              198 drivers/clocksource/timer-atmel-pit.c 	data->irq = irq_of_parse_and_map(node, 0);
node              184 drivers/clocksource/timer-atmel-st.c static int __init atmel_st_timer_init(struct device_node *node)
node              190 drivers/clocksource/timer-atmel-st.c 	regmap_st = syscon_node_to_regmap(node);
node              202 drivers/clocksource/timer-atmel-st.c 	irq  = irq_of_parse_and_map(node, 0);
node              217 drivers/clocksource/timer-atmel-st.c 	sclk = of_clk_get(node, 0);
node              357 drivers/clocksource/timer-atmel-tcb.c static int __init tcb_clksrc_init(struct device_node *node)
node              374 drivers/clocksource/timer-atmel-tcb.c 	tc.regs = of_iomap(node->parent, 0);
node              378 drivers/clocksource/timer-atmel-tcb.c 	t0_clk = of_clk_get_by_name(node->parent, "t0_clk");
node              382 drivers/clocksource/timer-atmel-tcb.c 	tc.slow_clk = of_clk_get_by_name(node->parent, "slow_clk");
node              387 drivers/clocksource/timer-atmel-tcb.c 	tc.clk[1] = of_clk_get_by_name(node->parent, "t1_clk");
node              390 drivers/clocksource/timer-atmel-tcb.c 	tc.clk[2] = of_clk_get_by_name(node->parent, "t2_clk");
node              394 drivers/clocksource/timer-atmel-tcb.c 	tc.irq[2] = of_irq_get(node->parent, 2);
node              396 drivers/clocksource/timer-atmel-tcb.c 		tc.irq[2] = of_irq_get(node->parent, 0);
node              401 drivers/clocksource/timer-atmel-tcb.c 	match = of_match_node(atmel_tcb_of_match, node->parent);
node              435 drivers/clocksource/timer-atmel-tcb.c 	clksrc.name = kbasename(node->parent->full_name);
node              436 drivers/clocksource/timer-atmel-tcb.c 	clkevt.clkevt.name = kbasename(node->parent->full_name);
node              152 drivers/clocksource/timer-digicolor.c static int __init digicolor_timer_init(struct device_node *node)
node              162 drivers/clocksource/timer-digicolor.c 	dc_timer_dev.base = of_iomap(node, 0);
node              168 drivers/clocksource/timer-digicolor.c 	irq = irq_of_parse_and_map(node, dc_timer_dev.timer_id);
node              174 drivers/clocksource/timer-digicolor.c 	clk = of_clk_get(node, 0);
node              188 drivers/clocksource/timer-digicolor.c 	clocksource_mmio_init(dc_timer_dev.base + COUNT(TIMER_B), node->name,
node              163 drivers/clocksource/timer-integrator-ap.c static int __init integrator_ap_timer_init_of(struct device_node *node)
node              173 drivers/clocksource/timer-integrator-ap.c 	base = of_io_request_and_map(node, 0, "integrator-timer");
node              177 drivers/clocksource/timer-integrator-ap.c 	clk = of_clk_get(node, 0);
node              179 drivers/clocksource/timer-integrator-ap.c 		pr_err("No clock for %pOFn\n", node);
node              202 drivers/clocksource/timer-integrator-ap.c 	if (node == alias_node)
node              217 drivers/clocksource/timer-integrator-ap.c 	if (node == alias_node) {
node              219 drivers/clocksource/timer-integrator-ap.c 		irq = irq_of_parse_and_map(node, 0);
node              258 drivers/clocksource/timer-mediatek.c static int __init mtk_syst_init(struct device_node *node)
node              269 drivers/clocksource/timer-mediatek.c 	ret = timer_of_init(node, &to);
node              279 drivers/clocksource/timer-mediatek.c static int __init mtk_gpt_init(struct device_node *node)
node              291 drivers/clocksource/timer-mediatek.c 	ret = timer_of_init(node, &to);
node              298 drivers/clocksource/timer-mediatek.c 			      node->name, timer_of_rate(&to), 300, 32,
node              160 drivers/clocksource/timer-meson6.c static int __init meson6_timer_init(struct device_node *node)
node              165 drivers/clocksource/timer-meson6.c 	timer_base = of_io_request_and_map(node, 0, "meson6-timer");
node              171 drivers/clocksource/timer-meson6.c 	irq = irq_of_parse_and_map(node, 0);
node              185 drivers/clocksource/timer-meson6.c 	clocksource_mmio_init(timer_base + MESON_ISA_TIMERE, node->name,
node              168 drivers/clocksource/timer-milbeaut.c static int __init mlb_timer_init(struct device_node *node)
node              173 drivers/clocksource/timer-milbeaut.c 	ret = timer_of_init(node, &to);
node              180 drivers/clocksource/timer-milbeaut.c 		node->name, rate, MLB_TIMER_RATING, 32,
node               49 drivers/clocksource/timer-nps.c static int __init nps_get_timer_clk(struct device_node *node,
node               55 drivers/clocksource/timer-nps.c 	*clk = of_clk_get(node, 0);
node               87 drivers/clocksource/timer-nps.c static int __init nps_setup_clocksource(struct device_node *node)
node               99 drivers/clocksource/timer-nps.c 	ret = nps_get_timer_clk(node, &nps_timer1_freq, &clk);
node              243 drivers/clocksource/timer-nps.c static int __init nps_setup_clockevent(struct device_node *node)
node              248 drivers/clocksource/timer-nps.c 	nps_timer0_irq = irq_of_parse_and_map(node, 0);
node              254 drivers/clocksource/timer-nps.c 	ret = nps_get_timer_clk(node, &nps_timer0_freq, &clk);
node              116 drivers/clocksource/timer-owl.c static int __init owl_timer_init(struct device_node *node)
node              122 drivers/clocksource/timer-owl.c 	owl_timer_base = of_io_request_and_map(node, 0, "owl-timer");
node              131 drivers/clocksource/timer-owl.c 	timer1_irq = of_irq_get_byname(node, "timer1");
node              137 drivers/clocksource/timer-owl.c 	clk = of_clk_get(node, 0);
node              147 drivers/clocksource/timer-owl.c 	clocksource_mmio_init(owl_clksrc_base + OWL_Tx_VAL, node->name,
node              151 drivers/clocksource/timer-pistachio.c static int __init pistachio_clksrc_of_init(struct device_node *node)
node              158 drivers/clocksource/timer-pistachio.c 	pcs_gpt.base = of_iomap(node, 0);
node              164 drivers/clocksource/timer-pistachio.c 	periph_regs = syscon_regmap_lookup_by_phandle(node, "img,cr-periph");
node              177 drivers/clocksource/timer-pistachio.c 	sys_clk = of_clk_get_by_name(node, "sys");
node              183 drivers/clocksource/timer-pistachio.c 	fast_clk = of_clk_get_by_name(node, "fast");
node              288 drivers/clocksource/timer-stm32.c static int __init stm32_timer_init(struct device_node *node)
node              301 drivers/clocksource/timer-stm32.c 	ret = timer_of_init(node, to);
node              312 drivers/clocksource/timer-stm32.c 	rstc = of_reset_control_get(node, NULL);
node              169 drivers/clocksource/timer-sun4i.c static int __init sun4i_timer_init(struct device_node *node)
node              174 drivers/clocksource/timer-sun4i.c 	ret = timer_of_init(node, &to);
node              195 drivers/clocksource/timer-sun4i.c 				    node->name, timer_of_rate(&to), 350, 32,
node              186 drivers/clocksource/timer-sun5i.c static int __init sun5i_setup_clocksource(struct device_node *node,
node              226 drivers/clocksource/timer-sun5i.c 	cs->clksrc.name = node->name;
node              264 drivers/clocksource/timer-sun5i.c static int __init sun5i_setup_clockevent(struct device_node *node, void __iomem *base,
node              301 drivers/clocksource/timer-sun5i.c 	ce->clkevt.name = node->name;
node              337 drivers/clocksource/timer-sun5i.c static int __init sun5i_timer_init(struct device_node *node)
node              344 drivers/clocksource/timer-sun5i.c 	timer_base = of_io_request_and_map(node, 0, of_node_full_name(node));
node              350 drivers/clocksource/timer-sun5i.c 	irq = irq_of_parse_and_map(node, 0);
node              356 drivers/clocksource/timer-sun5i.c 	clk = of_clk_get(node, 0);
node              362 drivers/clocksource/timer-sun5i.c 	rstc = of_reset_control_get(node, NULL);
node              366 drivers/clocksource/timer-sun5i.c 	ret = sun5i_setup_clocksource(node, timer_base, clk, irq);
node              370 drivers/clocksource/timer-sun5i.c 	return sun5i_setup_clockevent(node, timer_base, clk, irq);
node              325 drivers/clocksource/timer-ti-dm.c 	list_for_each_entry(t, &omap_timer_list, node) {
node              474 drivers/clocksource/timer-ti-dm.c 	list_for_each_entry(timer, &omap_timer_list, node) {
node              758 drivers/clocksource/timer-ti-dm.c 	list_for_each_entry(timer, &omap_timer_list, node) {
node              857 drivers/clocksource/timer-ti-dm.c 	list_add_tail(&timer->node, &omap_timer_list);
node              885 drivers/clocksource/timer-ti-dm.c 	list_for_each_entry(timer, &omap_timer_list, node)
node              888 drivers/clocksource/timer-ti-dm.c 			list_del(&timer->node);
node               21 drivers/clocksource/timer-versatile.c static int __init versatile_sched_clock_init(struct device_node *node)
node               23 drivers/clocksource/timer-versatile.c 	void __iomem *base = of_iomap(node, 0);
node              117 drivers/clocksource/timer-zevio.c static int __init zevio_timer_add(struct device_node *node)
node              127 drivers/clocksource/timer-zevio.c 	timer->base = of_iomap(node, 0);
node              135 drivers/clocksource/timer-zevio.c 	timer->clk = of_clk_get(node, 0);
node              142 drivers/clocksource/timer-zevio.c 	timer->interrupt_regs = of_iomap(node, 1);
node              143 drivers/clocksource/timer-zevio.c 	irqnr = irq_of_parse_and_map(node, 0);
node              145 drivers/clocksource/timer-zevio.c 	of_address_to_resource(node, 0, &res);
node              148 drivers/clocksource/timer-zevio.c 			(unsigned long long)res.start, node);
node              152 drivers/clocksource/timer-zevio.c 			(unsigned long long)res.start, node);
node              209 drivers/clocksource/timer-zevio.c static int __init zevio_timer_init(struct device_node *node)
node              211 drivers/clocksource/timer-zevio.c 	return zevio_timer_add(node);
node              285 drivers/counter/ftm-quaddec.c 	struct device_node *node = pdev->dev.of_node;
node              302 drivers/counter/ftm-quaddec.c 	ftm->big_endian = of_property_read_bool(node, "big-endian");
node              127 drivers/cpufreq/armada-8k-cpufreq.c 	struct device_node *node;
node              130 drivers/cpufreq/armada-8k-cpufreq.c 	node = of_find_compatible_node(NULL, NULL, "marvell,ap806-cpu-clock");
node              131 drivers/cpufreq/armada-8k-cpufreq.c 	if (!node || !of_device_is_available(node)) {
node              132 drivers/cpufreq/armada-8k-cpufreq.c 		of_node_put(node);
node              135 drivers/cpufreq/armada-8k-cpufreq.c 	of_node_put(node);
node               69 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c 	u8 node, slow_mode;
node               74 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c 	node = pmi_msg.data1;
node               77 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c 	cpu = cbe_node_to_cpu(node);
node               79 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c 	pr_debug("cbe_handle_pmi: node: %d max_freq: %d\n", node, slow_mode);
node               94 drivers/cpufreq/ppc_cbe_cpufreq_pmi.c 		pr_debug("limiting node %d to slow mode %d\n", node, slow_mode);
node               99 drivers/crypto/cavium/cpt/cptvf.h 	u8 node; /* Operating node: Bits (46:44) in BAR0 address */
node              176 drivers/crypto/cavium/cpt/cptvf_main.c 	struct hlist_node *node;
node              184 drivers/crypto/cavium/cpt/cptvf_main.c 		hlist_for_each_entry_safe(chunk, node, &cqinfo->queue[i].chead,
node              627 drivers/crypto/cavium/cpt/cptvf_main.c 	cpumask_set_cpu(cpumask_local_spread(cpu, cptvf->node),
node              710 drivers/crypto/cavium/cpt/cptvf_main.c 	cptvf->node = dev_to_node(&pdev->dev);
node              247 drivers/crypto/cavium/cpt/cptvf_reqmanager.c 		struct hlist_node *node;
node              249 drivers/crypto/cavium/cpt/cptvf_reqmanager.c 		hlist_for_each(node, &queue->chead) {
node              250 drivers/crypto/cavium/cpt/cptvf_reqmanager.c 			chunk = hlist_entry(node, struct command_chunk,
node              233 drivers/crypto/cavium/nitrox/nitrox_dev.h 	int node;
node              106 drivers/crypto/cavium/nitrox/nitrox_lib.c 		cmdq = kzalloc_node(sizeof(*cmdq), GFP_KERNEL, ndev->node);
node              157 drivers/crypto/cavium/nitrox/nitrox_lib.c 				     GFP_KERNEL, ndev->node);
node              475 drivers/crypto/cavium/nitrox/nitrox_main.c 	ndev->node = dev_to_node(&pdev->dev);
node              476 drivers/crypto/cavium/nitrox/nitrox_main.c 	if (ndev->node == NUMA_NO_NODE)
node              477 drivers/crypto/cavium/nitrox/nitrox_main.c 		ndev->node = 0;
node              100 drivers/crypto/cavium/zip/zip_main.c struct zip_device *zip_get_device(int node)
node              102 drivers/crypto/cavium/zip/zip_main.c 	if ((node < MAX_ZIP_DEVICES) && (node >= 0))
node              103 drivers/crypto/cavium/zip/zip_main.c 		return zip_dev[node];
node              105 drivers/crypto/cavium/zip/zip_main.c 	zip_err("ZIP device not found for node id %d\n", node);
node              391 drivers/crypto/hisilicon/sec/sec_drv.h 	u8 node[16];
node               72 drivers/crypto/hisilicon/zip/zip.h struct hisi_zip *find_zip_device(int node);
node               91 drivers/crypto/hisilicon/zip/zip_main.c static struct hisi_zip *find_zip_device_numa(int node)
node              100 drivers/crypto/hisilicon/zip/zip_main.c 		if (node_distance(dev->numa_node, node) < min_distance) {
node              102 drivers/crypto/hisilicon/zip/zip_main.c 			min_distance = node_distance(dev->numa_node, node);
node              110 drivers/crypto/hisilicon/zip/zip_main.c struct hisi_zip *find_zip_device(int node)
node              116 drivers/crypto/hisilicon/zip/zip_main.c 	zip = find_zip_device_numa(node);
node             1725 drivers/crypto/n2_core.c 			       u64 node, struct spu_queue *p,
node             1730 drivers/crypto/n2_core.c 	mdesc_for_each_arc(arc, mdesc, node, MDESC_ARC_TYPE_BACK) {
node             1752 drivers/crypto/n2_core.c 			    u64 node, const char *iname, unsigned long q_type,
node             1771 drivers/crypto/n2_core.c 	err = spu_mdesc_walk_arcs(mdesc, dev, node, p, table);
node             1788 drivers/crypto/n2_core.c 	u64 node;
node             1790 drivers/crypto/n2_core.c 	mdesc_for_each_node_by_name(mdesc, node, "exec-unit") {
node             1793 drivers/crypto/n2_core.c 		type = mdesc_get_property(mdesc, node, "type", NULL);
node             1797 drivers/crypto/n2_core.c 		err = handle_exec_unit(ip, list, dev, mdesc, node,
node             1808 drivers/crypto/n2_core.c static int get_irq_props(struct mdesc_handle *mdesc, u64 node,
node             1815 drivers/crypto/n2_core.c 	ino = mdesc_get_property(mdesc, node, "ino", &ino_len);
node             1843 drivers/crypto/n2_core.c 	u64 node;
node             1849 drivers/crypto/n2_core.c 	mdesc_for_each_node_by_name(mdesc, node, "virtual-device") {
node             1853 drivers/crypto/n2_core.c 		name = mdesc_get_property(mdesc, node, "name", NULL);
node             1856 drivers/crypto/n2_core.c 		chdl = mdesc_get_property(mdesc, node, "cfg-handle", NULL);
node             1860 drivers/crypto/n2_core.c 		return get_irq_props(mdesc, node, ip);
node              828 drivers/crypto/nx/nx-842-pseries.c 	struct device_node *node = NULL;
node              833 drivers/crypto/nx/nx-842-pseries.c 		node = local_devdata->dev->of_node;
node              837 drivers/crypto/nx/nx-842-pseries.c 			!strcmp(upd->dn->name, node->name)) {
node              984 drivers/crypto/omap-aes.c 	struct device_node *node = dev->of_node;
node              994 drivers/crypto/omap-aes.c 	err = of_address_to_resource(node, 0, res);
node             1934 drivers/crypto/omap-sham.c 	struct device_node *node = dev->of_node;
node             1944 drivers/crypto/omap-sham.c 	err = of_address_to_resource(node, 0, res);
node             1951 drivers/crypto/omap-sham.c 	dd->irq = irq_of_parse_and_map(node, 0);
node              164 drivers/crypto/qat/qat_common/adf_common_drv.h struct qat_crypto_instance *qat_crypto_get_instance_node(int node);
node              617 drivers/crypto/qat/qat_common/qat_algs.c 	int node = get_current_node();
node              621 drivers/crypto/qat/qat_common/qat_algs.c 	inst = qat_crypto_get_instance_node(node);
node              970 drivers/crypto/qat/qat_common/qat_algs.c 	int node = get_current_node();
node              973 drivers/crypto/qat/qat_common/qat_algs.c 	inst = qat_crypto_get_instance_node(node);
node               95 drivers/crypto/qat/qat_common/qat_crypto.c struct qat_crypto_instance *qat_crypto_get_instance_node(int node)
node              104 drivers/crypto/qat/qat_common/qat_crypto.c 		if ((node == dev_to_node(&GET_DEV(tmp_dev)) ||
node              117 drivers/crypto/qat/qat_common/qat_crypto.c 		pr_info("QAT: Could not find a device on node %d\n", node);
node              314 drivers/crypto/virtio/virtio_crypto_algs.c 		int node = virtio_crypto_get_current_node();
node              316 drivers/crypto/virtio/virtio_crypto_algs.c 				      virtcrypto_get_dev_node(node,
node              110 drivers/crypto/virtio/virtio_crypto_common.h struct virtio_crypto *virtcrypto_get_dev_node(int node,
node              123 drivers/crypto/virtio/virtio_crypto_common.h 	int cpu, node;
node              126 drivers/crypto/virtio/virtio_crypto_common.h 	node = topology_physical_package_id(cpu);
node              129 drivers/crypto/virtio/virtio_crypto_common.h 	return node;
node              184 drivers/crypto/virtio/virtio_crypto_mgr.c struct virtio_crypto *virtcrypto_get_dev_node(int node, uint32_t service,
node              194 drivers/crypto/virtio/virtio_crypto_mgr.c 		if ((node == dev_to_node(&tmp_dev->vdev->dev) ||
node              208 drivers/crypto/virtio/virtio_crypto_mgr.c 				node);
node               58 drivers/dca/dca-core.c 	list_del(&domain->node);
node               97 drivers/dca/dca-core.c 	domain = list_first_entry(&dca_domains, struct dca_domain, node);
node               99 drivers/dca/dca-core.c 	list_for_each_entry_safe(dca, _dca, &domain->dca_providers, node)
node              100 drivers/dca/dca-core.c 		list_move(&dca->node, &unregistered_providers);
node              106 drivers/dca/dca-core.c 	list_for_each_entry_safe(dca, _dca, &unregistered_providers, node) {
node              108 drivers/dca/dca-core.c 		list_del(&dca->node);
node              116 drivers/dca/dca-core.c 	list_for_each_entry(domain, &dca_domains, node)
node              154 drivers/dca/dca-core.c 						  node);
node              159 drivers/dca/dca-core.c 	list_for_each_entry(dca, &domain->dca_providers, node)
node              197 drivers/dca/dca-core.c 	list_for_each_entry(dca, &domain->dca_providers, node) {
node              380 drivers/dca/dca-core.c 			list_add(&domain->node, &dca_domains);
node              383 drivers/dca/dca-core.c 	list_add(&dca->node, &domain->dca_providers);
node              413 drivers/dca/dca-core.c 	list_del(&dca->node);
node              223 drivers/devfreq/devfreq-event.c 	struct device_node *node;
node              229 drivers/devfreq/devfreq-event.c 	node = of_parse_phandle(dev->of_node, "devfreq-events", index);
node              230 drivers/devfreq/devfreq-event.c 	if (!node)
node              234 drivers/devfreq/devfreq-event.c 	list_for_each_entry(edev, &devfreq_event_list, node) {
node              235 drivers/devfreq/devfreq-event.c 		if (edev->dev.parent && edev->dev.parent->of_node == node)
node              239 drivers/devfreq/devfreq-event.c 	list_for_each_entry(edev, &devfreq_event_list, node) {
node              240 drivers/devfreq/devfreq-event.c 		if (of_node_name_eq(node, edev->desc->name))
node              248 drivers/devfreq/devfreq-event.c 		of_node_put(node);
node              252 drivers/devfreq/devfreq-event.c 	of_node_put(node);
node              337 drivers/devfreq/devfreq-event.c 	INIT_LIST_HEAD(&edev->node);
node              340 drivers/devfreq/devfreq-event.c 	list_add(&edev->node, &devfreq_event_list);
node              361 drivers/devfreq/devfreq-event.c 	list_del(&edev->node);
node               65 drivers/devfreq/devfreq.c 	list_for_each_entry(tmp_devfreq, &devfreq_list, node) {
node              215 drivers/devfreq/devfreq.c 	list_for_each_entry(tmp_governor, &devfreq_governor_list, node) {
node              591 drivers/devfreq/devfreq.c 	list_del(&devfreq->node);
node              644 drivers/devfreq/devfreq.c 	INIT_LIST_HEAD(&devfreq->node);
node              733 drivers/devfreq/devfreq.c 	list_add(&devfreq->node, &devfreq_list);
node              832 drivers/devfreq/devfreq.c 	struct device_node *node;
node              841 drivers/devfreq/devfreq.c 	node = of_parse_phandle(dev->of_node, "devfreq", index);
node              842 drivers/devfreq/devfreq.c 	if (!node)
node              846 drivers/devfreq/devfreq.c 	list_for_each_entry(devfreq, &devfreq_list, node) {
node              848 drivers/devfreq/devfreq.c 			&& devfreq->dev.parent->of_node == node) {
node              850 drivers/devfreq/devfreq.c 			of_node_put(node);
node              855 drivers/devfreq/devfreq.c 	of_node_put(node);
node              967 drivers/devfreq/devfreq.c 	list_for_each_entry(devfreq, &devfreq_list, node) {
node              988 drivers/devfreq/devfreq.c 	list_for_each_entry(devfreq, &devfreq_list, node) {
node             1021 drivers/devfreq/devfreq.c 	list_add(&governor->node, &devfreq_governor_list);
node             1023 drivers/devfreq/devfreq.c 	list_for_each_entry(devfreq, &devfreq_list, node) {
node             1085 drivers/devfreq/devfreq.c 	list_for_each_entry(devfreq, &devfreq_list, node) {
node             1109 drivers/devfreq/devfreq.c 	list_del(&governor->node);
node             1219 drivers/devfreq/devfreq.c 		list_for_each_entry(governor, &devfreq_governor_list, node) {
node              503 drivers/devfreq/event/exynos-ppmu.c 	struct device_node *events_np, *node;
node              528 drivers/devfreq/event/exynos-ppmu.c 	for_each_child_of_node(events_np, node) {
node              533 drivers/devfreq/event/exynos-ppmu.c 			if (of_node_name_eq(node, ppmu_events[i].name))
node              540 drivers/devfreq/event/exynos-ppmu.c 				node);
node              555 drivers/devfreq/event/exynos-ppmu.c 		of_property_read_string(node, "event-name", &desc[j].name);
node              556 drivers/devfreq/event/exynos-ppmu.c 		ret = of_property_read_u32(node, "event-data-type",
node              182 drivers/devfreq/event/rockchip-dfi.c 	struct device_node *np = pdev->dev.of_node, *node;
node              200 drivers/devfreq/event/rockchip-dfi.c 	node = of_parse_phandle(np, "rockchip,pmu", 0);
node              201 drivers/devfreq/event/rockchip-dfi.c 	if (node) {
node              202 drivers/devfreq/event/rockchip-dfi.c 		data->regmap_pmu = syscon_node_to_regmap(node);
node              293 drivers/devfreq/exynos-bus.c 	struct device_node *np = dev->of_node, *node;
node              319 drivers/devfreq/exynos-bus.c 	node = of_parse_phandle(dev->of_node, "devfreq", 0);
node              320 drivers/devfreq/exynos-bus.c 	if (node) {
node              321 drivers/devfreq/exynos-bus.c 		of_node_put(node);
node               48 drivers/devfreq/governor.h 	struct list_head node;
node              310 drivers/devfreq/rk3399_dmc.c 	struct device_node *np = pdev->dev.of_node, *node;
node              372 drivers/devfreq/rk3399_dmc.c 	node = of_parse_phandle(np, "rockchip,pmu", 0);
node              373 drivers/devfreq/rk3399_dmc.c 	if (node) {
node              374 drivers/devfreq/rk3399_dmc.c 		data->regmap_pmu = syscon_node_to_regmap(node);
node              690 drivers/dma-buf/dma-buf.c 	list_add(&attach->node, &dmabuf->attachments);
node              720 drivers/dma-buf/dma-buf.c 	list_del(&attach->node);
node             1224 drivers/dma-buf/dma-buf.c 		list_for_each_entry(attach_obj, &buf_obj->attachments, node) {
node              147 drivers/dma-buf/dma-fence.c 	list_for_each_entry_safe(cur, tmp, &cb_list, node) {
node              148 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cur->node);
node              340 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cb->node);
node              362 drivers/dma-buf/dma-fence.c 		list_add_tail(&cb->node, &fence->cb_list);
node              364 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cb->node);
node              422 drivers/dma-buf/dma-fence.c 	ret = !list_empty(&cb->node);
node              424 drivers/dma-buf/dma-fence.c 		list_del_init(&cb->node);
node              498 drivers/dma-buf/dma-fence.c 	list_add(&cb.base.node, &fence->cb_list);
node              514 drivers/dma-buf/dma-fence.c 	if (!list_empty(&cb.base.node))
node              515 drivers/dma-buf/dma-fence.c 		list_del(&cb.base.node);
node              140 drivers/dma-buf/sw_sync.c 		rb_erase(&pt->node, &parent->pt_tree);
node              207 drivers/dma-buf/sw_sync.c 		rb_erase(&pt->node, &obj->pt_tree);
node              257 drivers/dma-buf/sw_sync.c 			other = rb_entry(parent, typeof(*pt), node);
node              273 drivers/dma-buf/sw_sync.c 		rb_link_node(&pt->node, parent, p);
node              274 drivers/dma-buf/sw_sync.c 		rb_insert_color(&pt->node, &obj->pt_tree);
node              276 drivers/dma-buf/sw_sync.c 		parent = rb_next(&pt->node);
node              278 drivers/dma-buf/sw_sync.c 			      parent ? &rb_entry(parent, typeof(*pt), node)->link : &obj->pt_list);
node               62 drivers/dma-buf/sync_debug.h 	struct rb_node node;
node               37 drivers/dma-buf/sync_file.c 	INIT_LIST_HEAD(&sync_file->cb.node);
node              313 drivers/dma-buf/sync_file.c 	if (list_empty(&sync_file->cb.node) &&
node               58 drivers/dma/acpi-dma.c 	list_for_each_entry(rentry, &resource_list, node) {
node              161 drivers/dma/altera-msgdma.c 	struct list_head node;
node              214 drivers/dma/altera-msgdma.c 	desc = list_first_entry(&mdev->free_list, struct msgdma_sw_desc, node);
node              215 drivers/dma/altera-msgdma.c 	list_del(&desc->node);
node              234 drivers/dma/altera-msgdma.c 	list_add_tail(&desc->node, &mdev->free_list);
node              235 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(child, next, &desc->tx_list, node) {
node              237 drivers/dma/altera-msgdma.c 		list_move_tail(&child->node, &mdev->free_list);
node              251 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(desc, next, list, node)
node              312 drivers/dma/altera-msgdma.c 	list_add_tail(&new->node, &mdev->pending_list);
node              364 drivers/dma/altera-msgdma.c 			list_add_tail(&new->node, &first->tx_list);
node              438 drivers/dma/altera-msgdma.c 			list_add_tail(&new->node, &first->tx_list);
node              538 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(sdesc, next, &desc->tx_list, node)
node              554 drivers/dma/altera-msgdma.c 					struct msgdma_sw_desc, node);
node              584 drivers/dma/altera-msgdma.c 	list_for_each_entry_safe(desc, next, &mdev->done_list, node) {
node              588 drivers/dma/altera-msgdma.c 		list_del(&desc->node);
node              612 drivers/dma/altera-msgdma.c 					struct msgdma_sw_desc, node);
node              615 drivers/dma/altera-msgdma.c 	list_del(&desc->node);
node              617 drivers/dma/altera-msgdma.c 	list_add_tail(&desc->node, &mdev->done_list);
node              671 drivers/dma/altera-msgdma.c 		list_add_tail(&desc->node, &mdev->free_list);
node              183 drivers/dma/amba-pl08x.c 	struct list_head node;
node              535 drivers/dma/amba-pl08x.c 	list_del(&txd->vd.node);
node             1284 drivers/dma/amba-pl08x.c 	list_for_each_entry(dsg, &txd->dsg_list, node) {
node             1500 drivers/dma/amba-pl08x.c 	list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) {
node             1501 drivers/dma/amba-pl08x.c 		list_del(&dsg->node);
node             1583 drivers/dma/amba-pl08x.c 			list_for_each_entry(dsg, &txd->dsg_list, node)
node             1911 drivers/dma/amba-pl08x.c 	list_add_tail(&dsg->node, &txd->dsg_list);
node             2035 drivers/dma/amba-pl08x.c 	list_add_tail(&dsg->node, &txd->dsg_list);
node              112 drivers/dma/bcm-sba-raid.c 	struct list_head node;
node              210 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_free_list, node) {
node              212 drivers/dma/bcm-sba-raid.c 			list_move_tail(&req->node, &sba->reqs_alloc_list);
node              248 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_pending_list);
node              264 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_active_list);
node              277 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_aborted_list);
node              289 drivers/dma/bcm-sba-raid.c 	list_move_tail(&req->node, &sba->reqs_free_list);
node              332 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_alloc_list, node)
node              336 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_active_list, node)
node              355 drivers/dma/bcm-sba-raid.c 	list_for_each_entry_safe(req, req1, &sba->reqs_pending_list, node)
node              398 drivers/dma/bcm-sba-raid.c 				       struct sba_request, node);
node              466 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_free_list, node)
node              470 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_alloc_list, node)
node              473 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_pending_list, node)
node              476 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_active_list, node)
node              479 drivers/dma/bcm-sba-raid.c 	list_for_each_entry(req, &sba->reqs_aborted_list, node)
node             1506 drivers/dma/bcm-sba-raid.c 		INIT_LIST_HEAD(&req->node);
node             1524 drivers/dma/bcm-sba-raid.c 		list_add_tail(&req->node, &sba->reqs_free_list);
node              450 drivers/dma/bcm2835-dma.c 	list_del(&vd->node);
node             1270 drivers/dma/coh901318.c 	struct list_head node;
node             1525 drivers/dma/coh901318.c 		INIT_LIST_HEAD(&desc->node);
node             1531 drivers/dma/coh901318.c 					node);
node             1532 drivers/dma/coh901318.c 		list_del(&desc->node);
node             1547 drivers/dma/coh901318.c 	list_add_tail(&cohd->node, &cohc->free);
node             1554 drivers/dma/coh901318.c 	list_add_tail(&desc->node, &cohc->active);
node             1561 drivers/dma/coh901318.c 					node);
node             1567 drivers/dma/coh901318.c 	list_del(&cohd->node);
node             1573 drivers/dma/coh901318.c 	list_add_tail(&desc->node, &cohc->queue);
node             1580 drivers/dma/coh901318.c 					node);
node             1624 drivers/dma/coh901318.c 		cohd = list_entry(pos, struct coh901318_desc, node);
node             1670 drivers/dma/coh901318.c 		cohd = list_entry(pos, struct coh901318_desc, node);
node              213 drivers/dma/dma-axi-dmac.c 		list_move_tail(&vdesc->node, &chan->active_descs);
node              269 drivers/dma/dma-axi-dmac.c 		struct axi_dmac_desc, vdesc.node);
node              294 drivers/dma/dma-axi-dmac.c 		list_for_each_entry(desc, &chan->active_descs, vdesc.node) {
node              388 drivers/dma/dma-axi-dmac.c 				list_del(&active->vdesc.node);
node              490 drivers/dma/dma-jz4780.c 		list_del(&vdesc->node);
node              377 drivers/dma/dmaengine.c 	int node = dev_to_node(chan->device->dev);
node              378 drivers/dma/dmaengine.c 	return node == NUMA_NO_NODE ||
node              379 drivers/dma/dmaengine.c 		cpumask_test_cpu(cpu, cpumask_of_node(node));
node              213 drivers/dma/dmatest.c 	struct list_head	node;
node              227 drivers/dma/dmatest.c 	struct list_head	node;
node              239 drivers/dma/dmatest.c 	list_for_each_entry(dtc, &info->channels, node) {
node              242 drivers/dma/dmatest.c 		list_for_each_entry(thread, &dtc->threads, node) {
node              255 drivers/dma/dmatest.c 	list_for_each_entry(dtc, &info->channels, node) {
node              258 drivers/dma/dmatest.c 		list_for_each_entry(thread, &dtc->threads, node) {
node              919 drivers/dma/dmatest.c 	list_for_each_entry_safe(thread, _thread, &dtc->threads, node) {
node              923 drivers/dma/dmatest.c 		list_del(&thread->node);
node              978 drivers/dma/dmatest.c 		list_add_tail(&thread->node, &dtc->threads);
node             1028 drivers/dma/dmatest.c 	list_add_tail(&dtc->node, &info->channels);
node             1101 drivers/dma/dmatest.c 	list_for_each_entry(dtc, &info->channels, node) {
node             1105 drivers/dma/dmatest.c 		list_for_each_entry(thread, &dtc->threads, node) {
node             1119 drivers/dma/dmatest.c 	list_for_each_entry_safe(dtc, _dtc, &info->channels, node) {
node             1120 drivers/dma/dmatest.c 		list_del(&dtc->node);
node             1201 drivers/dma/dmatest.c 		list_for_each_entry(dtc, &info->channels, node) {
node             1206 drivers/dma/dmatest.c 						      node);
node             1219 drivers/dma/dmatest.c 	dtc = list_last_entry(&info->channels, struct dmatest_chan, node);
node             1274 drivers/dma/dmatest.c 	list_for_each_entry(dtc, &info->channels, node) {
node             1278 drivers/dma/dmatest.c 		list_for_each_entry(thread, &dtc->threads, node) {
node              555 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	list_del(&vd->node);
node              586 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	list_del(&vd->node);
node              489 drivers/dma/dw-edma/dw-edma-core.c 				list_del(&vd->node);
node              496 drivers/dma/dw-edma/dw-edma-core.c 			list_del(&vd->node);
node              524 drivers/dma/dw-edma/dw-edma-core.c 		list_del(&vd->node);
node              129 drivers/dma/ep93xx_dma.c 	struct list_head		node;
node              243 drivers/dma/ep93xx_dma.c 	list_add_tail(&desc->node, &edmac->active);
node              248 drivers/dma/ep93xx_dma.c 			struct ep93xx_dma_desc, node);
node              259 drivers/dma/ep93xx_dma.c 		list_move_tail(&d->node, &edmac->active);
node              268 drivers/dma/ep93xx_dma.c 					struct ep93xx_dma_desc, node);
node              682 drivers/dma/ep93xx_dma.c 	list_for_each_entry_safe(desc, _desc, &edmac->free_list, node) {
node              684 drivers/dma/ep93xx_dma.c 			list_del_init(&desc->node);
node              711 drivers/dma/ep93xx_dma.c 		list_add(&desc->node, &edmac->free_list);
node              736 drivers/dma/ep93xx_dma.c 	new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node);
node              737 drivers/dma/ep93xx_dma.c 	list_del_init(&new->node);
node              776 drivers/dma/ep93xx_dma.c 	list_for_each_entry_safe(desc, d, &list, node) {
node              850 drivers/dma/ep93xx_dma.c 		list_add_tail(&desc->node, &edmac->queue);
node              967 drivers/dma/ep93xx_dma.c 	list_for_each_entry_safe(desc, d, &list, node)
node             1009 drivers/dma/ep93xx_dma.c 			list_add_tail(&desc->node, &first->tx_list);
node             1084 drivers/dma/ep93xx_dma.c 			list_add_tail(&desc->node, &first->tx_list);
node             1165 drivers/dma/ep93xx_dma.c 			list_add_tail(&desc->node, &first->tx_list);
node             1224 drivers/dma/ep93xx_dma.c 	list_for_each_entry_safe(desc, _d, &list, node)
node               49 drivers/dma/fsl-edma.c 				list_del(&fsl_chan->edesc->vdesc.node);
node              951 drivers/dma/fsl-qdma.c 	list_del(&vdesc->node);
node               98 drivers/dma/fsl_raid.c 	list_add_tail(&desc->node, &re_chan->submit_q);
node              118 drivers/dma/fsl_raid.c 	list_for_each_entry_safe(desc, _desc, &re_chan->submit_q, node) {
node              122 drivers/dma/fsl_raid.c 		list_move_tail(&desc->node, &re_chan->active_q);
node              148 drivers/dma/fsl_raid.c 	list_for_each_entry_safe(desc, _desc, &re_chan->ack_q, node) {
node              150 drivers/dma/fsl_raid.c 			list_move_tail(&desc->node, &re_chan->free_q);
node              176 drivers/dma/fsl_raid.c 					 node) {
node              187 drivers/dma/fsl_raid.c 			list_move_tail(&desc->node, &re_chan->ack_q);
node              258 drivers/dma/fsl_raid.c 	INIT_LIST_HEAD(&desc->node);
node              286 drivers/dma/fsl_raid.c 					struct fsl_re_desc, node);
node              287 drivers/dma/fsl_raid.c 		list_del(&desc->node);
node              594 drivers/dma/fsl_raid.c 		INIT_LIST_HEAD(&desc->node);
node              597 drivers/dma/fsl_raid.c 		list_add_tail(&desc->node, &re_chan->free_q);
node              612 drivers/dma/fsl_raid.c 					node);
node              614 drivers/dma/fsl_raid.c 		list_del(&desc->node);
node              295 drivers/dma/fsl_raid.h 	struct list_head node;
node              427 drivers/dma/fsldma.c 	list_for_each_entry(child, &desc->tx_list, node) {
node              447 drivers/dma/fsldma.c 	list_del(&desc->node);
node              492 drivers/dma/fsldma.c 	list_for_each_entry_safe(desc, _desc, &chan->ld_completed, node)
node              541 drivers/dma/fsldma.c 	list_del(&desc->node);
node              552 drivers/dma/fsldma.c 		list_add_tail(&desc->node, &chan->ld_completed);
node              599 drivers/dma/fsldma.c 	desc = list_first_entry(&chan->ld_pending, struct fsl_desc_sw, node);
node              645 drivers/dma/fsldma.c 	list_for_each_entry_safe(desc, _desc, &chan->ld_running, node) {
node              725 drivers/dma/fsldma.c 	list_for_each_entry_safe(desc, _desc, list, node)
node              734 drivers/dma/fsldma.c 	list_for_each_entry_safe_reverse(desc, _desc, list, node)
node              804 drivers/dma/fsldma.c 		list_add_tail(&new->node, &first->tx_list);
node             1106 drivers/dma/fsldma.c 	struct device_node *node, u32 feature, const char *compatible)
node             1120 drivers/dma/fsldma.c 	chan->regs = of_iomap(node, 0);
node             1127 drivers/dma/fsldma.c 	err = of_address_to_resource(node, 0, &res);
node             1187 drivers/dma/fsldma.c 	chan->irq = irq_of_parse_and_map(node, 0);
node              102 drivers/dma/fsldma.h 	struct list_head node;
node              191 drivers/dma/fsldma.h #define to_fsl_desc(lh) container_of(lh, struct fsl_desc_sw, node)
node              119 drivers/dma/hsu/hsu.c 	list_del(&vdesc->node);
node              124 drivers/dma/idma64.c 	list_del(&vdesc->node);
node              535 drivers/dma/img-mdc-dma.c 	list_del(&vd->node);
node              121 drivers/dma/imx-dma.c 	struct list_head		node;
node              248 drivers/dma/imx-dma.c 					node);
node              442 drivers/dma/imx-dma.c 				node);
node              630 drivers/dma/imx-dma.c 	desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, node);
node              651 drivers/dma/imx-dma.c 					     struct imxdma_desc, node);
node              780 drivers/dma/imx-dma.c 		list_add_tail(&desc->node, &imxdmac->ld_free);
node              805 drivers/dma/imx-dma.c 	list_for_each_entry_safe(desc, _desc, &imxdmac->ld_free, node) {
node              829 drivers/dma/imx-dma.c 	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node);
node              884 drivers/dma/imx-dma.c 	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node);
node              940 drivers/dma/imx-dma.c 	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node);
node              977 drivers/dma/imx-dma.c 	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node);
node             1010 drivers/dma/imx-dma.c 					struct imxdma_desc, node);
node              768 drivers/dma/imx-sdma.c 		list_del(&vd->node);
node               84 drivers/dma/k3dma.c 	struct list_head	node;
node              287 drivers/dma/k3dma.c 		list_del(&ds->vd.node);
node              333 drivers/dma/k3dma.c 				struct k3_dma_chan, node);
node              335 drivers/dma/k3dma.c 			list_del_init(&c->node);
node              368 drivers/dma/k3dma.c 	list_del_init(&c->node);
node              434 drivers/dma/k3dma.c 			if (list_empty(&c->node)) {
node              436 drivers/dma/k3dma.c 				list_add_tail(&c->node, &d->chan_pending);
node              733 drivers/dma/k3dma.c 	list_del_init(&c->node);
node              776 drivers/dma/k3dma.c 			list_del_init(&c->node);
node              799 drivers/dma/k3dma.c 			list_add_tail(&c->node, &d->chan_pending);
node              945 drivers/dma/k3dma.c 		INIT_LIST_HEAD(&c->node);
node               39 drivers/dma/mcf-edma.c 				list_del(&mcf_chan->edesc->vdesc.node);
node               87 drivers/dma/mediatek/mtk-cqdma.c 	struct list_head node;
node              276 drivers/dma/mediatek/mtk-cqdma.c 	list_for_each_entry_safe(vd, vd2, &cvc->vc.desc_issued, node) {
node              284 drivers/dma/mediatek/mtk-cqdma.c 		list_add_tail(&cvd->node, &pc->queue);
node              291 drivers/dma/mediatek/mtk-cqdma.c 		list_del(&vd->node);
node              303 drivers/dma/mediatek/mtk-cqdma.c 	list_for_each_entry(cvd, &cvc->pc->queue, node)
node              321 drivers/dma/mediatek/mtk-cqdma.c 				       struct mtk_cqdma_vdesc, node);
node              332 drivers/dma/mediatek/mtk-cqdma.c 	list_del(&cvd->node);
node              352 drivers/dma/mediatek/mtk-cqdma.c 				       struct mtk_cqdma_vdesc, node);
node              427 drivers/dma/mediatek/mtk-cqdma.c 	list_for_each_entry(vd, &cvc->pc->queue, node)
node              434 drivers/dma/mediatek/mtk-cqdma.c 	list_for_each_entry(vd, &cvc->vc.desc_issued, node)
node              503 drivers/dma/mediatek/mtk-hsdma.c 	list_for_each_entry_safe(vd, vd2, &hvc->vc.desc_issued, node) {
node              529 drivers/dma/mediatek/mtk-hsdma.c 		list_move_tail(&vd->node, &hvc->desc_hw_processing);
node              589 drivers/dma/mediatek/mtk-hsdma.c 			list_del(&cb->vd->node);
node              666 drivers/dma/mediatek/mtk-hsdma.c 	list_for_each_entry(vd, &hvc->desc_hw_processing, node)
node              670 drivers/dma/mediatek/mtk-hsdma.c 	list_for_each_entry(vd, &hvc->vc.desc_issued, node)
node              216 drivers/dma/mediatek/mtk-uart-apdma.c 	list_del(&d->vd.node);
node              249 drivers/dma/mediatek/mtk-uart-apdma.c 	list_del(&d->vd.node);
node               83 drivers/dma/mmp_pdma.c 	struct list_head node;
node              135 drivers/dma/mmp_pdma.c 	container_of(lh, struct mmp_pdma_desc_sw, node)
node              327 drivers/dma/mmp_pdma.c 				struct mmp_pdma_desc_sw, node);
node              351 drivers/dma/mmp_pdma.c 	list_for_each_entry(child, &desc->tx_list, node) {
node              420 drivers/dma/mmp_pdma.c 	list_for_each_entry_safe(desc, _desc, list, node) {
node              421 drivers/dma/mmp_pdma.c 		list_del(&desc->node);
node              505 drivers/dma/mmp_pdma.c 		list_add_tail(&new->node, &first->tx_list);
node              579 drivers/dma/mmp_pdma.c 			list_add_tail(&new->node, &first->tx_list);
node              674 drivers/dma/mmp_pdma.c 		list_add_tail(&new->node, &first->tx_list);
node              789 drivers/dma/mmp_pdma.c 	list_for_each_entry(sw, &chan->chain_running, node) {
node              898 drivers/dma/mmp_pdma.c 	list_for_each_entry_safe(desc, _desc, &chan->chain_running, node) {
node              903 drivers/dma/mmp_pdma.c 		list_move(&desc->node, &chain_cleanup);
node              929 drivers/dma/mmp_pdma.c 	list_for_each_entry_safe(desc, _desc, &chain_cleanup, node) {
node              933 drivers/dma/mmp_pdma.c 		list_del(&desc->node);
node              431 drivers/dma/moxart-dma.c 	list_del(&vd->node);
node              569 drivers/dma/moxart-dma.c 	struct device_node *node = dev->of_node;
node              581 drivers/dma/moxart-dma.c 	irq = irq_of_parse_and_map(node, 0);
node              627 drivers/dma/moxart-dma.c 	ret = of_dma_controller_register(node, moxart_of_xlate, mdc);
node              189 drivers/dma/mpc512x_dma.c 	struct list_head		node;
node              264 drivers/dma/mpc512x_dma.c 						struct mpc_dma_desc, node);
node              273 drivers/dma/mpc512x_dma.c 				list_move_tail(&mdesc->node, &mchan->active);
node              276 drivers/dma/mpc512x_dma.c 			list_move_tail(&mdesc->node, &mchan->active);
node              281 drivers/dma/mpc512x_dma.c 	list_for_each_entry(mdesc, &mchan->active, node) {
node              336 drivers/dma/mpc512x_dma.c 			list_for_each_entry(mdesc, &mchan->active, node)
node              399 drivers/dma/mpc512x_dma.c 		list_for_each_entry(mdesc, &list, node) {
node              472 drivers/dma/mpc512x_dma.c 	list_move_tail(&mdesc->node, &mchan->queued);
node              520 drivers/dma/mpc512x_dma.c 		list_add_tail(&mdesc->node, &descs);
node              575 drivers/dma/mpc512x_dma.c 	list_for_each_entry_safe(mdesc, tmp, &descs, node)
node              614 drivers/dma/mpc512x_dma.c 									node);
node              615 drivers/dma/mpc512x_dma.c 		list_del(&mdesc->node);
node              668 drivers/dma/mpc512x_dma.c 	list_add_tail(&mdesc->node, &mchan->prepared);
node              710 drivers/dma/mpc512x_dma.c 						struct mpc_dma_desc, node);
node              718 drivers/dma/mpc512x_dma.c 		list_del(&mdesc->node);
node              795 drivers/dma/mpc512x_dma.c 		list_add_tail(&mdesc->node, &mchan->prepared);
node              804 drivers/dma/mpc512x_dma.c 	list_add_tail(&mdesc->node, &mchan->free);
node              221 drivers/dma/mv_xor.c 				 node) {
node              224 drivers/dma/mv_xor.c 			list_move_tail(&iter->node, &mv_chan->free_slots);
node              246 drivers/dma/mv_xor.c 		list_move_tail(&desc->node, &mv_chan->completed_slots);
node              252 drivers/dma/mv_xor.c 		list_move_tail(&desc->node, &mv_chan->free_slots);
node              281 drivers/dma/mv_xor.c 				 node) {
node              313 drivers/dma/mv_xor.c 					  node);
node              316 drivers/dma/mv_xor.c 			if (!list_is_last(&iter->node, &mv_chan->chain)) {
node              321 drivers/dma/mv_xor.c 				iter = list_entry(iter->node.next,
node              323 drivers/dma/mv_xor.c 						  node);
node              358 drivers/dma/mv_xor.c 					node);
node              360 drivers/dma/mv_xor.c 		list_move_tail(&iter->node, &mv_chan->allocated_slots);
node              398 drivers/dma/mv_xor.c 		list_move_tail(&sw_desc->node, &mv_chan->chain);
node              404 drivers/dma/mv_xor.c 					    node);
node              405 drivers/dma/mv_xor.c 		list_move_tail(&sw_desc->node, &mv_chan->chain);
node              458 drivers/dma/mv_xor.c 		INIT_LIST_HEAD(&slot->node);
node              466 drivers/dma/mv_xor.c 		list_add_tail(&slot->node, &mv_chan->free_slots);
node              638 drivers/dma/mv_xor.c 					node) {
node              640 drivers/dma/mv_xor.c 		list_move_tail(&iter->node, &mv_chan->free_slots);
node              643 drivers/dma/mv_xor.c 				 node) {
node              645 drivers/dma/mv_xor.c 		list_move_tail(&iter->node, &mv_chan->free_slots);
node              648 drivers/dma/mv_xor.c 				 node) {
node              650 drivers/dma/mv_xor.c 		list_move_tail(&iter->node, &mv_chan->free_slots);
node              653 drivers/dma/mv_xor.c 		iter, _iter, &mv_chan->free_slots, node) {
node              654 drivers/dma/mv_xor.c 		list_del(&iter->node);
node              142 drivers/dma/mv_xor.h 	struct list_head	node;
node              139 drivers/dma/nbpfaxi.c 	struct list_head node;
node              156 drivers/dma/nbpfaxi.c 	struct list_head node;
node              168 drivers/dma/nbpfaxi.c 	struct list_head node;
node              384 drivers/dma/nbpfaxi.c 	struct nbpf_link_desc *ldesc = list_first_entry(&desc->sg, struct nbpf_link_desc, node);
node              610 drivers/dma/nbpfaxi.c 						struct nbpf_desc, node);
node              640 drivers/dma/nbpfaxi.c 			list_for_each_entry(desc, &chan->active, node)
node              647 drivers/dma/nbpfaxi.c 				list_for_each_entry(desc, &chan->queued, node)
node              675 drivers/dma/nbpfaxi.c 	list_add_tail(&desc->node, &chan->queued);
node              705 drivers/dma/nbpfaxi.c 		list_add_tail(&ldesc->node, &lhead);
node              720 drivers/dma/nbpfaxi.c 		list_add_tail(&desc->node, &head);
node              730 drivers/dma/nbpfaxi.c 	list_add(&dpage->node, &chan->desc_page);
node              743 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(ldesc, tmp, &desc->sg, node)
node              744 drivers/dma/nbpfaxi.c 		list_move(&ldesc->node, &chan->free_links);
node              746 drivers/dma/nbpfaxi.c 	list_add(&desc->node, &chan->free);
node              757 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(desc, tmp, &chan->done, node)
node              759 drivers/dma/nbpfaxi.c 			list_move(&desc->node, &head);
node              764 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(desc, tmp, &head, node) {
node              765 drivers/dma/nbpfaxi.c 		list_del(&desc->node);
node              797 drivers/dma/nbpfaxi.c 		desc = list_first_entry(&chan->free, struct nbpf_desc, node);
node              798 drivers/dma/nbpfaxi.c 		list_del(&desc->node);
node              814 drivers/dma/nbpfaxi.c 						 struct nbpf_link_desc, node);
node              820 drivers/dma/nbpfaxi.c 			list_move_tail(&ldesc->node, &desc->sg);
node              849 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(desc, tmp, &head, node) {
node              852 drivers/dma/nbpfaxi.c 		list_del(&desc->node);
node              957 drivers/dma/nbpfaxi.c 	list_for_each_entry(ldesc, &desc->sg, node) {
node             1068 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(dpage, tmp, &chan->desc_page, node) {
node             1071 drivers/dma/nbpfaxi.c 		list_del(&dpage->node);
node             1120 drivers/dma/nbpfaxi.c 		list_for_each_entry_safe(desc, tmp, &chan->done, node) {
node             1130 drivers/dma/nbpfaxi.c 				list_del(&desc->node);
node             1154 drivers/dma/nbpfaxi.c 			list_del(&desc->node);
node             1198 drivers/dma/nbpfaxi.c 	list_move_tail(&desc->node, &chan->done);
node             1203 drivers/dma/nbpfaxi.c 					struct nbpf_desc, node);
node              158 drivers/dma/owl-dma.c 	struct list_head	node;
node              324 drivers/dma/owl-dma.c 	list_del(&lli->node);
node              337 drivers/dma/owl-dma.c 	INIT_LIST_HEAD(&lli->node);
node              349 drivers/dma/owl-dma.c 		list_add_tail(&next->node, &txd->lli_list);
node              505 drivers/dma/owl-dma.c 	list_del(&vd->node);
node              514 drivers/dma/owl-dma.c 			       struct owl_dma_lli, node);
node              643 drivers/dma/owl-dma.c 	list_for_each_entry_safe(lli, _lli, &txd->lli_list, node)
node              750 drivers/dma/owl-dma.c 		list_for_each_entry(lli, &txd->lli_list, node) {
node              753 drivers/dma/owl-dma.c 				list_for_each_entry(lli, &txd->lli_list, node)
node              784 drivers/dma/owl-dma.c 		list_for_each_entry(lli, &txd->lli_list, node)
node              517 drivers/dma/pl330.c 	struct list_head node;
node             2035 drivers/dma/pl330.c 	list_for_each_entry(desc, &pch->work_list, node) {
node             2067 drivers/dma/pl330.c 	list_for_each_entry_safe(desc, _dt, &pch->work_list, node)
node             2071 drivers/dma/pl330.c 			list_move_tail(&desc->node, &pch->completed_list);
node             2094 drivers/dma/pl330.c 					struct dma_pl330_desc, node);
node             2100 drivers/dma/pl330.c 			list_move_tail(&desc->node, &pch->work_list);
node             2110 drivers/dma/pl330.c 			list_move_tail(&desc->node, &pch->dmac->desc_pool);
node             2292 drivers/dma/pl330.c 	list_for_each_entry(desc, &pch->submitted_list, node) {
node             2297 drivers/dma/pl330.c 	list_for_each_entry(desc, &pch->work_list , node) {
node             2418 drivers/dma/pl330.c 	list_for_each_entry(desc, &pch->work_list, node) {
node             2499 drivers/dma/pl330.c 	while (!list_empty(&last->node)) {
node             2500 drivers/dma/pl330.c 		desc = list_entry(last->node.next, struct dma_pl330_desc, node);
node             2509 drivers/dma/pl330.c 		list_move_tail(&desc->node, &pch->submitted_list);
node             2514 drivers/dma/pl330.c 	list_add_tail(&last->node, &pch->submitted_list);
node             2527 drivers/dma/pl330.c 	INIT_LIST_HEAD(&desc->node);
node             2546 drivers/dma/pl330.c 		list_add_tail(&desc[i].node, pool);
node             2564 drivers/dma/pl330.c 				struct dma_pl330_desc, node);
node             2566 drivers/dma/pl330.c 		list_del_init(&desc->node);
node             2701 drivers/dma/pl330.c 			while (!list_empty(&first->node)) {
node             2702 drivers/dma/pl330.c 				desc = list_entry(first->node.next,
node             2703 drivers/dma/pl330.c 						struct dma_pl330_desc, node);
node             2704 drivers/dma/pl330.c 				list_move_tail(&desc->node, &pl330->desc_pool);
node             2707 drivers/dma/pl330.c 			list_move_tail(&first->node, &pl330->desc_pool);
node             2740 drivers/dma/pl330.c 			list_add_tail(&desc->node, &first->node);
node             2817 drivers/dma/pl330.c 	while (!list_empty(&first->node)) {
node             2818 drivers/dma/pl330.c 		desc = list_entry(first->node.next,
node             2819 drivers/dma/pl330.c 				struct dma_pl330_desc, node);
node             2820 drivers/dma/pl330.c 		list_move_tail(&desc->node, &pl330->desc_pool);
node             2823 drivers/dma/pl330.c 	list_move_tail(&first->node, &pl330->desc_pool);
node             2865 drivers/dma/pl330.c 			list_add_tail(&desc->node, &first->node);
node               68 drivers/dma/ppc4xx/adma.c 	struct list_head node;
node             1401 drivers/dma/ppc4xx/adma.c 	list_for_each_entry(ref, &ppc440spe_adma_chan_list, node) {
node             4171 drivers/dma/ppc4xx/adma.c 		INIT_LIST_HEAD(&ref->node);
node             4172 drivers/dma/ppc4xx/adma.c 		list_add_tail(&ref->node, &ppc440spe_adma_chan_list);
node             4197 drivers/dma/ppc4xx/adma.c 	list_for_each_entry_safe(ref, _ref, &ppc440spe_adma_chan_list, node) {
node             4199 drivers/dma/ppc4xx/adma.c 			list_del(&ref->node);
node             4265 drivers/dma/ppc4xx/adma.c 					 node) {
node             4268 drivers/dma/ppc4xx/adma.c 				list_del(&ref->node);
node              575 drivers/dma/pxa_dma.c 					    struct virt_dma_desc, node);
node              620 drivers/dma/pxa_dma.c 	list_for_each_entry_safe(vd, tmp, &chan->vc.desc_issued, node) {
node              632 drivers/dma/pxa_dma.c 			list_del(&vd->node);
node              657 drivers/dma/pxa_dma.c 					      struct virt_dma_desc, node);
node              791 drivers/dma/pxa_dma.c 		list_move_tail(&vd->node, &vc->desc_issued);
node              803 drivers/dma/pxa_dma.c 					struct virt_dma_desc, node);
node              818 drivers/dma/pxa_dma.c 	list_move_tail(&vd->node, &vc->desc_submitted);
node              837 drivers/dma/pxa_dma.c 				    struct virt_dma_desc, node);
node              855 drivers/dma/pxa_dma.c 	INIT_LIST_HEAD(&vd->node);
node             1100 drivers/dma/pxa_dma.c 	list_for_each_entry(vd, &head, node) {
node              372 drivers/dma/qcom/bam_dma.c 	struct list_head node;
node              718 drivers/dma/qcom/bam_dma.c 		list_add(&async_desc->vd.node, &bchan->vc.desc_issued);
node              846 drivers/dma/qcom/bam_dma.c 				list_add(&async_desc->vd.node,
node             1000 drivers/dma/qcom/bam_dma.c 		list_del(&vd->node);
node              133 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(mdesc, next, &list, node) {
node              159 drivers/dma/qcom/hidma.c 		list_move(&mdesc->node, &mchan->free);
node              181 drivers/dma/qcom/hidma.c 	if (mdesc->node.next) {
node              183 drivers/dma/qcom/hidma.c 		list_move_tail(&mdesc->node, &mchan->completed);
node              188 drivers/dma/qcom/hidma.c 						  struct hidma_desc, node);
node              244 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(qdesc, next, &mchan->queued, node) {
node              246 drivers/dma/qcom/hidma.c 		list_move_tail(&qdesc->node, &mchan->active);
node              252 drivers/dma/qcom/hidma.c 							   node);
node              338 drivers/dma/qcom/hidma.c 	list_move_tail(&mdesc->node, &mchan->queued);
node              380 drivers/dma/qcom/hidma.c 		list_add_tail(&mdesc->node, &descs);
node              385 drivers/dma/qcom/hidma.c 		list_for_each_entry_safe(mdesc, tmp, &descs, node) {
node              411 drivers/dma/qcom/hidma.c 		mdesc = list_first_entry(&mchan->free, struct hidma_desc, node);
node              412 drivers/dma/qcom/hidma.c 		list_del(&mdesc->node);
node              426 drivers/dma/qcom/hidma.c 	list_add_tail(&mdesc->node, &mchan->prepared);
node              444 drivers/dma/qcom/hidma.c 		mdesc = list_first_entry(&mchan->free, struct hidma_desc, node);
node              445 drivers/dma/qcom/hidma.c 		list_del(&mdesc->node);
node              459 drivers/dma/qcom/hidma.c 	list_add_tail(&mdesc->node, &mchan->prepared);
node              494 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(mdesc, tmp, &list, node) {
node              502 drivers/dma/qcom/hidma.c 		list_move(&mdesc->node, &mchan->free);
node              547 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(mdesc, tmp, &descs, node) {
node              549 drivers/dma/qcom/hidma.c 		list_del(&mdesc->node);
node               80 drivers/dma/qcom/hidma.h 	struct list_head		node;
node               94 drivers/dma/qcom/hidma_dbg.c 	list_for_each_entry(mdesc, &mchan->prepared, node)
node               98 drivers/dma/qcom/hidma_dbg.c 	list_for_each_entry(mdesc, &mchan->active, node)
node              102 drivers/dma/qcom/hidma_dbg.c 	list_for_each_entry(mdesc, &mchan->completed, node)
node              152 drivers/dma/s3c24xx-dma.c 	struct list_head node;
node              431 drivers/dma/s3c24xx-dma.c 	list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) {
node              432 drivers/dma/s3c24xx-dma.c 		list_del(&dsg->node);
node              445 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_sg *dsg = list_entry(txd->at, struct s3c24xx_sg, node);
node              509 drivers/dma/s3c24xx-dma.c 	list_del(&txd->vd.node);
node              785 drivers/dma/s3c24xx-dma.c 		list_for_each_entry(dsg, &txd->dsg_list, node)
node              794 drivers/dma/s3c24xx-dma.c 		dsg = list_entry(txd->at, struct s3c24xx_sg, node);
node              795 drivers/dma/s3c24xx-dma.c 		list_for_each_entry_from(dsg, &txd->dsg_list, node)
node              842 drivers/dma/s3c24xx-dma.c 	list_add_tail(&dsg->node, &txd->dsg_list);
node              956 drivers/dma/s3c24xx-dma.c 		list_add_tail(&dsg->node, &txd->dsg_list);
node             1048 drivers/dma/s3c24xx-dma.c 		list_add_tail(&dsg->node, &txd->dsg_list);
node               94 drivers/dma/sa11x0-dma.c 	struct list_head	node;
node              150 drivers/dma/sa11x0-dma.c 	list_del(&txd->vd.node);
node              358 drivers/dma/sa11x0-dma.c 				struct sa11x0_dma_chan, node);
node              359 drivers/dma/sa11x0-dma.c 			list_del_init(&c->node);
node              395 drivers/dma/sa11x0-dma.c 	list_del_init(&c->node);
node              510 drivers/dma/sa11x0-dma.c 			if (list_empty(&c->node)) {
node              511 drivers/dma/sa11x0-dma.c 				list_add_tail(&c->node, &d->chan_pending);
node              717 drivers/dma/sa11x0-dma.c 			list_del_init(&c->node);
node              743 drivers/dma/sa11x0-dma.c 			list_add_tail(&c->node, &d->chan_pending);
node              776 drivers/dma/sa11x0-dma.c 				list_add_tail(&p->txd_load->vd.node, &head);
node              780 drivers/dma/sa11x0-dma.c 			list_add_tail(&p->txd_done->vd.node, &head);
node              862 drivers/dma/sa11x0-dma.c 		INIT_LIST_HEAD(&c->node);
node               35 drivers/dma/sh/rcar-dmac.c 	struct list_head node;
node               78 drivers/dma/sh/rcar-dmac.c 	struct list_head node;
node              103 drivers/dma/sh/rcar-dmac.c 	struct list_head node;
node              352 drivers/dma/sh/rcar-dmac.c 					 struct rcar_dmac_xfer_chunk, node);
node              474 drivers/dma/sh/rcar-dmac.c 	list_add_tail(&desc->node, &chan->desc.pending);
node              476 drivers/dma/sh/rcar-dmac.c 					 struct rcar_dmac_xfer_chunk, node);
node              510 drivers/dma/sh/rcar-dmac.c 		list_add_tail(&desc->node, &list);
node              515 drivers/dma/sh/rcar-dmac.c 	list_add_tail(&page->node, &chan->desc.pages);
node              540 drivers/dma/sh/rcar-dmac.c 	list_add(&desc->node, &chan->desc.free);
node              560 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry_safe(desc, _desc, &list, node) {
node              562 drivers/dma/sh/rcar-dmac.c 			list_del(&desc->node);
node              610 drivers/dma/sh/rcar-dmac.c 	desc = list_first_entry(&chan->desc.free, struct rcar_dmac_desc, node);
node              611 drivers/dma/sh/rcar-dmac.c 	list_del(&desc->node);
node              637 drivers/dma/sh/rcar-dmac.c 		list_add_tail(&chunk->node, &list);
node              642 drivers/dma/sh/rcar-dmac.c 	list_add_tail(&page->node, &chan->desc.pages);
node              681 drivers/dma/sh/rcar-dmac.c 				 struct rcar_dmac_xfer_chunk, node);
node              682 drivers/dma/sh/rcar-dmac.c 	list_del(&chunk->node);
node              733 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry(chunk, &desc->chunks, node) {
node              805 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry_safe(desc, _desc, &descs, node) {
node              806 drivers/dma/sh/rcar-dmac.c 		list_del(&desc->node);
node              991 drivers/dma/sh/rcar-dmac.c 			list_add_tail(&chunk->node, &desc->chunks);
node             1074 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry(desc, &list, node)
node             1077 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry_safe(page, _page, &rchan->desc.pages, node) {
node             1078 drivers/dma/sh/rcar-dmac.c 		list_del(&page->node);
node             1318 drivers/dma/sh/rcar-dmac.c 		list_for_each_entry(desc, &chan->desc.done, node) {
node             1322 drivers/dma/sh/rcar-dmac.c 		list_for_each_entry(desc, &chan->desc.pending, node) {
node             1326 drivers/dma/sh/rcar-dmac.c 		list_for_each_entry(desc, &chan->desc.active, node) {
node             1375 drivers/dma/sh/rcar-dmac.c 	list_for_each_entry_reverse(chunk, &desc->chunks, node) {
node             1437 drivers/dma/sh/rcar-dmac.c 					struct rcar_dmac_desc, node);
node             1505 drivers/dma/sh/rcar-dmac.c 		if (!list_is_last(&desc->running->node, &desc->chunks)) {
node             1506 drivers/dma/sh/rcar-dmac.c 			desc->running = list_next_entry(desc->running, node);
node             1520 drivers/dma/sh/rcar-dmac.c 						 node);
node             1526 drivers/dma/sh/rcar-dmac.c 	list_move_tail(&desc->node, &chan->desc.done);
node             1532 drivers/dma/sh/rcar-dmac.c 						      node);
node             1619 drivers/dma/sh/rcar-dmac.c 					node);
node             1621 drivers/dma/sh/rcar-dmac.c 		list_del(&desc->node);
node             1635 drivers/dma/sh/rcar-dmac.c 		list_add_tail(&desc->node, &chan->desc.wait);
node               63 drivers/dma/sh/shdma-base.c 	list_for_each_entry(sdesc, &schan->ld_queue, node)
node               86 drivers/dma/sh/shdma-base.c 	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
node               94 drivers/dma/sh/shdma-base.c 				      &chunk->node == &schan->ld_free))
node              105 drivers/dma/sh/shdma-base.c 		list_move_tail(&chunk->node, &schan->ld_queue);
node              161 drivers/dma/sh/shdma-base.c 	list_for_each_entry(sdesc, &schan->ld_free, node)
node              164 drivers/dma/sh/shdma-base.c 			list_del(&sdesc->node);
node              242 drivers/dma/sh/shdma-base.c 		list_add(&desc->node, &schan->ld_free);
node              336 drivers/dma/sh/shdma-base.c 	list_for_each_entry_safe(desc, _desc, &schan->ld_queue, node) {
node              403 drivers/dma/sh/shdma-base.c 				list_move(&desc->node, &schan->ld_free);
node              407 drivers/dma/sh/shdma-base.c 				list_move_tail(&desc->node, &cyclic_list);
node              615 drivers/dma/sh/shdma-base.c 			list_add_tail(&new->node, &tx_list);
node              630 drivers/dma/sh/shdma-base.c 	list_for_each_entry(new, &tx_list, node)
node              766 drivers/dma/sh/shdma-base.c 							   struct shdma_desc, node);
node              839 drivers/dma/sh/shdma-base.c 		list_for_each_entry(sdesc, &schan->ld_queue, node)
node              883 drivers/dma/sh/shdma-base.c 		list_for_each_entry(sdesc, &dl, node) {
node              925 drivers/dma/sh/shdma-base.c 	list_for_each_entry(sdesc, &schan->ld_queue, node) {
node               37 drivers/dma/sh/shdma.h 	struct list_head node;
node               56 drivers/dma/sh/shdma.h #define to_sh_desc(lh) container_of(lh, struct sh_desc, node)
node              495 drivers/dma/sh/shdmac.c 	list_for_each_entry_rcu(shdev, &sh_dmae_devices, node) {
node              772 drivers/dma/sh/shdmac.c 	list_add_tail_rcu(&shdev->node, &sh_dmae_devices);
node              874 drivers/dma/sh/shdmac.c 	list_del_rcu(&shdev->node);
node              895 drivers/dma/sh/shdmac.c 	list_del_rcu(&shdev->node);
node               58 drivers/dma/sh/usb-dmac.c 	struct list_head node;
node              237 drivers/dma/sh/usb-dmac.c 	list_del(&vd->node);
node              274 drivers/dma/sh/usb-dmac.c 	INIT_LIST_HEAD(&desc->node);
node              277 drivers/dma/sh/usb-dmac.c 	list_add_tail(&desc->node, &chan->desc_freed);
node              291 drivers/dma/sh/usb-dmac.c 	list_for_each_entry_safe(desc, _desc, &list, node) {
node              292 drivers/dma/sh/usb-dmac.c 		list_del(&desc->node);
node              306 drivers/dma/sh/usb-dmac.c 	list_for_each_entry(desc, &chan->desc_freed, node) {
node              308 drivers/dma/sh/usb-dmac.c 			list_move_tail(&desc->node, &chan->desc_got);
node              320 drivers/dma/sh/usb-dmac.c 				       node);
node              321 drivers/dma/sh/usb-dmac.c 		list_move_tail(&desc->node, &chan->desc_got);
node              335 drivers/dma/sh/usb-dmac.c 	list_move_tail(&desc->node, &chan->desc_freed);
node              459 drivers/dma/sh/usb-dmac.c 	list_for_each_entry_safe(desc, _desc, &list, node)
node              460 drivers/dma/sh/usb-dmac.c 		list_move_tail(&desc->node, &uchan->desc_freed);
node              493 drivers/dma/sh/usb-dmac.c 	list_for_each_entry_reverse(desc, &chan->desc_freed, node) {
node               83 drivers/dma/sirf-dma.c 	struct list_head		node;
node              249 drivers/dma/sirf-dma.c 				 node);
node              251 drivers/dma/sirf-dma.c 	list_move_tail(&sdesc->node, &schan->active);
node              285 drivers/dma/sirf-dma.c 						 struct sirfsoc_dma_desc, node);
node              307 drivers/dma/sirf-dma.c 					 struct sirfsoc_dma_desc, node);
node              359 drivers/dma/sirf-dma.c 			list_for_each_entry(sdesc, &list, node) {
node              380 drivers/dma/sirf-dma.c 				struct sirfsoc_dma_desc, node);
node              416 drivers/dma/sirf-dma.c 	list_move_tail(&sdesc->node, &schan->queued);
node              580 drivers/dma/sirf-dma.c 		list_add_tail(&sdesc->node, &descs);
node              618 drivers/dma/sirf-dma.c 	list_for_each_entry_safe(sdesc, tmp, &descs, node)
node              661 drivers/dma/sirf-dma.c 	sdesc = list_first_entry(&schan->active, struct sirfsoc_dma_desc, node);
node              707 drivers/dma/sirf-dma.c 			node);
node              708 drivers/dma/sirf-dma.c 		list_del(&sdesc->node);
node              740 drivers/dma/sirf-dma.c 		list_add_tail(&sdesc->node, &schan->prepared);
node              783 drivers/dma/sirf-dma.c 			node);
node              784 drivers/dma/sirf-dma.c 		list_del(&sdesc->node);
node              798 drivers/dma/sirf-dma.c 	list_add_tail(&sdesc->node, &schan->prepared);
node             1090 drivers/dma/sirf-dma.c 			node);
node              516 drivers/dma/sprd-dma.c 	list_del(&vd->node);
node               83 drivers/dma/st_fdma.c 	nbytes = fchan->fdesc->node[0].desc->nbytes;
node               85 drivers/dma/st_fdma.c 	ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START;
node              143 drivers/dma/st_fdma.c 				list_del(&fchan->fdesc->vdesc.node);
node              231 drivers/dma/st_fdma.c 		dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc,
node              232 drivers/dma/st_fdma.c 			      fdesc->node[i].pdesc);
node              242 drivers/dma/st_fdma.c 	fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT);
node              249 drivers/dma/st_fdma.c 		fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
node              250 drivers/dma/st_fdma.c 				GFP_NOWAIT, &fdesc->node[i].pdesc);
node              251 drivers/dma/st_fdma.c 		if (!fdesc->node[i].desc)
node              258 drivers/dma/st_fdma.c 		dma_pool_free(fchan->node_pool, fdesc->node[i].desc,
node              259 drivers/dma/st_fdma.c 			      fdesc->node[i].pdesc);
node              329 drivers/dma/st_fdma.c 	hw_node = fdesc->node[0].desc;
node              483 drivers/dma/st_fdma.c 		struct st_fdma_hw_node *hw_node = fdesc->node[i].desc;
node              485 drivers/dma/st_fdma.c 		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
node              532 drivers/dma/st_fdma.c 		hw_node = fdesc->node[i].desc;
node              534 drivers/dma/st_fdma.c 		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
node              569 drivers/dma/st_fdma.c 		if (cur_addr == fdesc->node[i].pdesc) {
node              573 drivers/dma/st_fdma.c 		residue += fdesc->node[i].desc->nbytes;
node              100 drivers/dma/st_fdma.h 	struct st_fdma_sw_node node[];
node              374 drivers/dma/ste_dma40.c 	struct list_head		 node;
node              763 drivers/dma/ste_dma40.c 	list_del(&d40d->node);
node              774 drivers/dma/ste_dma40.c 		list_for_each_entry_safe(d, _d, &d40c->client, node) {
node              788 drivers/dma/ste_dma40.c 		INIT_LIST_HEAD(&desc->node);
node              803 drivers/dma/ste_dma40.c 	list_add_tail(&desc->node, &d40c->active);
node              825 drivers/dma/ste_dma40.c 	list_add_tail(&desc->node, &d40c->done);
node              951 drivers/dma/ste_dma40.c 	return list_first_entry_or_null(&d40c->active, struct d40_desc, node);
node              959 drivers/dma/ste_dma40.c 	list_add_tail(&desc->node, &d40c->pending_queue);
node              965 drivers/dma/ste_dma40.c 					node);
node              970 drivers/dma/ste_dma40.c 	return list_first_entry_or_null(&d40c->queue, struct d40_desc, node);
node              975 drivers/dma/ste_dma40.c 	return list_first_entry_or_null(&d40c->done, struct d40_desc, node);
node             1139 drivers/dma/ste_dma40.c 		list_for_each_entry_safe(d40d, _d, &d40c->client, node) {
node             1147 drivers/dma/ste_dma40.c 					 &d40c->prepare_queue, node) {
node             1616 drivers/dma/ste_dma40.c 			list_add_tail(&d40d->node, &d40c->client);
node             2260 drivers/dma/ste_dma40.c 	list_add_tail(&desc->node, &chan->prepare_queue);
node              625 drivers/dma/stm32-dma.c 				list_del(&chan->desc->vdesc.node);
node              176 drivers/dma/stm32-dmamux.c 	struct device_node *node = pdev->dev.of_node;
node              185 drivers/dma/stm32-dmamux.c 	if (!node)
node              201 drivers/dma/stm32-dmamux.c 		dma_node = of_parse_phandle(node, "dma-masters", i - 1);
node              290 drivers/dma/stm32-dmamux.c 	return of_dma_router_register(node, stm32_dmamux_route_allocate,
node              255 drivers/dma/stm32-mdma.c 	struct stm32_mdma_desc_node node[];
node              343 drivers/dma/stm32-mdma.c 	desc = kzalloc(offsetof(typeof(*desc), node[count]), GFP_NOWAIT);
node              348 drivers/dma/stm32-mdma.c 		desc->node[i].hwdesc =
node              350 drivers/dma/stm32-mdma.c 				       &desc->node[i].hwdesc_phys);
node              351 drivers/dma/stm32-mdma.c 		if (!desc->node[i].hwdesc)
node              362 drivers/dma/stm32-mdma.c 		dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
node              363 drivers/dma/stm32-mdma.c 			      desc->node[i].hwdesc_phys);
node              375 drivers/dma/stm32-mdma.c 		dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
node              376 drivers/dma/stm32-mdma.c 			      desc->node[i].hwdesc_phys);
node              675 drivers/dma/stm32-mdma.c 				   struct stm32_mdma_desc_node *node)
node              677 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "hwdesc:  %pad\n", &node->hwdesc_phys);
node              678 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CTCR:    0x%08x\n", node->hwdesc->ctcr);
node              679 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CBNDTR:  0x%08x\n", node->hwdesc->cbndtr);
node              680 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CSAR:    0x%08x\n", node->hwdesc->csar);
node              681 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CDAR:    0x%08x\n", node->hwdesc->cdar);
node              682 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CBRUR:   0x%08x\n", node->hwdesc->cbrur);
node              683 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CLAR:    0x%08x\n", node->hwdesc->clar);
node              684 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CTBR:    0x%08x\n", node->hwdesc->ctbr);
node              685 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CMAR:    0x%08x\n", node->hwdesc->cmar);
node              686 drivers/dma/stm32-mdma.c 	dev_dbg(chan2dev(chan), "CMDR:    0x%08x\n\n", node->hwdesc->cmdr);
node              700 drivers/dma/stm32-mdma.c 	hwdesc = desc->node[count].hwdesc;
node              716 drivers/dma/stm32-mdma.c 			hwdesc->clar = desc->node[0].hwdesc_phys;
node              720 drivers/dma/stm32-mdma.c 		hwdesc->clar = desc->node[next].hwdesc_phys;
node              723 drivers/dma/stm32-mdma.c 	stm32_mdma_dump_hwdesc(chan, &desc->node[count]);
node              814 drivers/dma/stm32-mdma.c 		dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
node              815 drivers/dma/stm32-mdma.c 			      desc->node[i].hwdesc_phys);
node              905 drivers/dma/stm32-mdma.c 		dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
node              906 drivers/dma/stm32-mdma.c 			      desc->node[i].hwdesc_phys);
node             1020 drivers/dma/stm32-mdma.c 		hwdesc = desc->node[0].hwdesc;
node             1031 drivers/dma/stm32-mdma.c 		stm32_mdma_dump_hwdesc(chan, &desc->node[0]);
node             1131 drivers/dma/stm32-mdma.c 	hwdesc = chan->desc->node[0].hwdesc;
node             1209 drivers/dma/stm32-mdma.c 	hwdesc = chan->desc->node[chan->curr_hwdesc].hwdesc;
node             1279 drivers/dma/stm32-mdma.c 	struct stm32_mdma_hwdesc *hwdesc = desc->node[0].hwdesc;
node             1285 drivers/dma/stm32-mdma.c 		hwdesc = desc->node[i].hwdesc;
node             1334 drivers/dma/stm32-mdma.c 	list_del(&chan->desc->vdesc.node);
node              359 drivers/dma/sun4i-dma.c 			list_del(&contract->vd.node);
node              177 drivers/dma/sun6i-dma.c 	struct list_head	node;
node              443 drivers/dma/sun6i-dma.c 	list_del(&desc->node);
node              507 drivers/dma/sun6i-dma.c 					 struct sun6i_vchan, node);
node              510 drivers/dma/sun6i-dma.c 		list_del_init(&vchan->node);
node              852 drivers/dma/sun6i-dma.c 		list_del_init(&vchan->node);
node              875 drivers/dma/sun6i-dma.c 		list_add_tail(&vchan->node, &sdev->pending);
node              893 drivers/dma/sun6i-dma.c 	list_del_init(&vchan->node);
node              904 drivers/dma/sun6i-dma.c 			list_add_tail(&vd->node, &vc->desc_completed);
node              976 drivers/dma/sun6i-dma.c 		if (!vchan->phy && list_empty(&vchan->node)) {
node              977 drivers/dma/sun6i-dma.c 			list_add_tail(&vchan->node, &sdev->pending);
node              999 drivers/dma/sun6i-dma.c 	list_del_init(&vchan->node);
node             1358 drivers/dma/sun6i-dma.c 		INIT_LIST_HEAD(&vchan->node);
node              153 drivers/dma/tegra20-apb-dma.c 	struct list_head		node;
node              168 drivers/dma/tegra20-apb-dma.c 	struct list_head		node;
node              283 drivers/dma/tegra20-apb-dma.c 	list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) {
node              285 drivers/dma/tegra20-apb-dma.c 			list_del(&dma_desc->node);
node              313 drivers/dma/tegra20-apb-dma.c 	list_add_tail(&dma_desc->node, &tdc->free_dma_desc);
node              326 drivers/dma/tegra20-apb-dma.c 					typeof(*sg_req), node);
node              327 drivers/dma/tegra20-apb-dma.c 		list_del(&sg_req->node);
node              513 drivers/dma/tegra20-apb-dma.c 					typeof(*sg_req), node);
node              528 drivers/dma/tegra20-apb-dma.c 	hsgreq = list_first_entry(&tdc->pending_sg_req, typeof(*hsgreq), node);
node              529 drivers/dma/tegra20-apb-dma.c 	if (!list_is_last(&hsgreq->node, &tdc->pending_sg_req)) {
node              530 drivers/dma/tegra20-apb-dma.c 		hnsgreq = list_first_entry(&hsgreq->node,
node              531 drivers/dma/tegra20-apb-dma.c 					typeof(*hnsgreq), node);
node              549 drivers/dma/tegra20-apb-dma.c 						typeof(*sgreq), node);
node              550 drivers/dma/tegra20-apb-dma.c 		list_move_tail(&sgreq->node, &tdc->free_sg_req);
node              554 drivers/dma/tegra20-apb-dma.c 			list_add_tail(&dma_desc->node, &tdc->free_dma_desc);
node              582 drivers/dma/tegra20-apb-dma.c 	hsgreq = list_first_entry(&tdc->pending_sg_req, typeof(*hsgreq), node);
node              603 drivers/dma/tegra20-apb-dma.c 	sgreq = list_first_entry(&tdc->pending_sg_req, typeof(*sgreq), node);
node              607 drivers/dma/tegra20-apb-dma.c 	list_del(&sgreq->node);
node              614 drivers/dma/tegra20-apb-dma.c 		list_add_tail(&dma_desc->node, &tdc->free_dma_desc);
node              616 drivers/dma/tegra20-apb-dma.c 	list_add_tail(&sgreq->node, &tdc->free_sg_req);
node              632 drivers/dma/tegra20-apb-dma.c 	sgreq = list_first_entry(&tdc->pending_sg_req, typeof(*sgreq), node);
node              647 drivers/dma/tegra20-apb-dma.c 	if (!list_is_last(&sgreq->node, &tdc->pending_sg_req)) {
node              648 drivers/dma/tegra20-apb-dma.c 		list_move_tail(&sgreq->node, &tdc->pending_sg_req);
node              782 drivers/dma/tegra20-apb-dma.c 					typeof(*sgreq), node);
node              806 drivers/dma/tegra20-apb-dma.c 	if (!list_is_first(&sg_req->node, &tdc->pending_sg_req))
node              878 drivers/dma/tegra20-apb-dma.c 	list_for_each_entry(dma_desc, &tdc->free_dma_desc, node) {
node              886 drivers/dma/tegra20-apb-dma.c 	list_for_each_entry(sg_req, &tdc->pending_sg_req, node) {
node             1099 drivers/dma/tegra20-apb-dma.c 		list_add_tail(&sg_req->node, &dma_desc->tx_list);
node             1233 drivers/dma/tegra20-apb-dma.c 		list_add_tail(&sg_req->node, &dma_desc->tx_list);
node             1304 drivers/dma/tegra20-apb-dma.c 					typeof(*dma_desc), node);
node             1305 drivers/dma/tegra20-apb-dma.c 		list_del(&dma_desc->node);
node             1310 drivers/dma/tegra20-apb-dma.c 		sg_req = list_first_entry(&sg_req_list, typeof(*sg_req), node);
node             1311 drivers/dma/tegra20-apb-dma.c 		list_del(&sg_req->node);
node              360 drivers/dma/tegra210-adma.c 	list_del(&vd->node);
node              104 drivers/dma/ti/cppi41.c 	struct list_head node;		/* Node for pending list */
node              484 drivers/dma/ti/cppi41.c 	list_for_each_entry_safe(c, _c, &cdd->pending, node) {
node              486 drivers/dma/ti/cppi41.c 		list_del(&c->node);
node              507 drivers/dma/ti/cppi41.c 	list_add_tail(&c->node, &cdd->pending);
node              754 drivers/dma/ti/cppi41.c 		list_for_each_entry_safe(cc, _ct, &cdd->pending, node) {
node              757 drivers/dma/ti/cppi41.c 			list_del(&cc->node);
node              132 drivers/dma/ti/dma-crossbar.c 	struct device_node *node = pdev->dev.of_node;
node              140 drivers/dma/ti/dma-crossbar.c 	if (!node)
node              147 drivers/dma/ti/dma-crossbar.c 	dma_node = of_parse_phandle(node, "dma-masters", 0);
node              169 drivers/dma/ti/dma-crossbar.c 	if (of_property_read_u32(node, "dma-requests", &xbar->xbar_events)) {
node              192 drivers/dma/ti/dma-crossbar.c 	ret = of_dma_router_register(node, ti_am335x_xbar_route_allocate,
node              321 drivers/dma/ti/dma-crossbar.c 	struct device_node *node = pdev->dev.of_node;
node              332 drivers/dma/ti/dma-crossbar.c 	if (!node)
node              339 drivers/dma/ti/dma-crossbar.c 	dma_node = of_parse_phandle(node, "dma-masters", 0);
node              367 drivers/dma/ti/dma-crossbar.c 	if (of_property_read_u32(node, "dma-requests", &xbar->xbar_requests)) {
node              374 drivers/dma/ti/dma-crossbar.c 	if (!of_property_read_u32(node, "ti,dma-safe-map", &safe_val))
node              378 drivers/dma/ti/dma-crossbar.c 	prop = of_find_property(node, "ti,reserved-dma-request-ranges", &sz);
node              392 drivers/dma/ti/dma-crossbar.c 		ret = of_property_read_u32_array(node, pname, (u32 *)rsv_events,
node              426 drivers/dma/ti/dma-crossbar.c 	ret = of_dma_router_register(node, ti_dra7_xbar_route_allocate,
node              180 drivers/dma/ti/edma.c 	struct list_head		node;
node              220 drivers/dma/ti/edma.c 	struct device_node		*node;
node              226 drivers/dma/ti/edma.c 	struct list_head		node;
node              775 drivers/dma/ti/edma.c 		list_del(&vdesc->node);
node             1978 drivers/dma/ti/edma.c 		INIT_LIST_HEAD(&echan->node);
node             2258 drivers/dma/ti/edma.c 	struct device_node	*node = pdev->dev.of_node;
node             2264 drivers/dma/ti/edma.c 	if (node) {
node             2267 drivers/dma/ti/edma.c 		match = of_match_node(edma_of_ids, node);
node             2361 drivers/dma/ti/edma.c 	if (irq < 0 && node)
node             2362 drivers/dma/ti/edma.c 		irq = irq_of_parse_and_map(node, 0);
node             2377 drivers/dma/ti/edma.c 	if (irq < 0 && node)
node             2378 drivers/dma/ti/edma.c 		irq = irq_of_parse_and_map(node, 2);
node             2412 drivers/dma/ti/edma.c 			ret = of_parse_phandle_with_fixed_args(node, "ti,tptcs",
node             2417 drivers/dma/ti/edma.c 			ecc->tc_list[i].node = tc_args.np;
node             2469 drivers/dma/ti/edma.c 	if (node)
node             2470 drivers/dma/ti/edma.c 		of_dma_controller_register(node, of_edma_xlate, ecc);
node              540 drivers/dma/ti/omap-dma.c 	list_del(&vd->node);
node               98 drivers/dma/uniphier-mdmac.c 	list_del(&vd->node);
node               29 drivers/dma/virt-dma.c 	list_move_tail(&vd->node, &vc->desc_submitted);
node               56 drivers/dma/virt-dma.c 	list_del(&vd->node);
node               71 drivers/dma/virt-dma.c 	list_for_each_entry(vd, &vc->desc_issued, node)
node              103 drivers/dma/virt-dma.c 	list_for_each_entry_safe(vd, _vd, &head, node) {
node              106 drivers/dma/virt-dma.c 		list_del(&vd->node);
node              116 drivers/dma/virt-dma.c 	list_for_each_entry_safe(vd, _vd, head, node) {
node              118 drivers/dma/virt-dma.c 			list_move_tail(&vd->node, &vc->desc_allocated);
node              121 drivers/dma/virt-dma.c 			list_del(&vd->node);
node               19 drivers/dma/virt-dma.h 	struct list_head node;
node               70 drivers/dma/virt-dma.h 	list_add_tail(&vd->node, &vc->desc_allocated);
node              103 drivers/dma/virt-dma.h 	list_add_tail(&vd->node, &vc->desc_completed);
node              117 drivers/dma/virt-dma.h 		list_add(&vd->node, &vc->desc_allocated);
node              162 drivers/dma/virt-dma.h 					struct virt_dma_desc, node);
node              192 drivers/dma/virt-dma.h 	list_for_each_entry(vd, &head, node)
node              239 drivers/dma/xgene-dma.c 	struct list_head node;
node              481 drivers/dma/xgene-dma.c 	list_del(&desc->node);
node              520 drivers/dma/xgene-dma.c 	list_for_each_entry_safe(desc, _desc, &chan->ld_completed, node) {
node              572 drivers/dma/xgene-dma.c 	list_del(&desc->node);
node              583 drivers/dma/xgene-dma.c 		list_add_tail(&desc->node, &chan->ld_completed);
node              655 drivers/dma/xgene-dma.c 	list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_pending, node) {
node              671 drivers/dma/xgene-dma.c 		list_move_tail(&desc_sw->node, &chan->ld_running);
node              700 drivers/dma/xgene-dma.c 	list_for_each_entry_safe(desc_sw, _desc_sw, &chan->ld_running, node) {
node              753 drivers/dma/xgene-dma.c 		list_move_tail(&desc_sw->node, &ld_completed);
node              766 drivers/dma/xgene-dma.c 	list_for_each_entry_safe(desc_sw, _desc_sw, &ld_completed, node) {
node              805 drivers/dma/xgene-dma.c 	list_for_each_entry_safe(desc, _desc, list, node)
node              866 drivers/dma/xgene-dma.c 		list_add_tail(&new->node, &first->tx_list);
node              924 drivers/dma/xgene-dma.c 		list_add_tail(&new->node, &first->tx_list);
node              272 drivers/dma/xilinx/xilinx_dma.c 	struct list_head node;
node              284 drivers/dma/xilinx/xilinx_dma.c 	struct list_head node;
node              296 drivers/dma/xilinx/xilinx_dma.c 	struct list_head node;
node              310 drivers/dma/xilinx/xilinx_dma.c 	struct list_head node;
node              611 drivers/dma/xilinx/xilinx_dma.c 					   node);
node              612 drivers/dma/xilinx/xilinx_dma.c 		list_del(&segment->node);
node              640 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&segment->node, &chan->free_seg_list);
node              702 drivers/dma/xilinx/xilinx_dma.c 		list_for_each_entry_safe(segment, next, &desc->segments, node) {
node              703 drivers/dma/xilinx/xilinx_dma.c 			list_del(&segment->node);
node              708 drivers/dma/xilinx/xilinx_dma.c 					 &desc->segments, node) {
node              709 drivers/dma/xilinx/xilinx_dma.c 			list_del(&cdma_segment->node);
node              714 drivers/dma/xilinx/xilinx_dma.c 					 &desc->segments, node) {
node              715 drivers/dma/xilinx/xilinx_dma.c 			list_del(&axidma_segment->node);
node              735 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, list, node) {
node              736 drivers/dma/xilinx/xilinx_dma.c 		list_del(&desc->node);
node              825 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, &chan->done_list, node) {
node              834 drivers/dma/xilinx/xilinx_dma.c 		list_del(&desc->node);
node              922 drivers/dma/xilinx/xilinx_dma.c 			list_add_tail(&chan->seg_v[i].node,
node             1020 drivers/dma/xilinx/xilinx_dma.c 				       struct xilinx_dma_tx_descriptor, node);
node             1022 drivers/dma/xilinx/xilinx_dma.c 			list_for_each_entry(segment, &desc->segments, node) {
node             1117 drivers/dma/xilinx/xilinx_dma.c 				struct xilinx_dma_tx_descriptor, node);
node             1164 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry(segment, &desc->segments, node) {
node             1189 drivers/dma/xilinx/xilinx_dma.c 	list_del(&desc->node);
node             1190 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&desc->node, &chan->active_list);
node             1217 drivers/dma/xilinx/xilinx_dma.c 				     struct xilinx_dma_tx_descriptor, node);
node             1219 drivers/dma/xilinx/xilinx_dma.c 				    struct xilinx_dma_tx_descriptor, node);
node             1221 drivers/dma/xilinx/xilinx_dma.c 				       struct xilinx_cdma_tx_segment, node);
node             1250 drivers/dma/xilinx/xilinx_dma.c 					   node);
node             1289 drivers/dma/xilinx/xilinx_dma.c 				     struct xilinx_dma_tx_descriptor, node);
node             1291 drivers/dma/xilinx/xilinx_dma.c 				    struct xilinx_dma_tx_descriptor, node);
node             1293 drivers/dma/xilinx/xilinx_dma.c 				       struct xilinx_axidma_tx_segment, node);
node             1357 drivers/dma/xilinx/xilinx_dma.c 					   node);
node             1401 drivers/dma/xilinx/xilinx_dma.c 	list_for_each_entry_safe(desc, next, &chan->active_list, node) {
node             1402 drivers/dma/xilinx/xilinx_dma.c 		list_del(&desc->node);
node             1405 drivers/dma/xilinx/xilinx_dma.c 		list_add_tail(&desc->node, &chan->done_list);
node             1549 drivers/dma/xilinx/xilinx_dma.c 				    struct xilinx_dma_tx_descriptor, node);
node             1553 drivers/dma/xilinx/xilinx_dma.c 					       node);
node             1558 drivers/dma/xilinx/xilinx_dma.c 						node);
node             1563 drivers/dma/xilinx/xilinx_dma.c 					       node);
node             1572 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&desc->node, &chan->pending_list);
node             1694 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&segment->node, &desc->segments);
node             1698 drivers/dma/xilinx/xilinx_dma.c 				   struct xilinx_vdma_tx_segment, node);
node             1752 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&segment->node, &desc->segments);
node             1839 drivers/dma/xilinx/xilinx_dma.c 			list_add_tail(&segment->node, &desc->segments);
node             1844 drivers/dma/xilinx/xilinx_dma.c 				   struct xilinx_axidma_tx_segment, node);
node             1852 drivers/dma/xilinx/xilinx_dma.c 					  node);
node             1939 drivers/dma/xilinx/xilinx_dma.c 			list_add_tail(&segment->node, &desc->segments);
node             1944 drivers/dma/xilinx/xilinx_dma.c 				   struct xilinx_axidma_tx_segment, node);
node             1954 drivers/dma/xilinx/xilinx_dma.c 				  node);
node             2031 drivers/dma/xilinx/xilinx_dma.c 	list_add_tail(&segment->node, &desc->segments);
node             2035 drivers/dma/xilinx/xilinx_dma.c 				   struct xilinx_axidma_tx_segment, node);
node             2043 drivers/dma/xilinx/xilinx_dma.c 					  node);
node             2388 drivers/dma/xilinx/xilinx_dma.c 				  struct device_node *node, int chan_id)
node             2418 drivers/dma/xilinx/xilinx_dma.c 	has_dre = of_property_read_bool(node, "xlnx,include-dre");
node             2420 drivers/dma/xilinx/xilinx_dma.c 	chan->genlock = of_property_read_bool(node, "xlnx,genlock-mode");
node             2422 drivers/dma/xilinx/xilinx_dma.c 	err = of_property_read_u32(node, "xlnx,datawidth", &value);
node             2436 drivers/dma/xilinx/xilinx_dma.c 	if (of_device_is_compatible(node, "xlnx,axi-vdma-mm2s-channel") ||
node             2437 drivers/dma/xilinx/xilinx_dma.c 	    of_device_is_compatible(node, "xlnx,axi-dma-mm2s-channel") ||
node             2438 drivers/dma/xilinx/xilinx_dma.c 	    of_device_is_compatible(node, "xlnx,axi-cdma-channel")) {
node             2452 drivers/dma/xilinx/xilinx_dma.c 	} else if (of_device_is_compatible(node,
node             2454 drivers/dma/xilinx/xilinx_dma.c 		   of_device_is_compatible(node,
node             2459 drivers/dma/xilinx/xilinx_dma.c 		chan->has_vflip = of_property_read_bool(node,
node             2482 drivers/dma/xilinx/xilinx_dma.c 	chan->irq = irq_of_parse_and_map(node, 0);
node             2544 drivers/dma/xilinx/xilinx_dma.c 				    struct device_node *node)
node             2548 drivers/dma/xilinx/xilinx_dma.c 	ret = of_property_read_u32(node, "dma-channels", &nr_channels);
node             2553 drivers/dma/xilinx/xilinx_dma.c 		xilinx_dma_chan_probe(xdev, node, xdev->chan_id++);
node             2613 drivers/dma/xilinx/xilinx_dma.c 	struct device_node *node = pdev->dev.of_node;
node             2651 drivers/dma/xilinx/xilinx_dma.c 		xdev->mcdma = of_property_read_bool(node, "xlnx,mcdma");
node             2652 drivers/dma/xilinx/xilinx_dma.c 		if (!of_property_read_u32(node, "xlnx,sg-length-width",
node             2668 drivers/dma/xilinx/xilinx_dma.c 		err = of_property_read_u32(node, "xlnx,num-fstores",
node             2676 drivers/dma/xilinx/xilinx_dma.c 		err = of_property_read_u32(node, "xlnx,flush-fsync",
node             2683 drivers/dma/xilinx/xilinx_dma.c 	err = of_property_read_u32(node, "xlnx,addrwidth", &addr_width);
node             2732 drivers/dma/xilinx/xilinx_dma.c 	for_each_child_of_node(node, child) {
node             2747 drivers/dma/xilinx/xilinx_dma.c 	err = of_dma_controller_register(node, of_dma_xilinx_xlate,
node              181 drivers/dma/xilinx/zynqmp_dma.c 	struct list_head node;
node              382 drivers/dma/xilinx/zynqmp_dma.c 				     struct zynqmp_dma_desc_sw, node);
node              385 drivers/dma/xilinx/zynqmp_dma.c 					       struct zynqmp_dma_desc_sw, node);
node              392 drivers/dma/xilinx/zynqmp_dma.c 	list_add_tail(&new->node, &chan->pending_list);
node              412 drivers/dma/xilinx/zynqmp_dma.c 				struct zynqmp_dma_desc_sw, node);
node              413 drivers/dma/xilinx/zynqmp_dma.c 	list_del(&desc->node);
node              435 drivers/dma/xilinx/zynqmp_dma.c 	list_add_tail(&sdesc->node, &chan->free_list);
node              436 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
node              438 drivers/dma/xilinx/zynqmp_dma.c 		list_move_tail(&child->node, &chan->free_list);
node              452 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(desc, next, list, node)
node              486 drivers/dma/xilinx/zynqmp_dma.c 		list_add_tail(&desc->node, &chan->free_list);
node              583 drivers/dma/xilinx/zynqmp_dma.c 					struct zynqmp_dma_desc_sw, node);
node              601 drivers/dma/xilinx/zynqmp_dma.c 	list_for_each_entry_safe(desc, next, &chan->done_list, node) {
node              605 drivers/dma/xilinx/zynqmp_dma.c 		list_del(&desc->node);
node              629 drivers/dma/xilinx/zynqmp_dma.c 					struct zynqmp_dma_desc_sw, node);
node              632 drivers/dma/xilinx/zynqmp_dma.c 	list_del(&desc->node);
node              634 drivers/dma/xilinx/zynqmp_dma.c 	list_add_tail(&desc->node, &chan->done_list);
node              839 drivers/dma/xilinx/zynqmp_dma.c 			list_add_tail(&new->node, &first->tx_list);
node              875 drivers/dma/xilinx/zynqmp_dma.c 	struct device_node *node = pdev->dev.of_node;
node              892 drivers/dma/xilinx/zynqmp_dma.c 	err = of_property_read_u32(node, "xlnx,bus-width", &chan->bus_width);
node              904 drivers/dma/xilinx/zynqmp_dma.c 	chan->is_dmacoherent =  of_property_read_bool(node, "dma-coherent");
node              104 drivers/dma/zx_dma.c 	struct list_head	node;
node              205 drivers/dma/zx_dma.c 		list_del(&ds->vd.node);
node              243 drivers/dma/zx_dma.c 				     struct zx_dma_chan, node);
node              247 drivers/dma/zx_dma.c 			list_del_init(&c->node);
node              324 drivers/dma/zx_dma.c 	list_del_init(&c->node);
node              390 drivers/dma/zx_dma.c 		if (!c->phy && list_empty(&c->node)) {
node              392 drivers/dma/zx_dma.c 			list_add_tail(&c->node, &d->chan_pending);
node              671 drivers/dma/zx_dma.c 	list_del_init(&c->node);
node              843 drivers/dma/zx_dma.c 		INIT_LIST_HEAD(&c->node);
node               27 drivers/edac/cell_edac.c 	int				node;
node               41 drivers/edac/cell_edac.c 		priv->node, chan, ar);
node               64 drivers/edac/cell_edac.c 		priv->node, chan, ar);
node              146 drivers/edac/cell_edac.c 		if (of_node_to_nid(np) != priv->node)
node              161 drivers/edac/cell_edac.c 			priv->node, priv->chanmask,
node              214 drivers/edac/cell_edac.c 	priv->node = pdev->id;
node              335 drivers/edac/ghes_edac.c 		p += sprintf(p, "node:%d ", mem_err->node);
node              193 drivers/edac/thunderx_edac.c 	u64 node;
node              418 drivers/edac/thunderx_edac.c 	lmc->mem = alloc_pages_node(lmc->node, GFP_KERNEL, 0);
node              496 drivers/edac/thunderx_edac.c 	addr |= lmc->node << 40;
node              520 drivers/edac/thunderx_edac.c static unsigned int thunderx_get_num_lmcs(unsigned int node)
node              531 drivers/edac/thunderx_edac.c 			if (pdev->dev.numa_node == node)
node              653 drivers/edac/thunderx_edac.c 	int node = dev_to_node(&pdev->dev);
node              656 drivers/edac/thunderx_edac.c 	ret += max(node, 0) << 3;
node              743 drivers/edac/thunderx_edac.c 	lmc->node = FIELD_GET(THUNDERX_NODE, pci_resource_start(pdev, 0));
node              745 drivers/edac/thunderx_edac.c 	lmc->xbits = thunderx_get_num_lmcs(lmc->node) >> 1;
node              760 drivers/edac/thunderx_edac.c 	l2c_ioaddr = ioremap(L2C_CTL | FIELD_PREP(THUNDERX_NODE, lmc->node), PAGE_SIZE);
node              201 drivers/edac/ti_edac.c static int _emif_get_id(struct device_node *node)
node              208 drivers/edac/ti_edac.c 	addrp = of_get_address(node, 0, NULL, NULL);
node              209 drivers/edac/ti_edac.c 	my_addr = (u32)of_translate_address(node, addrp);
node              212 drivers/edac/ti_edac.c 		if (np == node)
node             1818 drivers/edac/xgene_edac.c 	struct xgene_edac_dev_ctx *node;
node             1838 drivers/edac/xgene_edac.c 	list_for_each_entry(node, &ctx->l3s, next)
node             1839 drivers/edac/xgene_edac.c 		xgene_edac_l3_check(node->edac_dev);
node             1841 drivers/edac/xgene_edac.c 	list_for_each_entry(node, &ctx->socs, next)
node             1842 drivers/edac/xgene_edac.c 		xgene_edac_soc_check(node->edac_dev);
node             1973 drivers/edac/xgene_edac.c 	struct xgene_edac_dev_ctx *node;
node             1982 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(node, temp_node, &edac->l3s, next)
node             1983 drivers/edac/xgene_edac.c 		xgene_edac_l3_remove(node);
node             1985 drivers/edac/xgene_edac.c 	list_for_each_entry_safe(node, temp_node, &edac->socs, next)
node             1986 drivers/edac/xgene_edac.c 		xgene_edac_soc_remove(node);
node              178 drivers/extcon/extcon-palmas.c 	struct device_node *node = pdev->dev.of_node;
node              191 drivers/extcon/extcon-palmas.c 	if (node && !pdata) {
node              192 drivers/extcon/extcon-palmas.c 		palmas_usb->wakeup = of_property_read_bool(node, "ti,wakeup");
node              193 drivers/extcon/extcon-palmas.c 		palmas_usb->enable_id_detection = of_property_read_bool(node,
node              195 drivers/extcon/extcon-palmas.c 		palmas_usb->enable_vbus_detection = of_property_read_bool(node,
node              233 drivers/extcon/extcon-palmas.c 		if (of_property_read_u32(node, "debounce-delay-ms", &debounce))
node             1336 drivers/extcon/extcon.c struct extcon_dev *extcon_find_edev_by_node(struct device_node *node)
node             1342 drivers/extcon/extcon.c 		if (edev->dev.parent && edev->dev.parent->of_node == node)
node             1360 drivers/extcon/extcon.c 	struct device_node *node;
node             1371 drivers/extcon/extcon.c 	node = of_parse_phandle(dev->of_node, "extcon", index);
node             1372 drivers/extcon/extcon.c 	if (!node) {
node             1378 drivers/extcon/extcon.c 	edev = extcon_find_edev_by_node(node);
node             1379 drivers/extcon/extcon.c 	of_node_put(node);
node             1386 drivers/extcon/extcon.c struct extcon_dev *extcon_find_edev_by_node(struct device_node *node)
node              544 drivers/firewire/core-device.c 	device->max_speed = device->node->max_speed;
node              830 drivers/firewire/core-device.c 	device->node->data = NULL;
node              833 drivers/firewire/core-device.c 	fw_node_put(device->node);
node              894 drivers/firewire/core-device.c 		struct fw_node *current_node = new->node;
node              895 drivers/firewire/core-device.c 		struct fw_node *obsolete_node = old->node;
node              897 drivers/firewire/core-device.c 		new->node = obsolete_node;
node              898 drivers/firewire/core-device.c 		new->node->data = new;
node              899 drivers/firewire/core-device.c 		old->node = current_node;
node              900 drivers/firewire/core-device.c 		old->node->data = old;
node             1005 drivers/firewire/core-device.c 			if (device->node->link_on)
node             1009 drivers/firewire/core-device.c 			if (device->node == card->root_node)
node             1088 drivers/firewire/core-device.c 	if (device->node == card->root_node)
node             1203 drivers/firewire/core-device.c void fw_node_event(struct fw_card *card, struct fw_node *node, int event)
node             1232 drivers/firewire/core-device.c 		device->node = fw_node_get(node);
node             1233 drivers/firewire/core-device.c 		device->node_id = node->node_id;
node             1235 drivers/firewire/core-device.c 		device->is_local = node == card->local_node;
node             1244 drivers/firewire/core-device.c 		node->data = device;
node             1259 drivers/firewire/core-device.c 		device = node->data;
node             1263 drivers/firewire/core-device.c 		device->node_id = node->node_id;
node             1276 drivers/firewire/core-device.c 		device = node->data;
node             1280 drivers/firewire/core-device.c 		device->node_id = node->node_id;
node             1291 drivers/firewire/core-device.c 		if (!node->data)
node             1306 drivers/firewire/core-device.c 		device = node->data;
node              101 drivers/firewire/core-topology.c 	struct fw_node *node;
node              103 drivers/firewire/core-topology.c 	node = kzalloc(struct_size(node, ports, port_count), GFP_ATOMIC);
node              104 drivers/firewire/core-topology.c 	if (node == NULL)
node              107 drivers/firewire/core-topology.c 	node->color = color;
node              108 drivers/firewire/core-topology.c 	node->node_id = LOCAL_BUS | SELF_ID_PHY_ID(sid);
node              109 drivers/firewire/core-topology.c 	node->link_on = SELF_ID_LINK_ON(sid);
node              110 drivers/firewire/core-topology.c 	node->phy_speed = SELF_ID_PHY_SPEED(sid);
node              111 drivers/firewire/core-topology.c 	node->initiated_reset = SELF_ID_PHY_INITIATOR(sid);
node              112 drivers/firewire/core-topology.c 	node->port_count = port_count;
node              114 drivers/firewire/core-topology.c 	refcount_set(&node->ref_count, 1);
node              115 drivers/firewire/core-topology.c 	INIT_LIST_HEAD(&node->link);
node              117 drivers/firewire/core-topology.c 	return node;
node              134 drivers/firewire/core-topology.c static void update_hop_count(struct fw_node *node)
node              140 drivers/firewire/core-topology.c 	for (i = 0; i < node->port_count; i++) {
node              141 drivers/firewire/core-topology.c 		if (node->ports[i] == NULL)
node              144 drivers/firewire/core-topology.c 		if (node->ports[i]->max_hops > max_child_hops)
node              145 drivers/firewire/core-topology.c 			max_child_hops = node->ports[i]->max_hops;
node              147 drivers/firewire/core-topology.c 		if (node->ports[i]->max_depth > depths[0]) {
node              149 drivers/firewire/core-topology.c 			depths[0] = node->ports[i]->max_depth;
node              150 drivers/firewire/core-topology.c 		} else if (node->ports[i]->max_depth > depths[1])
node              151 drivers/firewire/core-topology.c 			depths[1] = node->ports[i]->max_depth;
node              154 drivers/firewire/core-topology.c 	node->max_depth = depths[0] + 1;
node              155 drivers/firewire/core-topology.c 	node->max_hops = max(max_child_hops, depths[0] + depths[1] + 2);
node              173 drivers/firewire/core-topology.c 	struct fw_node *node, *child, *local_node, *irm_node;
node              181 drivers/firewire/core-topology.c 	node = NULL;
node              222 drivers/firewire/core-topology.c 		node = fw_node_create(q, port_count, card->color);
node              223 drivers/firewire/core-topology.c 		if (node == NULL) {
node              229 drivers/firewire/core-topology.c 			local_node = node;
node              232 drivers/firewire/core-topology.c 			irm_node = node;
node              250 drivers/firewire/core-topology.c 				node->color = i;
node              254 drivers/firewire/core-topology.c 				node->ports[i] = child;
node              259 drivers/firewire/core-topology.c 				child->ports[child->color] = node;
node              280 drivers/firewire/core-topology.c 		list_add_tail(&node->link, &stack);
node              283 drivers/firewire/core-topology.c 		if (node->phy_speed == SCODE_BETA &&
node              294 drivers/firewire/core-topology.c 		update_hop_count(node);
node              300 drivers/firewire/core-topology.c 	card->root_node = node;
node              309 drivers/firewire/core-topology.c 				   struct fw_node * node,
node              316 drivers/firewire/core-topology.c 	struct fw_node *node, *next, *child, *parent;
node              324 drivers/firewire/core-topology.c 	list_for_each_entry(node, &list, link) {
node              325 drivers/firewire/core-topology.c 		node->color = card->color;
node              327 drivers/firewire/core-topology.c 		for (i = 0; i < node->port_count; i++) {
node              328 drivers/firewire/core-topology.c 			child = node->ports[i];
node              339 drivers/firewire/core-topology.c 		callback(card, node, parent);
node              342 drivers/firewire/core-topology.c 	list_for_each_entry_safe(node, next, &list, link)
node              343 drivers/firewire/core-topology.c 		fw_node_put(node);
node              347 drivers/firewire/core-topology.c 			     struct fw_node *node, struct fw_node *parent)
node              349 drivers/firewire/core-topology.c 	fw_node_event(card, node, FW_NODE_DESTROYED);
node              350 drivers/firewire/core-topology.c 	fw_node_put(node);
node              357 drivers/firewire/core-topology.c 			      struct fw_node *node, struct fw_node *parent)
node              359 drivers/firewire/core-topology.c 	int b_path = (node->phy_speed == SCODE_BETA);
node              363 drivers/firewire/core-topology.c 		node->max_speed = parent->max_speed < node->phy_speed ?
node              364 drivers/firewire/core-topology.c 					parent->max_speed : node->phy_speed;
node              365 drivers/firewire/core-topology.c 		node->b_path = parent->b_path && b_path;
node              367 drivers/firewire/core-topology.c 		node->max_speed = node->phy_speed;
node              368 drivers/firewire/core-topology.c 		node->b_path = b_path;
node              371 drivers/firewire/core-topology.c 	fw_node_event(card, node, FW_NODE_CREATED);
node              153 drivers/firewire/core.h void fw_node_event(struct fw_card *card, struct fw_node *node, int event);
node              199 drivers/firewire/core.h static inline struct fw_node *fw_node_get(struct fw_node *node)
node              201 drivers/firewire/core.h 	refcount_inc(&node->ref_count);
node              203 drivers/firewire/core.h 	return node;
node              206 drivers/firewire/core.h static inline void fw_node_put(struct fw_node *node)
node              208 drivers/firewire/core.h 	if (refcount_dec_and_test(&node->ref_count))
node              209 drivers/firewire/core.h 		kfree(node);
node              137 drivers/firmware/arm_scmi/driver.c 	struct list_head node;
node              635 drivers/firmware/arm_scmi/driver.c 		info = list_entry(p, struct scmi_info, node);
node              852 drivers/firmware/arm_scmi/driver.c 	INIT_LIST_HEAD(&info->node);
node              877 drivers/firmware/arm_scmi/driver.c 	list_add_tail(&info->node, &scmi_list);
node              926 drivers/firmware/arm_scmi/driver.c 		list_del(&info->node);
node              228 drivers/firmware/arm_scpi.c 	struct list_head node;
node              370 drivers/firmware/arm_scpi.c 					 node);
node              371 drivers/firmware/arm_scpi.c 		list_del(&match->node);
node              373 drivers/firmware/arm_scpi.c 		list_for_each_entry(t, &ch->rx_pending, node)
node              375 drivers/firmware/arm_scpi.c 				list_del(&t->node);
node              440 drivers/firmware/arm_scpi.c 		list_add_tail(&t->node, &ch->rx_pending);
node              457 drivers/firmware/arm_scpi.c 	t = list_first_entry(&ch->xfers_list, struct scpi_xfer, node);
node              458 drivers/firmware/arm_scpi.c 	list_del(&t->node);
node              466 drivers/firmware/arm_scpi.c 	list_add_tail(&t->node, &ch->xfers_list);
node              889 drivers/firmware/arm_scpi.c 		list_add_tail(&xfers->node, &ch->xfers_list);
node              226 drivers/firmware/efi/cper.c 		n += scnprintf(msg + n, len - n, "node: %d ", mem->node);
node              284 drivers/firmware/efi/cper.c 	cmem->node = mem->node;
node               34 drivers/firmware/efi/dev-path-parser.c static long __init parse_acpi_path(struct efi_dev_path *node,
node               40 drivers/firmware/efi/dev-path-parser.c 	if (node->length != 12)
node               44 drivers/firmware/efi/dev-path-parser.c 		'A' + ((node->acpi.hid >> 10) & 0x1f) - 1,
node               45 drivers/firmware/efi/dev-path-parser.c 		'A' + ((node->acpi.hid >>  5) & 0x1f) - 1,
node               46 drivers/firmware/efi/dev-path-parser.c 		'A' + ((node->acpi.hid >>  0) & 0x1f) - 1,
node               47 drivers/firmware/efi/dev-path-parser.c 			node->acpi.hid >> 16);
node               48 drivers/firmware/efi/dev-path-parser.c 	sprintf(hid_uid.uid, "%u", node->acpi.uid);
node               72 drivers/firmware/efi/dev-path-parser.c static long __init parse_pci_path(struct efi_dev_path *node,
node               77 drivers/firmware/efi/dev-path-parser.c 	if (node->length != 6)
node               82 drivers/firmware/efi/dev-path-parser.c 	devfn = PCI_DEVFN(node->pci.dev, node->pci.fn);
node              108 drivers/firmware/efi/dev-path-parser.c static long __init parse_end_path(struct efi_dev_path *node,
node              111 drivers/firmware/efi/dev-path-parser.c 	if (node->length != 4)
node              113 drivers/firmware/efi/dev-path-parser.c 	if (node->sub_type != EFI_DEV_END_INSTANCE &&
node              114 drivers/firmware/efi/dev-path-parser.c 	    node->sub_type != EFI_DEV_END_ENTIRE)
node              120 drivers/firmware/efi/dev-path-parser.c 	return node->sub_type;
node              159 drivers/firmware/efi/dev-path-parser.c struct device * __init efi_get_device_by_path(struct efi_dev_path **node,
node              169 drivers/firmware/efi/dev-path-parser.c 		if (*len < 4 || *len < (*node)->length)
node              171 drivers/firmware/efi/dev-path-parser.c 		else if ((*node)->type     == EFI_DEV_ACPI &&
node              172 drivers/firmware/efi/dev-path-parser.c 			 (*node)->sub_type == EFI_DEV_BASIC_ACPI)
node              173 drivers/firmware/efi/dev-path-parser.c 			ret = parse_acpi_path(*node, parent, &child);
node              174 drivers/firmware/efi/dev-path-parser.c 		else if ((*node)->type     == EFI_DEV_HW &&
node              175 drivers/firmware/efi/dev-path-parser.c 			 (*node)->sub_type == EFI_DEV_PCI)
node              176 drivers/firmware/efi/dev-path-parser.c 			ret = parse_pci_path(*node, parent, &child);
node              177 drivers/firmware/efi/dev-path-parser.c 		else if (((*node)->type    == EFI_DEV_END_PATH ||
node              178 drivers/firmware/efi/dev-path-parser.c 			  (*node)->type    == EFI_DEV_END_PATH2))
node              179 drivers/firmware/efi/dev-path-parser.c 			ret = parse_end_path(*node, parent, &child);
node              188 drivers/firmware/efi/dev-path-parser.c 		*node  = (void *)*node + (*node)->length;
node              189 drivers/firmware/efi/dev-path-parser.c 		*len  -= (*node)->length;
node              719 drivers/firmware/efi/efi.c static int __init __find_uefi_params(unsigned long node,
node              729 drivers/firmware/efi/efi.c 		prop = of_get_flat_dt_prop(node, params[i].propname, &len);
node              753 drivers/firmware/efi/efi.c static int __init fdt_find_uefi_params(unsigned long node, const char *uname,
node              768 drivers/firmware/efi/efi.c 			int err = of_get_flat_dt_subnode_by_name(node, subnode);
node              773 drivers/firmware/efi/efi.c 			node = err;
node              776 drivers/firmware/efi/efi.c 		return __find_uefi_params(node, info, dt_params[i].params);
node               35 drivers/firmware/efi/libstub/fdt.c 	int node, num_rsv;
node               80 drivers/firmware/efi/libstub/fdt.c 	node = fdt_subnode_offset(fdt, 0, "chosen");
node               81 drivers/firmware/efi/libstub/fdt.c 	if (node < 0) {
node               82 drivers/firmware/efi/libstub/fdt.c 		node = fdt_add_subnode(fdt, 0, "chosen");
node               83 drivers/firmware/efi/libstub/fdt.c 		if (node < 0) {
node               85 drivers/firmware/efi/libstub/fdt.c 			status = node;
node               91 drivers/firmware/efi/libstub/fdt.c 		status = fdt_setprop(fdt, node, "bootargs", cmdline_ptr,
node              102 drivers/firmware/efi/libstub/fdt.c 		status = fdt_setprop_var(fdt, node, "linux,initrd-start", initrd_image_start);
node              107 drivers/firmware/efi/libstub/fdt.c 		status = fdt_setprop_var(fdt, node, "linux,initrd-end", initrd_image_end);
node              113 drivers/firmware/efi/libstub/fdt.c 	node = fdt_subnode_offset(fdt, 0, "chosen");
node              116 drivers/firmware/efi/libstub/fdt.c 	status = fdt_setprop_var(fdt, node, "linux,uefi-system-table", fdt_val64);
node              122 drivers/firmware/efi/libstub/fdt.c 	status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-start", fdt_val64);
node              128 drivers/firmware/efi/libstub/fdt.c 	status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-size", fdt_val32);
node              132 drivers/firmware/efi/libstub/fdt.c 	status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-desc-size", fdt_val32);
node              136 drivers/firmware/efi/libstub/fdt.c 	status = fdt_setprop_var(fdt, node, "linux,uefi-mmap-desc-ver", fdt_val32);
node              146 drivers/firmware/efi/libstub/fdt.c 			status = fdt_setprop_var(fdt, node, "kaslr-seed", fdt_val64);
node              168 drivers/firmware/efi/libstub/fdt.c 	int node = fdt_path_offset(fdt, "/chosen");
node              173 drivers/firmware/efi/libstub/fdt.c 	if (node < 0)
node              178 drivers/firmware/efi/libstub/fdt.c 	err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-start", fdt_val64);
node              184 drivers/firmware/efi/libstub/fdt.c 	err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-size", fdt_val32);
node              190 drivers/firmware/efi/libstub/fdt.c 	err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-desc-size", fdt_val32);
node              196 drivers/firmware/efi/libstub/fdt.c 	err = fdt_setprop_inplace_var(fdt, node, "linux,uefi-mmap-desc-ver", fdt_val32);
node               43 drivers/firmware/efi/vars.c 	struct efi_generic_dev_path *node;
node               46 drivers/firmware/efi/vars.c 	node = (struct efi_generic_dev_path *)buffer;
node               48 drivers/firmware/efi/vars.c 	if (len < sizeof(*node))
node               51 drivers/firmware/efi/vars.c 	while (offset <= len - sizeof(*node) &&
node               52 drivers/firmware/efi/vars.c 	       node->length >= sizeof(*node) &&
node               53 drivers/firmware/efi/vars.c 		node->length <= len - offset) {
node               54 drivers/firmware/efi/vars.c 		offset += node->length;
node               56 drivers/firmware/efi/vars.c 		if ((node->type == EFI_DEV_END_PATH ||
node               57 drivers/firmware/efi/vars.c 		     node->type == EFI_DEV_END_PATH2) &&
node               58 drivers/firmware/efi/vars.c 		    node->sub_type == EFI_DEV_END_ENTIRE)
node               61 drivers/firmware/efi/vars.c 		node = (struct efi_generic_dev_path *)(buffer + offset);
node               94 drivers/firmware/stratix10-svc.c 	struct list_head node;
node              139 drivers/firmware/stratix10-svc.c 	struct list_head node;
node              180 drivers/firmware/stratix10-svc.c 	list_for_each_entry(pmem, &svc_data_mem, node)
node              739 drivers/firmware/stratix10-svc.c 				      struct stratix10_svc_controller, node);
node              838 drivers/firmware/stratix10-svc.c 		list_for_each_entry(p_mem, &svc_data_mem, node)
node              919 drivers/firmware/stratix10-svc.c 	list_add_tail(&pmem->node, &svc_data_mem);
node              939 drivers/firmware/stratix10-svc.c 	list_for_each_entry(pmem, &svc_data_mem, node)
node              947 drivers/firmware/stratix10-svc.c 	list_del(&pmem->node);
node             1024 drivers/firmware/stratix10-svc.c 	list_add_tail(&controller->node, &svc_ctrl);
node             1064 drivers/firmware/stratix10-svc.c 	list_del(&ctrl->node);
node              131 drivers/firmware/ti_sci.c 	struct list_head node;
node             3023 drivers/firmware/ti_sci.c 		info = list_entry(p, struct ti_sci_info, node);
node             3146 drivers/firmware/ti_sci.c 		info = list_entry(p, struct ti_sci_info, node);
node             3428 drivers/firmware/ti_sci.c 	INIT_LIST_HEAD(&info->node);
node             3517 drivers/firmware/ti_sci.c 	list_add_tail(&info->node, &ti_sci_list);
node             3547 drivers/firmware/ti_sci.c 		list_del(&info->node);
node              145 drivers/firmware/trusted_foundations.c 	struct device_node *node;
node              149 drivers/firmware/trusted_foundations.c 	node = of_find_compatible_node(NULL, NULL, "tlm,trusted-foundations");
node              150 drivers/firmware/trusted_foundations.c 	if (!node)
node              153 drivers/firmware/trusted_foundations.c 	err = of_property_read_u32(node, "tlm,version-major",
node              157 drivers/firmware/trusted_foundations.c 	err = of_property_read_u32(node, "tlm,version-minor",
node              624 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_request_node(const u32 node, const u32 capabilities,
node              628 drivers/firmware/xilinx/zynqmp.c 	return zynqmp_pm_invoke_fn(PM_REQUEST_NODE, node, capabilities,
node              642 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_release_node(const u32 node)
node              644 drivers/firmware/xilinx/zynqmp.c 	return zynqmp_pm_invoke_fn(PM_RELEASE_NODE, node, 0, 0, 0, NULL);
node              659 drivers/firmware/xilinx/zynqmp.c static int zynqmp_pm_set_requirement(const u32 node, const u32 capabilities,
node              663 drivers/firmware/xilinx/zynqmp.c 	return zynqmp_pm_invoke_fn(PM_SET_REQUIREMENT, node, capabilities,
node              166 drivers/fpga/dfl-afu-dma-region.c 		this = container_of(*new, struct dfl_afu_dma_region, node);
node              181 drivers/fpga/dfl-afu-dma-region.c 	rb_link_node(&region->node, parent, new);
node              182 drivers/fpga/dfl-afu-dma-region.c 	rb_insert_color(&region->node, &afu->dma_regions);
node              203 drivers/fpga/dfl-afu-dma-region.c 	rb_erase(&region->node, &afu->dma_regions);
node              215 drivers/fpga/dfl-afu-dma-region.c 	struct rb_node *node = rb_first(&afu->dma_regions);
node              218 drivers/fpga/dfl-afu-dma-region.c 	while (node) {
node              219 drivers/fpga/dfl-afu-dma-region.c 		region = container_of(node, struct dfl_afu_dma_region, node);
node              224 drivers/fpga/dfl-afu-dma-region.c 		rb_erase(node, &afu->dma_regions);
node              234 drivers/fpga/dfl-afu-dma-region.c 		node = rb_next(node);
node              257 drivers/fpga/dfl-afu-dma-region.c 	struct rb_node *node = afu->dma_regions.rb_node;
node              260 drivers/fpga/dfl-afu-dma-region.c 	while (node) {
node              263 drivers/fpga/dfl-afu-dma-region.c 		region = container_of(node, struct dfl_afu_dma_region, node);
node              272 drivers/fpga/dfl-afu-dma-region.c 			node = node->rb_left;
node              274 drivers/fpga/dfl-afu-dma-region.c 			node = node->rb_right;
node               25 drivers/fpga/dfl-afu-region.c 	list_for_each_entry((region), &(afu)->regions, node)
node               78 drivers/fpga/dfl-afu-region.c 	list_add(&region->node, &afu->regions);
node              100 drivers/fpga/dfl-afu-region.c 	list_for_each_entry_safe(region, tmp, &afu->regions, node)
node               40 drivers/fpga/dfl-afu.h 	struct list_head node;
node               58 drivers/fpga/dfl-afu.h 	struct rb_node node;
node               37 drivers/fpga/dfl-fme-pr.c 	list_for_each_entry(fme_region, &fme->region_list, node)
node              288 drivers/fpga/dfl-fme-pr.c 	list_for_each_entry_safe(fbridge, tmp, &priv->bridge_list, node) {
node              289 drivers/fpga/dfl-fme-pr.c 		list_del(&fbridge->node);
node              367 drivers/fpga/dfl-fme-pr.c 	list_for_each_entry_safe(fme_region, tmp, &priv->region_list, node) {
node              368 drivers/fpga/dfl-fme-pr.c 		list_del(&fme_region->node);
node              418 drivers/fpga/dfl-fme-pr.c 		list_add(&fme_br->node, &priv->bridge_list);
node              428 drivers/fpga/dfl-fme-pr.c 		list_add(&fme_region->node, &priv->region_list);
node               32 drivers/fpga/dfl-fme-pr.h 	struct list_head node;
node               57 drivers/fpga/dfl-fme-pr.h 	struct list_head node;
node              174 drivers/fpga/dfl.c 	list_for_each_entry(ops, &dfl_port_ops_list, node) {
node              208 drivers/fpga/dfl.c 	list_add_tail(&ops->node, &dfl_port_ops_list);
node              220 drivers/fpga/dfl.c 	list_del(&ops->node);
node              450 drivers/fpga/dfl.c 	struct list_head node;
node              459 drivers/fpga/dfl.c 	list_add(&pdata->node, &cdev->port_dev_list);
node              521 drivers/fpga/dfl.c 	list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
node              530 drivers/fpga/dfl.c 		list_del(&finfo->node);
node              605 drivers/fpga/dfl.c 		list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
node              606 drivers/fpga/dfl.c 			list_del(&finfo->node);
node              672 drivers/fpga/dfl.c 	list_add_tail(&finfo->node, &binfo->sub_features);
node              851 drivers/fpga/dfl.c 	list_for_each_entry_safe(dfl, tmp, &info->dfls, node) {
node              852 drivers/fpga/dfl.c 		list_del(&dfl->node);
node              889 drivers/fpga/dfl.c 	list_add_tail(&dfl->node, &info->dfls);
node              966 drivers/fpga/dfl.c 	list_for_each_entry(dfl, &info->dfls, node) {
node             1002 drivers/fpga/dfl.c 	list_for_each_entry_safe(pdata, ptmp, &cdev->port_dev_list, node) {
node             1012 drivers/fpga/dfl.c 		list_del(&pdata->node);
node             1045 drivers/fpga/dfl.c 	list_for_each_entry(pdata, &cdev->port_dev_list, node) {
node             1184 drivers/fpga/dfl.c 	list_for_each_entry(pdata, &cdev->port_dev_list, node) {
node             1221 drivers/fpga/dfl.c 	list_for_each_entry(pdata, &cdev->port_dev_list, node) {
node              160 drivers/fpga/dfl.h 	struct list_head node;
node              228 drivers/fpga/dfl.h 	struct list_head node;
node              392 drivers/fpga/dfl.h 	struct list_head node;
node              167 drivers/fpga/fpga-bridge.c 	list_for_each_entry(bridge, bridge_list, node) {
node              191 drivers/fpga/fpga-bridge.c 	list_for_each_entry(bridge, bridge_list, node) {
node              214 drivers/fpga/fpga-bridge.c 	list_for_each_entry_safe(bridge, next, bridge_list, node) {
node              218 drivers/fpga/fpga-bridge.c 		list_del(&bridge->node);
node              247 drivers/fpga/fpga-bridge.c 	list_add(&bridge->node, bridge_list);
node              277 drivers/fpga/fpga-bridge.c 	list_add(&bridge->node, bridge_list);
node              349 drivers/fpga/fpga-bridge.c 	INIT_LIST_HEAD(&bridge->node);
node              492 drivers/fpga/fpga-mgr.c struct fpga_manager *of_fpga_mgr_get(struct device_node *node)
node              496 drivers/fpga/fpga-mgr.c 	dev = class_find_device_by_of_node(fpga_mgr_class, node);
node              113 drivers/gnss/serial.c 	struct device_node *node = serdev->dev.of_node;
node              116 drivers/gnss/serial.c 	of_property_read_u32(node, "current-speed", &speed);
node              385 drivers/gnss/sirf.c 	struct device_node *node = serdev->dev.of_node;
node              388 drivers/gnss/sirf.c 	of_property_read_u32(node, "current-speed", &speed);
node              238 drivers/gpio/gpio-altera.c 	struct device_node *node = pdev->dev.of_node;
node              249 drivers/gpio/gpio-altera.c 	if (of_property_read_u32(node, "altr,ngpio", &reg))
node              274 drivers/gpio/gpio-altera.c 	if (of_property_read_u32(node, "altr,interrupt-type", &reg)) {
node              304 drivers/gpio/gpio-altera.c 	ret = of_mm_gpiochip_add_data(node, &altera_gc->mmchip, altera_gc);
node               47 drivers/gpio/gpio-brcmstb.c 	struct list_head node;
node              304 drivers/gpio/gpio-brcmstb.c 	list_for_each_entry(bank, &priv->bank_list, node)
node              316 drivers/gpio/gpio-brcmstb.c 	list_for_each_entry_reverse(bank, &priv->bank_list, node) {
node              412 drivers/gpio/gpio-brcmstb.c 	list_for_each_entry(bank, &priv->bank_list, node)
node              535 drivers/gpio/gpio-brcmstb.c 	list_for_each_entry(bank, &priv->bank_list, node) {
node              581 drivers/gpio/gpio-brcmstb.c 	list_for_each_entry(bank, &priv->bank_list, node) {
node              741 drivers/gpio/gpio-brcmstb.c 		list_add(&bank->node, &priv->bank_list);
node              620 drivers/gpio/gpio-max732x.c 	struct device_node *node;
node              627 drivers/gpio/gpio-max732x.c 	node = client->dev.of_node;
node              629 drivers/gpio/gpio-max732x.c 	if (!pdata && node)
node              211 drivers/gpio/gpio-mt7621.c 			 struct device_node *node, int bank)
node              222 drivers/gpio/gpio-mt7621.c 	rg->chip.of_node = node;
node               59 drivers/gpio/gpio-mxc.c 	struct list_head node;
node              301 drivers/gpio/gpio-mxc.c 	list_for_each_entry(port, &mxc_gpio_ports, node) {
node              519 drivers/gpio/gpio-mxc.c 	list_add_tail(&port->node, &mxc_gpio_ports);
node              564 drivers/gpio/gpio-mxc.c 	list_for_each_entry(port, &mxc_gpio_ports, node) {
node              578 drivers/gpio/gpio-mxc.c 	list_for_each_entry(port, &mxc_gpio_ports, node) {
node             1346 drivers/gpio/gpio-omap.c 	struct device_node *node = dev->of_node;
node             1399 drivers/gpio/gpio-omap.c 	bank->chip.of_node = of_node_get(node);
node             1402 drivers/gpio/gpio-omap.c 	if (node) {
node             1403 drivers/gpio/gpio-omap.c 		if (!of_property_read_bool(node, "ti,gpio-always-on"))
node               92 drivers/gpio/gpio-sodaville.c static int sdv_xlate(struct irq_domain *h, struct device_node *node,
node               98 drivers/gpio/gpio-sodaville.c 	if (node != irq_domain_get_of_node(h))
node               24 drivers/gpio/gpio-ts4800.c 	struct device_node *node;
node               38 drivers/gpio/gpio-ts4800.c 	node = pdev->dev.of_node;
node               39 drivers/gpio/gpio-ts4800.c 	if (!node)
node               42 drivers/gpio/gpio-ts4800.c 	retval = of_property_read_u32(node, "ngpios", &ngpios);
node              500 drivers/gpio/gpio-twl4030.c 	struct device_node *node = pdev->dev.of_node;
node              522 drivers/gpio/gpio-twl4030.c 	irq_domain_add_legacy(node, TWL4030_GPIO_MAX, irq_base, 0,
node              539 drivers/gpio/gpio-twl4030.c 	if (node)
node               54 drivers/gpio/gpiolib-acpi.c 	struct list_head node;
node               66 drivers/gpio/gpiolib-acpi.c 	struct list_head node;
node              204 drivers/gpio/gpiolib-acpi.c 	list_for_each_entry(event, &acpi_gpio->events, node)
node              343 drivers/gpio/gpiolib-acpi.c 	list_add_tail(&event->node, &acpi_gpio->events);
node              430 drivers/gpio/gpiolib-acpi.c 	list_for_each_entry_safe_reverse(event, ep, &acpi_gpio->events, node) {
node              440 drivers/gpio/gpiolib-acpi.c 		list_del(&event->node);
node             1011 drivers/gpio/gpiolib-acpi.c 		list_for_each_entry(conn, &achip->conns, node) {
node             1028 drivers/gpio/gpiolib-acpi.c 			list_for_each_entry(event, &achip->events, node) {
node             1060 drivers/gpio/gpiolib-acpi.c 			list_add_tail(&conn->node, &achip->conns);
node             1108 drivers/gpio/gpiolib-acpi.c 	list_for_each_entry_safe_reverse(conn, tmp, &achip->conns, node) {
node             1110 drivers/gpio/gpiolib-acpi.c 		list_del(&conn->node);
node              148 drivers/gpio/gpiolib-devres.c 					      struct device_node *node,
node              156 drivers/gpio/gpiolib-devres.c 	desc = gpiod_get_from_of_node(node, propname, index, dflags, label);
node              316 drivers/gpio/gpiolib-of.c struct gpio_desc *gpiod_get_from_of_node(struct device_node *node,
node              330 drivers/gpio/gpiolib-of.c 	desc = of_get_named_gpiod_flags(node, propname,
node             2581 drivers/gpio/gpiolib.c 	list_add_tail(&pin_range->node, &gdev->pin_ranges);
node             2638 drivers/gpio/gpiolib.c 	list_add_tail(&pin_range->node, &gdev->pin_ranges);
node             2653 drivers/gpio/gpiolib.c 	list_for_each_entry_safe(pin_range, tmp, &gdev->pin_ranges, node) {
node             2654 drivers/gpio/gpiolib.c 		list_del(&pin_range->node);
node              857 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              858 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              897 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              898 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              907 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              908 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              917 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              918 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              983 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	list_for_each_entry(s_job, &sched->ring_mirror_list, node) {
node             1011 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	list_for_each_entry(s_job, &sched->ring_mirror_list, node) {
node              699 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              700 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	struct drm_device *dev = node->minor->dev;
node              748 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              749 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	struct drm_device *dev = node->minor->dev;
node              849 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              850 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	struct drm_device *dev = node->minor->dev;
node               34 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct drm_mm_node node;
node              153 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct amdgpu_gtt_node *node = mem->mm_node;
node              155 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	return (node->node.start != AMDGPU_BO_INVALID_OFFSET);
node              175 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct amdgpu_gtt_node *node = mem->mm_node;
node              198 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	r = drm_mm_insert_node_in_range(&mgr->mm, &node->node, mem->num_pages,
node              204 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		mem->start = node->node.start;
node              225 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct amdgpu_gtt_node *node;
node              237 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              238 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	if (!node) {
node              243 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	node->node.start = AMDGPU_BO_INVALID_OFFSET;
node              244 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	node->node.size = mem->num_pages;
node              245 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	node->tbo = tbo;
node              246 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	mem->mm_node = node;
node              251 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 			kfree(node);
node              257 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		mem->start = node->node.start;
node              281 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct amdgpu_gtt_node *node = mem->mm_node;
node              283 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	if (!node)
node              287 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	if (node->node.start != AMDGPU_BO_INVALID_OFFSET)
node              288 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		drm_mm_remove_node(&node->node);
node              292 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	kfree(node);
node              314 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct amdgpu_gtt_node *node;
node              320 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		node = container_of(mm_node, struct amdgpu_gtt_node, node);
node              321 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		r = amdgpu_ttm_recover_gart(node->tbo);
node              408 drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              409 drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c 	struct drm_device *dev = node->minor->dev;
node             1226 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             1227 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	struct drm_device *dev = node->minor->dev;
node               77 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	struct amdgpu_mn_node *node, *next_node;
node               82 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	hash_del(&amn->node);
node               83 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	rbtree_postorder_for_each_entry_safe(node, next_node,
node               85 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		list_for_each_entry_safe(bo, next_bo, &node->bos, mn_list) {
node               89 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		kfree(node);
node              170 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c static void amdgpu_mn_invalidate_node(struct amdgpu_mn_node *node,
node              177 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	list_for_each_entry(bo, &node->bos, mn_list) {
node              219 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		struct amdgpu_mn_node *node;
node              226 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		node = container_of(it, struct amdgpu_mn_node, it);
node              229 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		amdgpu_mn_invalidate_node(node, start, end);
node              265 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		struct amdgpu_mn_node *node;
node              273 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		node = container_of(it, struct amdgpu_mn_node, it);
node              276 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		list_for_each_entry(bo, &node->bos, mn_list) {
node              329 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	hash_for_each_possible(adev->mn_hash, amn, node, key)
node              350 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	hash_add(adev->mn_hash, &amn->node, AMDGPU_MN_KEY(mm, type));
node              382 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	struct amdgpu_mn_node *node = NULL, *new_node;
node              399 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		kfree(node);
node              400 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		node = container_of(it, struct amdgpu_mn_node, it);
node              401 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		interval_tree_remove(&node->it, &amn->objects);
node              404 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		list_splice(&node->bos, &bos);
node              407 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	if (!node)
node              408 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		node = new_node;
node              414 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	node->it.start = addr;
node              415 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	node->it.last = end;
node              416 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	INIT_LIST_HEAD(&node->bos);
node              417 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	list_splice(&bos, &node->bos);
node              418 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	list_add(&bo->mn_list, &node->bos);
node              420 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	interval_tree_insert(&node->it, &amn->objects);
node              457 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		struct amdgpu_mn_node *node;
node              459 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		node = container_of(head, struct amdgpu_mn_node, bos);
node              460 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		interval_tree_remove(&node->it, &amn->objects);
node              461 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		kfree(node);
node               62 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.h 	struct hlist_node	node;
node              204 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	struct drm_mm_node *node = bo->tbo.mem.mm_node;
node              211 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	     pages_left -= node->size, node++)
node              212 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 		if (node->start < fpfn)
node             3145 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             3146 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c 	struct drm_device *dev = node->minor->dev;
node              318 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 		list_del(&obj->node);
node              344 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_add(&obj->node, &con->head);
node              538 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_for_each_entry_safe(obj, tmp, &con->head, node) {
node              695 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_for_each_entry(obj, &con->head, node) {
node              914 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_for_each_entry_safe(obj, tmp, &con->head, node) {
node              973 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_for_each_entry_safe(obj, tmp, &con->head, node) {
node             1143 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 	list_for_each_entry_safe(obj, tmp, &con->head, node) {
node             1505 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c 		list_for_each_entry_safe(obj, tmp, &con->head, node) {
node              391 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h 	struct list_head node;
node               36 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	struct hlist_node	node;
node              136 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_possible(sync->fences, e, node, f->context) {
node              177 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_add(sync->fences, &e->node, f->context);
node              265 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
node              270 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 			hash_del(&e->node);
node              307 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
node              313 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 		hash_del(&e->node);
node              340 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(source->fences, i, tmp, e, node) {
node              347 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 			hash_del(&e->node);
node              365 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
node              370 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 		hash_del(&e->node);
node              391 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
node              392 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 		hash_del(&e->node);
node             1478 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct drm_mm_node *node = bo->mem.mm_node;
node             1511 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 			if (place->fpfn < (node->start + node->size) &&
node             1512 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 			    !(place->lpfn && place->lpfn <= node->start))
node             1515 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 			num_pages -= node->size;
node             1516 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 			++node;
node             2133 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node             2134 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	unsigned ttm_pl = (uintptr_t)node->info_ent->data;
node             2135 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct drm_device *dev = node->minor->dev;
node               60 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define START(node) ((node)->start)
node               61 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define LAST(node) ((node)->last)
node              195 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 				    struct drm_mm_node *node)
node              197 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 	uint64_t start = node->start << PAGE_SHIFT;
node              198 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 	uint64_t end = (node->size + node->start) << PAGE_SHIFT;
node              245 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 				       struct drm_mm_node *node)
node              249 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 	start = node->start + node->size;
node              170 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 	char node[10] = { 0 };
node              195 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 	sprintf(node, "node%d", hive->number_devices);
node              197 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c 	ret = sysfs_create_link(hive->kobj, &adev->dev->kobj, node);
node             1029 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node             1030 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node             1046 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node             1047 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node             1058 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node             1059 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              115 drivers/gpu/drm/arc/arcpgu_drv.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              116 drivers/gpu/drm/arc/arcpgu_drv.c 	struct drm_device *drm = node->minor->dev;
node              105 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 	struct komeda_plane_state *node, *last;
node              120 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 	list_for_each_entry(node, zorder_list, zlist_node) {
node              121 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 		if (new->base.zpos < node->base.zpos) {
node              122 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 			list_add_tail(&new->zlist_node, &node->zlist_node);
node              124 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 		} else if (node->base.zpos == new->base.zpos) {
node              125 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 			struct drm_plane *a = node->base.plane;
node              132 drivers/gpu/drm/arm/display/komeda/komeda_kms.c 					 a->name, b->name, node->base.zpos);
node              198 drivers/gpu/drm/arm/hdlcd_drv.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              199 drivers/gpu/drm/arm/hdlcd_drv.c 	struct drm_device *drm = node->minor->dev;
node              211 drivers/gpu/drm/arm/hdlcd_drv.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              212 drivers/gpu/drm/arm/hdlcd_drv.c 	struct drm_device *drm = node->minor->dev;
node               20 drivers/gpu/drm/armada/armada_debugfs.c 	struct drm_info_node *node = m->private;
node               21 drivers/gpu/drm/armada/armada_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              132 drivers/gpu/drm/armada/armada_gem.c 		struct drm_mm_node *node;
node              137 drivers/gpu/drm/armada/armada_gem.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node              138 drivers/gpu/drm/armada/armada_gem.c 		if (!node)
node              142 drivers/gpu/drm/armada/armada_gem.c 		ret = drm_mm_insert_node_generic(&priv->linear, node,
node              146 drivers/gpu/drm/armada/armada_gem.c 			kfree(node);
node              150 drivers/gpu/drm/armada/armada_gem.c 		obj->linear = node;
node              147 drivers/gpu/drm/bridge/adv7511/adv7533.c 						   .node = NULL,
node               39 drivers/gpu/drm/bridge/analogix/analogix_dp_core.c 	struct device_node *node;
node              111 drivers/gpu/drm/bridge/sil-sii8620.c 	struct list_head node;
node              270 drivers/gpu/drm/bridge/sil-sii8620.c 	list_for_each_entry_safe(msg, n, &ctx->mt_queue, node) {
node              271 drivers/gpu/drm/bridge/sil-sii8620.c 		list_del(&msg->node);
node              289 drivers/gpu/drm/bridge/sil-sii8620.c 				       node);
node              290 drivers/gpu/drm/bridge/sil-sii8620.c 		list_del(&msg->node);
node              302 drivers/gpu/drm/bridge/sil-sii8620.c 	msg = list_first_entry(&ctx->mt_queue, struct sii8620_mt_msg, node);
node              391 drivers/gpu/drm/bridge/sil-sii8620.c 		list_add_tail(&msg->node, &ctx->mt_queue);
node              407 drivers/gpu/drm/bridge/sil-sii8620.c 	msg = list_last_entry(&ctx->mt_queue, struct sii8620_mt_msg, node);
node             1813 drivers/gpu/drm/bridge/sil-sii8620.c 	return list_first_entry(&ctx->mt_queue, struct sii8620_mt_msg, node);
node              274 drivers/gpu/drm/bridge/ti-sn65dsi86.c 						   .node = NULL,
node             1511 drivers/gpu/drm/drm_atomic.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             1512 drivers/gpu/drm/drm_atomic.c 	struct drm_device *dev = node->minor->dev;
node              443 drivers/gpu/drm/drm_client.c 	struct drm_info_node *node = m->private;
node              444 drivers/gpu/drm/drm_client.c 	struct drm_device *dev = node->minor->dev;
node               53 drivers/gpu/drm/drm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               54 drivers/gpu/drm/drm_debugfs.c 	struct drm_minor *minor = node->minor;
node               75 drivers/gpu/drm/drm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               76 drivers/gpu/drm/drm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              126 drivers/gpu/drm/drm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              127 drivers/gpu/drm/drm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              148 drivers/gpu/drm/drm_debugfs.c 	struct drm_info_node *node = inode->i_private;
node              150 drivers/gpu/drm/drm_debugfs.c 	return single_open(file, node->info_ent->show, node);
node              287 drivers/gpu/drm/drm_debugfs.c 	struct drm_info_node *node, *tmp;
node              290 drivers/gpu/drm/drm_debugfs.c 	list_for_each_entry_safe(node, tmp, &minor->debugfs_list, list) {
node              291 drivers/gpu/drm/drm_debugfs.c 		debugfs_remove(node->dent);
node              292 drivers/gpu/drm/drm_debugfs.c 		list_del(&node->list);
node              293 drivers/gpu/drm/drm_debugfs.c 		kfree(node);
node             1842 drivers/gpu/drm/drm_fb_helper.c 		 info->node, info->fix.id);
node              345 drivers/gpu/drm/drm_file.c 				struct pci_bus, node);
node               63 drivers/gpu/drm/drm_flip_work.c 	list_add_tail(&task->node, &work->queued);
node              132 drivers/gpu/drm/drm_flip_work.c 		list_for_each_entry_safe(task, tmp, &tasks, node) {
node             1069 drivers/gpu/drm/drm_framebuffer.c 	struct drm_info_node *node = m->private;
node             1070 drivers/gpu/drm/drm_framebuffer.c 	struct drm_device *dev = node->minor->dev;
node             1151 drivers/gpu/drm/drm_gem.c 	struct drm_vma_offset_node *node;
node             1158 drivers/gpu/drm/drm_gem.c 	node = drm_vma_offset_exact_lookup_locked(dev->vma_offset_manager,
node             1161 drivers/gpu/drm/drm_gem.c 	if (likely(node)) {
node             1162 drivers/gpu/drm/drm_gem.c 		obj = container_of(node, struct drm_gem_object, vma_node);
node             1181 drivers/gpu/drm/drm_gem.c 	if (!drm_vma_node_is_allowed(node, priv)) {
node             1186 drivers/gpu/drm/drm_gem.c 	if (node->readonly) {
node             1195 drivers/gpu/drm/drm_gem.c 	ret = drm_gem_mmap_obj(obj, drm_vma_node_size(node) << PAGE_SHIFT,
node              352 drivers/gpu/drm/drm_gem_cma_helper.c 	struct drm_vma_offset_node *node;
node              358 drivers/gpu/drm/drm_gem_cma_helper.c 	node = drm_vma_offset_exact_lookup_locked(dev->vma_offset_manager,
node              361 drivers/gpu/drm/drm_gem_cma_helper.c 	if (likely(node)) {
node              362 drivers/gpu/drm/drm_gem_cma_helper.c 		obj = container_of(node, struct drm_gem_object, vma_node);
node              382 drivers/gpu/drm/drm_gem_cma_helper.c 	if (!drm_vma_node_is_allowed(node, priv)) {
node              155 drivers/gpu/drm/drm_mipi_dsi.c of_mipi_dsi_device_add(struct mipi_dsi_host *host, struct device_node *node)
node              162 drivers/gpu/drm/drm_mipi_dsi.c 	if (of_modalias_node(node, info.type, sizeof(info.type)) < 0) {
node              163 drivers/gpu/drm/drm_mipi_dsi.c 		dev_err(dev, "modalias failure on %pOF\n", node);
node              167 drivers/gpu/drm/drm_mipi_dsi.c 	ret = of_property_read_u32(node, "reg", &reg);
node              170 drivers/gpu/drm/drm_mipi_dsi.c 			node, ret);
node              175 drivers/gpu/drm/drm_mipi_dsi.c 	info.node = of_node_get(node);
node              181 drivers/gpu/drm/drm_mipi_dsi.c of_mipi_dsi_device_add(struct mipi_dsi_host *host, struct device_node *node)
node              224 drivers/gpu/drm/drm_mipi_dsi.c 	dsi->dev.of_node = info->node;
node              261 drivers/gpu/drm/drm_mipi_dsi.c struct mipi_dsi_host *of_find_mipi_dsi_host_by_node(struct device_node *node)
node              268 drivers/gpu/drm/drm_mipi_dsi.c 		if (host->dev->of_node == node) {
node              282 drivers/gpu/drm/drm_mipi_dsi.c 	struct device_node *node;
node              284 drivers/gpu/drm/drm_mipi_dsi.c 	for_each_available_child_of_node(host->dev->of_node, node) {
node              286 drivers/gpu/drm/drm_mipi_dsi.c 		if (!of_find_property(node, "reg", NULL))
node              288 drivers/gpu/drm/drm_mipi_dsi.c 		of_mipi_dsi_device_add(host, node);
node              107 drivers/gpu/drm/drm_mm.c static noinline void save_stack(struct drm_mm_node *node)
node              115 drivers/gpu/drm/drm_mm.c 	node->stack = stack_depot_save(entries, n, GFP_NOWAIT);
node              120 drivers/gpu/drm/drm_mm.c 	struct drm_mm_node *node;
node              129 drivers/gpu/drm/drm_mm.c 	list_for_each_entry(node, drm_mm_nodes(mm), node_list) {
node              130 drivers/gpu/drm/drm_mm.c 		if (!node->stack) {
node              132 drivers/gpu/drm/drm_mm.c 				  node->start, node->size);
node              136 drivers/gpu/drm/drm_mm.c 		nr_entries = stack_depot_fetch(node->stack, &entries);
node              139 drivers/gpu/drm/drm_mm.c 			  node->start, node->size, buf);
node              148 drivers/gpu/drm/drm_mm.c static void save_stack(struct drm_mm_node *node) { }
node              152 drivers/gpu/drm/drm_mm.c #define START(node) ((node)->start)
node              153 drivers/gpu/drm/drm_mm.c #define LAST(node)  ((node)->start + (node)->size - 1)
node              168 drivers/gpu/drm/drm_mm.c 					  struct drm_mm_node *node)
node              175 drivers/gpu/drm/drm_mm.c 	node->__subtree_last = LAST(node);
node              181 drivers/gpu/drm/drm_mm.c 			if (parent->__subtree_last >= node->__subtree_last)
node              184 drivers/gpu/drm/drm_mm.c 			parent->__subtree_last = node->__subtree_last;
node              200 drivers/gpu/drm/drm_mm.c 		if (parent->__subtree_last < node->__subtree_last)
node              201 drivers/gpu/drm/drm_mm.c 			parent->__subtree_last = node->__subtree_last;
node              202 drivers/gpu/drm/drm_mm.c 		if (node->start < parent->start) {
node              210 drivers/gpu/drm/drm_mm.c 	rb_link_node(&node->rb, rb, link);
node              211 drivers/gpu/drm/drm_mm.c 	rb_insert_augmented_cached(&node->rb, &mm->interval_tree, leftmost,
node              217 drivers/gpu/drm/drm_mm.c 	u64 x = expr(node); \
node              225 drivers/gpu/drm/drm_mm.c 	rb_link_node(&node->member, rb, link); \
node              226 drivers/gpu/drm/drm_mm.c 	rb_insert_color(&node->member, &root); \
node              238 drivers/gpu/drm/drm_mm.c 			     struct drm_mm_node *node)
node              241 drivers/gpu/drm/drm_mm.c 	u64 x = node->hole_size;
node              254 drivers/gpu/drm/drm_mm.c 	rb_link_node(&node->rb_hole_size, rb, link);
node              255 drivers/gpu/drm/drm_mm.c 	rb_insert_color_cached(&node->rb_hole_size, root, first);
node              258 drivers/gpu/drm/drm_mm.c static void add_hole(struct drm_mm_node *node)
node              260 drivers/gpu/drm/drm_mm.c 	struct drm_mm *mm = node->mm;
node              262 drivers/gpu/drm/drm_mm.c 	node->hole_size =
node              263 drivers/gpu/drm/drm_mm.c 		__drm_mm_hole_node_end(node) - __drm_mm_hole_node_start(node);
node              264 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!drm_mm_hole_follows(node));
node              266 drivers/gpu/drm/drm_mm.c 	insert_hole_size(&mm->holes_size, node);
node              269 drivers/gpu/drm/drm_mm.c 	list_add(&node->hole_stack, &mm->hole_stack);
node              272 drivers/gpu/drm/drm_mm.c static void rm_hole(struct drm_mm_node *node)
node              274 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!drm_mm_hole_follows(node));
node              276 drivers/gpu/drm/drm_mm.c 	list_del(&node->hole_stack);
node              277 drivers/gpu/drm/drm_mm.c 	rb_erase_cached(&node->rb_hole_size, &node->mm->holes_size);
node              278 drivers/gpu/drm/drm_mm.c 	rb_erase(&node->rb_hole_addr, &node->mm->holes_addr);
node              279 drivers/gpu/drm/drm_mm.c 	node->hole_size = 0;
node              281 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(drm_mm_hole_follows(node));
node              305 drivers/gpu/drm/drm_mm.c 		struct drm_mm_node *node =
node              308 drivers/gpu/drm/drm_mm.c 		if (size <= node->hole_size) {
node              309 drivers/gpu/drm/drm_mm.c 			best = node;
node              322 drivers/gpu/drm/drm_mm.c 	struct drm_mm_node *node = NULL;
node              327 drivers/gpu/drm/drm_mm.c 		node = rb_hole_addr_to_node(rb);
node              328 drivers/gpu/drm/drm_mm.c 		hole_start = __drm_mm_hole_node_start(node);
node              331 drivers/gpu/drm/drm_mm.c 			rb = node->rb_hole_addr.rb_left;
node              332 drivers/gpu/drm/drm_mm.c 		else if (addr > hole_start + node->hole_size)
node              333 drivers/gpu/drm/drm_mm.c 			rb = node->rb_hole_addr.rb_right;
node              338 drivers/gpu/drm/drm_mm.c 	return node;
node              366 drivers/gpu/drm/drm_mm.c 	  struct drm_mm_node *node,
node              372 drivers/gpu/drm/drm_mm.c 		return rb_hole_size_to_node(rb_prev(&node->rb_hole_size));
node              375 drivers/gpu/drm/drm_mm.c 		return rb_hole_addr_to_node(rb_next(&node->rb_hole_addr));
node              378 drivers/gpu/drm/drm_mm.c 		return rb_hole_addr_to_node(rb_prev(&node->rb_hole_addr));
node              381 drivers/gpu/drm/drm_mm.c 		node = list_next_entry(node, hole_stack);
node              382 drivers/gpu/drm/drm_mm.c 		return &node->hole_stack == &mm->hole_stack ? NULL : node;
node              400 drivers/gpu/drm/drm_mm.c int drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node)
node              402 drivers/gpu/drm/drm_mm.c 	u64 end = node->start + node->size;
node              407 drivers/gpu/drm/drm_mm.c 	end = node->start + node->size;
node              408 drivers/gpu/drm/drm_mm.c 	if (unlikely(end <= node->start))
node              412 drivers/gpu/drm/drm_mm.c 	hole = find_hole(mm, node->start);
node              420 drivers/gpu/drm/drm_mm.c 		mm->color_adjust(hole, node->color, &adj_start, &adj_end);
node              422 drivers/gpu/drm/drm_mm.c 	if (adj_start > node->start || adj_end < end)
node              425 drivers/gpu/drm/drm_mm.c 	node->mm = mm;
node              427 drivers/gpu/drm/drm_mm.c 	list_add(&node->node_list, &hole->node_list);
node              428 drivers/gpu/drm/drm_mm.c 	drm_mm_interval_tree_add_node(hole, node);
node              429 drivers/gpu/drm/drm_mm.c 	node->allocated = true;
node              430 drivers/gpu/drm/drm_mm.c 	node->hole_size = 0;
node              433 drivers/gpu/drm/drm_mm.c 	if (node->start > hole_start)
node              436 drivers/gpu/drm/drm_mm.c 		add_hole(node);
node              438 drivers/gpu/drm/drm_mm.c 	save_stack(node);
node              465 drivers/gpu/drm/drm_mm.c 				struct drm_mm_node * const node,
node              540 drivers/gpu/drm/drm_mm.c 		node->mm = mm;
node              541 drivers/gpu/drm/drm_mm.c 		node->size = size;
node              542 drivers/gpu/drm/drm_mm.c 		node->start = adj_start;
node              543 drivers/gpu/drm/drm_mm.c 		node->color = color;
node              544 drivers/gpu/drm/drm_mm.c 		node->hole_size = 0;
node              546 drivers/gpu/drm/drm_mm.c 		list_add(&node->node_list, &hole->node_list);
node              547 drivers/gpu/drm/drm_mm.c 		drm_mm_interval_tree_add_node(hole, node);
node              548 drivers/gpu/drm/drm_mm.c 		node->allocated = true;
node              554 drivers/gpu/drm/drm_mm.c 			add_hole(node);
node              556 drivers/gpu/drm/drm_mm.c 		save_stack(node);
node              572 drivers/gpu/drm/drm_mm.c void drm_mm_remove_node(struct drm_mm_node *node)
node              574 drivers/gpu/drm/drm_mm.c 	struct drm_mm *mm = node->mm;
node              577 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!node->allocated);
node              578 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(node->scanned_block);
node              580 drivers/gpu/drm/drm_mm.c 	prev_node = list_prev_entry(node, node_list);
node              582 drivers/gpu/drm/drm_mm.c 	if (drm_mm_hole_follows(node))
node              583 drivers/gpu/drm/drm_mm.c 		rm_hole(node);
node              585 drivers/gpu/drm/drm_mm.c 	drm_mm_interval_tree_remove(node, &mm->interval_tree);
node              586 drivers/gpu/drm/drm_mm.c 	list_del(&node->node_list);
node              587 drivers/gpu/drm/drm_mm.c 	node->allocated = false;
node              725 drivers/gpu/drm/drm_mm.c 			   struct drm_mm_node *node)
node              733 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(node->mm != mm);
node              734 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!node->allocated);
node              735 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(node->scanned_block);
node              736 drivers/gpu/drm/drm_mm.c 	node->scanned_block = true;
node              744 drivers/gpu/drm/drm_mm.c 	hole = list_prev_entry(node, node_list);
node              745 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(list_next_entry(hole, node_list) != node);
node              746 drivers/gpu/drm/drm_mm.c 	__list_del_entry(&node->node_list);
node              816 drivers/gpu/drm/drm_mm.c 			      struct drm_mm_node *node)
node              820 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(node->mm != scan->mm);
node              821 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!node->scanned_block);
node              822 drivers/gpu/drm/drm_mm.c 	node->scanned_block = false;
node              824 drivers/gpu/drm/drm_mm.c 	DRM_MM_BUG_ON(!node->mm->scan_active);
node              825 drivers/gpu/drm/drm_mm.c 	node->mm->scan_active--;
node              835 drivers/gpu/drm/drm_mm.c 	prev_node = list_prev_entry(node, node_list);
node              837 drivers/gpu/drm/drm_mm.c 		      list_next_entry(node, node_list));
node              838 drivers/gpu/drm/drm_mm.c 	list_add(&node->node_list, &prev_node->node_list);
node              840 drivers/gpu/drm/drm_mm.c 	return (node->start + node->size > scan->hit_start &&
node              841 drivers/gpu/drm/drm_mm.c 		node->start < scan->hit_end);
node               94 drivers/gpu/drm/drm_of.c 				struct device_node *node)
node               96 drivers/gpu/drm/drm_of.c 	of_node_get(node);
node               98 drivers/gpu/drm/drm_of.c 				    compare, node);
node              196 drivers/gpu/drm/drm_of.c int drm_of_encoder_active_endpoint(struct device_node *node,
node              205 drivers/gpu/drm/drm_of.c 	if (!node || !crtc)
node              208 drivers/gpu/drm/drm_of.c 	for_each_endpoint_of_node(node, ep) {
node              142 drivers/gpu/drm/drm_syncobj.c 	struct list_head node;
node              194 drivers/gpu/drm/drm_syncobj.c 		list_add_tail(&wait->node, &syncobj->cb_list);
node              206 drivers/gpu/drm/drm_syncobj.c 	if (!wait->node.next)
node              210 drivers/gpu/drm/drm_syncobj.c 	list_del_init(&wait->node);
node              242 drivers/gpu/drm/drm_syncobj.c 	list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node)
node              275 drivers/gpu/drm/drm_syncobj.c 		list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node)
node              370 drivers/gpu/drm/drm_syncobj.c 	if (wait.node.next)
node              866 drivers/gpu/drm/drm_syncobj.c 	list_del_init(&wait->node);
node              144 drivers/gpu/drm/drm_vma_manager.c 	struct drm_mm_node *node, *best;
node              152 drivers/gpu/drm/drm_vma_manager.c 		node = rb_entry(iter, struct drm_mm_node, rb);
node              153 drivers/gpu/drm/drm_vma_manager.c 		offset = node->start;
node              156 drivers/gpu/drm/drm_vma_manager.c 			best = node;
node              202 drivers/gpu/drm/drm_vma_manager.c 		       struct drm_vma_offset_node *node, unsigned long pages)
node              208 drivers/gpu/drm/drm_vma_manager.c 	if (!drm_mm_node_allocated(&node->vm_node))
node              210 drivers/gpu/drm/drm_vma_manager.c 					 &node->vm_node, pages);
node              230 drivers/gpu/drm/drm_vma_manager.c 			   struct drm_vma_offset_node *node)
node              234 drivers/gpu/drm/drm_vma_manager.c 	if (drm_mm_node_allocated(&node->vm_node)) {
node              235 drivers/gpu/drm/drm_vma_manager.c 		drm_mm_remove_node(&node->vm_node);
node              236 drivers/gpu/drm/drm_vma_manager.c 		memset(&node->vm_node, 0, sizeof(node->vm_node));
node              263 drivers/gpu/drm/drm_vma_manager.c int drm_vma_node_allow(struct drm_vma_offset_node *node, struct drm_file *tag)
node              276 drivers/gpu/drm/drm_vma_manager.c 	write_lock(&node->vm_lock);
node              278 drivers/gpu/drm/drm_vma_manager.c 	iter = &node->vm_files.rb_node;
node              302 drivers/gpu/drm/drm_vma_manager.c 	rb_insert_color(&new->vm_rb, &node->vm_files);
node              306 drivers/gpu/drm/drm_vma_manager.c 	write_unlock(&node->vm_lock);
node              325 drivers/gpu/drm/drm_vma_manager.c void drm_vma_node_revoke(struct drm_vma_offset_node *node,
node              331 drivers/gpu/drm/drm_vma_manager.c 	write_lock(&node->vm_lock);
node              333 drivers/gpu/drm/drm_vma_manager.c 	iter = node->vm_files.rb_node;
node              338 drivers/gpu/drm/drm_vma_manager.c 				rb_erase(&entry->vm_rb, &node->vm_files);
node              349 drivers/gpu/drm/drm_vma_manager.c 	write_unlock(&node->vm_lock);
node              366 drivers/gpu/drm/drm_vma_manager.c bool drm_vma_node_is_allowed(struct drm_vma_offset_node *node,
node              372 drivers/gpu/drm/drm_vma_manager.c 	read_lock(&node->vm_lock);
node              374 drivers/gpu/drm/drm_vma_manager.c 	iter = node->vm_files.rb_node;
node              385 drivers/gpu/drm/drm_vma_manager.c 	read_unlock(&node->vm_lock);
node              193 drivers/gpu/drm/etnaviv/etnaviv_drv.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              194 drivers/gpu/drm/etnaviv/etnaviv_drv.c 	struct drm_device *dev = node->minor->dev;
node              196 drivers/gpu/drm/etnaviv/etnaviv_drv.c 			node->info_ent->data;
node              203 drivers/gpu/drm/etnaviv/etnaviv_drv.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              204 drivers/gpu/drm/etnaviv/etnaviv_drv.c 	struct drm_device *dev = node->minor->dev;
node              208 drivers/gpu/drm/etnaviv/etnaviv_drv.c 			node->info_ent->data;
node              100 drivers/gpu/drm/etnaviv/etnaviv_gem.h 	struct list_head node; /* GPU active submit list */
node              140 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 				   struct drm_mm_node *node, size_t size)
node              154 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = drm_mm_insert_node_in_range(&context->mm, node,
node              224 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		   struct drm_mm_node *node, size_t size, u64 va)
node              226 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	return drm_mm_insert_node_in_range(&context->mm, node, size, 0, 0, va,
node              235 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	struct drm_mm_node *node;
node              256 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	node = &mapping->vram_node;
node              259 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_iommu_insert_exact(context, node,
node              262 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_iommu_find_iova(context, node,
node              267 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	mapping->iova = node->start;
node              268 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size,
node              272 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		drm_mm_remove_node(node);
node              378 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		struct drm_mm_node *node = &mapping->vram_node;
node              381 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_iommu_find_iova(context, node, size);
node              387 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		mapping->iova = node->start;
node              388 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_context_map(context, node->start, paddr, size,
node              391 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 			drm_mm_remove_node(node);
node              410 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	struct drm_mm_node *node = &mapping->vram_node;
node              420 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	etnaviv_context_unmap(context, node->start, node->size);
node              421 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	drm_mm_remove_node(node);
node             1654 drivers/gpu/drm/exynos/exynos_drm_dsi.c 	struct device_node *node = dev->of_node;
node             1657 drivers/gpu/drm/exynos/exynos_drm_dsi.c 	ret = exynos_dsi_of_read_u32(node, "samsung,pll-clock-frequency",
node             1662 drivers/gpu/drm/exynos/exynos_drm_dsi.c 	ret = exynos_dsi_of_read_u32(node, "samsung,burst-clock-frequency",
node             1667 drivers/gpu/drm/exynos/exynos_drm_dsi.c 	ret = exynos_dsi_of_read_u32(node, "samsung,esc-clock-frequency",
node             1672 drivers/gpu/drm/exynos/exynos_drm_dsi.c 	dsi->in_bridge_node = of_graph_get_remote_node(node, DSI_PORT_IN, 0);
node              271 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node;
node              287 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	node = kcalloc(G2D_CMDLIST_NUM, sizeof(*node), GFP_KERNEL);
node              288 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (!node) {
node              296 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		node[nr].cmdlist =
node              298 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		node[nr].dma_addr =
node              301 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		buf_info = &node[nr].buf_info;
node              305 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		list_add_tail(&node[nr].list, &g2d->free_cmdlist);
node              332 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node;
node              341 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	node = list_first_entry(&g2d->free_cmdlist, struct g2d_cmdlist_node,
node              343 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_del_init(&node->list);
node              346 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	return node;
node              349 drivers/gpu/drm/exynos/exynos_drm_g2d.c static void g2d_put_cmdlist(struct g2d_data *g2d, struct g2d_cmdlist_node *node)
node              352 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_move_tail(&node->list, &g2d->free_cmdlist);
node              357 drivers/gpu/drm/exynos/exynos_drm_g2d.c 				     struct g2d_cmdlist_node *node)
node              367 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	lnode->cmdlist->data[lnode->cmdlist->last] = node->dma_addr;
node              370 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_add_tail(&node->list, &file_priv->inuse_cmdlist);
node              372 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (node->event)
node              373 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		list_add_tail(&node->event->base.link, &file_priv->event_list);
node              687 drivers/gpu/drm/exynos/exynos_drm_g2d.c 				struct g2d_cmdlist_node *node,
node              691 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist *cmdlist = node->cmdlist;
node              692 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_buf_info *buf_info = &node->buf_info;
node              774 drivers/gpu/drm/exynos/exynos_drm_g2d.c 				  struct g2d_cmdlist_node *node,
node              777 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_buf_info *buf_info = &node->buf_info;
node              807 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node =
node              812 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	writel_relaxed(node->dma_addr, g2d->regs + G2D_DMA_SFR_BASE_ADDR);
node              832 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node;
node              839 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_for_each_entry(node, &runqueue_node->run_cmdlist, list)
node              840 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		g2d_unmap_cmdlist_gem(g2d, node, runqueue_node->filp);
node              856 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_runqueue_node *node, *n;
node              861 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_for_each_entry_safe(node, n, &g2d->runqueue, list) {
node              862 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		if (file && node->filp != file)
node              865 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		list_del_init(&node->list);
node              866 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		g2d_free_runqueue_node(g2d, node);
node             1021 drivers/gpu/drm/exynos/exynos_drm_g2d.c 				struct g2d_cmdlist_node *node,
node             1024 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist *cmdlist = node->cmdlist;
node             1030 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		struct g2d_buf_info *buf_info = &node->buf_info;
node             1145 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node;
node             1150 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	node = g2d_get_cmdlist(g2d);
node             1151 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (!node)
node             1165 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	node->event = NULL;
node             1168 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		e = kzalloc(sizeof(*node->event), GFP_KERNEL);
node             1184 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		node->event = e;
node             1187 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	cmdlist = node->cmdlist;
node             1212 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (node->event) {
node             1244 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	ret = g2d_check_reg_offset(g2d, node, req->cmd_nr, false);
node             1248 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	node->buf_info.map_nr = req->cmd_buf_nr;
node             1263 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		ret = g2d_check_reg_offset(g2d, node, req->cmd_buf_nr, true);
node             1267 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		ret = g2d_map_cmdlist_gem(g2d, node, drm_dev, file);
node             1281 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_add_cmdlist_to_inuse(file_priv, node);
node             1286 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_unmap_cmdlist_gem(g2d, node, file);
node             1288 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (node->event)
node             1289 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		drm_event_cancel_free(drm_dev, &node->event->base);
node             1291 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_put_cmdlist(g2d, node);
node             1361 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct g2d_cmdlist_node *node, *n;
node             1387 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	list_for_each_entry_safe(node, n, &file_priv->inuse_cmdlist, list) {
node             1388 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		g2d_unmap_cmdlist_gem(g2d, node, file);
node             1389 drivers/gpu/drm/exynos/exynos_drm_g2d.c 		list_move_tail(&node->list, &g2d->free_cmdlist);
node              442 drivers/gpu/drm/i915/display/intel_fbc.c 				      struct drm_mm_node *node,
node              467 drivers/gpu/drm/i915/display/intel_fbc.c 	ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size <<= 1,
node              478 drivers/gpu/drm/i915/display/intel_fbc.c 	ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size >>= 1,
node              240 drivers/gpu/drm/i915/display/intel_fbdev.c 		(unsigned long)(ggtt->gmadr.start + vma->node.start);
node              241 drivers/gpu/drm/i915/display/intel_fbdev.c 	info->fix.smem_len = vma->node.size;
node              250 drivers/gpu/drm/i915/display/intel_fbdev.c 	info->screen_size = vma->node.size;
node              219 drivers/gpu/drm/i915/gem/i915_gem_client_blt.c 					   batch->node.start, batch->node.size,
node              193 drivers/gpu/drm/i915/gem/i915_gem_domain.c 		if (!drm_mm_node_allocated(&vma->node))
node              288 drivers/gpu/drm/i915/gem/i915_gem_domain.c 			if (!drm_mm_node_allocated(&vma->node))
node              298 drivers/gpu/drm/i915/gem/i915_gem_domain.c 		vma->node.color = cache_level;
node              491 drivers/gpu/drm/i915/gem/i915_gem_domain.c 		if (!drm_mm_node_allocated(&vma->node))
node              246 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		struct drm_mm_node node; /** temporary GTT binding */
node              355 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (vma->node.size < entry->pad_to_size)
node              358 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->alignment && !IS_ALIGNED(vma->node.start, entry->alignment))
node              362 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	    vma->node.start != entry->offset)
node              366 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	    vma->node.start < BATCH_OFFSET_BIAS)
node              370 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	    (vma->node.start + vma->node.size - 1) >> 32)
node              388 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (vma->node.size)
node              389 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		pin_flags = vma->node.start;
node              547 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (entry->offset != vma->node.start) {
node              548 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			entry->offset = vma->node.start | UPDATE;
node              555 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (drm_mm_node_allocated(&vma->node))
node              615 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (entry->offset != vma->node.start) {
node              616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		entry->offset = vma->node.start | UPDATE;
node              812 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		GEM_BUG_ON(drm_mm_node_allocated(&vma->node) &&
node              894 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	return gen8_canonical_addr((int)reloc->delta + target->node.start);
node              908 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	cache->node.allocated = false;
node              964 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		i915_gem_object_finish_access((struct drm_i915_gem_object *)cache->node.mm);
node              971 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (cache->node.allocated) {
node              973 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					     cache->node.start,
node              974 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					     cache->node.size);
node              975 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			drm_mm_remove_node(&cache->node);
node              977 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			i915_vma_unpin((struct i915_vma *)cache->node.mm);
node             1005 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		cache->node.mm = (void *)obj;
node             1049 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			memset(&cache->node, 0, sizeof(cache->node));
node             1051 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 				(&ggtt->vm.mm, &cache->node,
node             1058 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			cache->node.start = vma->node.start;
node             1059 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			cache->node.mm = (void *)vma;
node             1063 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	offset = cache->node.start;
node             1064 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (cache->node.allocated) {
node             1191 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					batch->node.start, PAGE_SIZE,
node             1288 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		addr = gen8_canonical_addr(vma->node.start + offset);
node             1345 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	return target->node.start | UPDATE;
node             1408 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	    gen8_canonical_addr(target->node.start) == reloc->presumed_offset)
node             2006 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	batch_start = gen8_canonical_addr(eb->batch->node.start) +
node             2009 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	shadow_batch_start = gen8_canonical_addr(vma->node.start);
node             2098 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					eb->batch->node.start +
node              301 drivers/gpu/drm/i915/gem/i915_gem_mman.c 			       (ggtt->gmadr.start + vma->node.start) >> PAGE_SHIFT,
node               48 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c 	offset = vma->node.start;
node              184 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c 					batch->node.start, batch->node.size,
node              232 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c 	src_offset = src->node.start;
node              233 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c 	dst_offset = dst->node.start;
node              381 drivers/gpu/drm/i915/gem/i915_gem_object_blt.c 					batch->node.start, batch->node.size,
node               16 drivers/gpu/drm/i915/gem/i915_gem_pm.c 	struct llist_node *node, *next;
node               18 drivers/gpu/drm/i915/gem/i915_gem_pm.c 	llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) {
node               20 drivers/gpu/drm/i915/gem/i915_gem_pm.c 			container_of((struct list_head *)node,
node              444 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 		unsigned long count = vma->node.size >> PAGE_SHIFT;
node               29 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 					 struct drm_mm_node *node, u64 size,
node               42 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	ret = drm_mm_insert_node_in_range(&dev_priv->mm.stolen, node,
node               51 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 				struct drm_mm_node *node, u64 size,
node               54 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	return i915_gem_stolen_insert_node_in_range(dev_priv, node, size,
node               59 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 				 struct drm_mm_node *node)
node               62 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	drm_mm_remove_node(node);
node              677 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	ret = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
node              685 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node               16 drivers/gpu/drm/i915/gem/i915_gem_stolen.h 				struct drm_mm_node *node, u64 size,
node               19 drivers/gpu/drm/i915/gem/i915_gem_stolen.h 					 struct drm_mm_node *node, u64 size,
node               23 drivers/gpu/drm/i915/gem/i915_gem_stolen.h 				 struct drm_mm_node *node);
node              169 drivers/gpu/drm/i915/gem/i915_gem_tiling.c 	if (vma->node.size < size)
node              173 drivers/gpu/drm/i915/gem/i915_gem_tiling.c 	if (!IS_ALIGNED(vma->node.start, alignment))
node               24 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 	struct hlist_node node;
node               34 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 	struct hlist_node node;
node              335 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 	hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real)
node              378 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 			 &mm->node, (unsigned long)mm->mm);
node              403 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 	hash_del(&mm->node);
node              678 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 			if (!IS_ALIGNED(vma->node.start,
node              681 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 				       vma->node.start);
node              686 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 			if (!IS_ALIGNED(vma->node.size,
node              689 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 				       vma->node.size);
node              834 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 				if (!IS_ALIGNED(vma->node.start,
node              837 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 					       vma->node.start);
node              842 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 				if (!IS_ALIGNED(vma->node.size,
node              845 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 					       vma->node.size);
node              600 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	*cmd++ = lower_32_bits(vma->node.start);
node              601 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	*cmd++ = upper_32_bits(vma->node.start);
node              663 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 					batch->node.start, batch->node.size,
node             1151 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	struct drm_mm_node *node =
node             1154 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	if (!node || node->start > offset)
node             1157 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	GEM_BUG_ON(offset >= node->start + node->size);
node             1220 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	err = engine->emit_bb_start(rq, vma->node.start, vma->node.size, 0);
node             1319 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	err = engine->emit_bb_start(rq, vma->node.start, vma->node.size, 0);
node               62 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c 	GEM_BUG_ON(offset + (count - 1) * PAGE_SIZE > vma->node.size);
node               63 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c 	offset += vma->node.start;
node              136 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c 				    batch->node.start, batch->node.size,
node              125 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	list_for_each_entry_safe(cur, tmp, list, node) {
node              126 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 		INIT_LIST_HEAD(&cur->node);
node             1329 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		   rq->batch ? upper_32_bits(rq->batch->node.start) : ~0u,
node             1330 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		   rq->batch ? lower_32_bits(rq->batch->node.start) : ~0u);
node               35 drivers/gpu/drm/i915/gt/intel_engine_pool.c static void node_free(struct intel_engine_pool_node *node)
node               37 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_gem_object_put(node->obj);
node               38 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_active_fini(&node->active);
node               39 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	kfree(node);
node               44 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct intel_engine_pool_node *node =
node               45 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		container_of(ref, typeof(*node), active);
node               46 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct dma_resv *resv = node->obj->base.resv;
node               54 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	err = i915_gem_object_pin_pages(node->obj);
node               59 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_gem_object_make_unshrinkable(node->obj);
node               66 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct intel_engine_pool_node *node =
node               67 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		container_of(ref, typeof(*node), active);
node               68 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct intel_engine_pool *pool = node->pool;
node               69 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct list_head *list = bucket_for_size(pool, node->obj->base.size);
node               74 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_gem_object_unpin_pages(node->obj);
node               77 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_gem_object_make_purgeable(node->obj);
node               80 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	list_add(&node->link, list);
node               88 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct intel_engine_pool_node *node;
node               91 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	node = kmalloc(sizeof(*node),
node               93 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	if (!node)
node               96 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	node->pool = pool;
node               97 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	i915_active_init(engine->i915, &node->active, pool_active, pool_retire);
node              101 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		i915_active_fini(&node->active);
node              102 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		kfree(node);
node              108 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	node->obj = obj;
node              109 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	return node;
node              115 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	struct intel_engine_pool_node *node;
node              126 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	list_for_each_entry(node, list, link) {
node              127 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		if (node->obj->base.size < size)
node              129 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		list_del(&node->link);
node              134 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	if (&node->link == list) {
node              135 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		node = node_create(pool, size);
node              136 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		if (IS_ERR(node))
node              137 drivers/gpu/drm/i915/gt/intel_engine_pool.c 			return node;
node              140 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	ret = i915_active_acquire(&node->active);
node              142 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		node_free(node);
node              146 drivers/gpu/drm/i915/gt/intel_engine_pool.c 	return node;
node              164 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		struct intel_engine_pool_node *node, *nn;
node              166 drivers/gpu/drm/i915/gt/intel_engine_pool.c 		list_for_each_entry_safe(node, nn, list, link)
node              167 drivers/gpu/drm/i915/gt/intel_engine_pool.c 			node_free(node);
node               18 drivers/gpu/drm/i915/gt/intel_engine_pool.h intel_engine_pool_mark_active(struct intel_engine_pool_node *node,
node               21 drivers/gpu/drm/i915/gt/intel_engine_pool.h 	return i915_active_ref(&node->active, rq->timeline, rq);
node               25 drivers/gpu/drm/i915/gt/intel_engine_pool.h intel_engine_pool_put(struct intel_engine_pool_node *node)
node               27 drivers/gpu/drm/i915/gt/intel_engine_pool.h 	i915_active_release(&node->active);
node              266 drivers/gpu/drm/i915/gt/intel_lrc.c 	return rb_entry(rb, struct i915_priolist, node);
node             1334 drivers/gpu/drm/i915/gt/intel_lrc.c 		rb_erase_cached(&p->node, &execlists->queue);
node             1655 drivers/gpu/drm/i915/gt/intel_lrc.c 			  struct i915_sched_node *node,
node             1658 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(!list_empty(&node->link));
node             1659 drivers/gpu/drm/i915/gt/intel_lrc.c 	list_add_tail(&node->link, i915_sched_lookup_priolist(engine, prio));
node             2608 drivers/gpu/drm/i915/gt/intel_lrc.c 		rb_erase_cached(&p->node, &execlists->queue);
node             3430 drivers/gpu/drm/i915/gt/intel_lrc.c 		struct rb_node *node = &ve->nodes[sibling->id].rb;
node             3432 drivers/gpu/drm/i915/gt/intel_lrc.c 		if (RB_EMPTY_NODE(node))
node             3438 drivers/gpu/drm/i915/gt/intel_lrc.c 		if (!RB_EMPTY_NODE(node))
node             3439 drivers/gpu/drm/i915/gt/intel_lrc.c 			rb_erase_cached(node, &sibling->execlists.virtual);
node             3566 drivers/gpu/drm/i915/gt/intel_lrc.c 		struct ve_node * const node = &ve->nodes[sibling->id];
node             3571 drivers/gpu/drm/i915/gt/intel_lrc.c 			if (!RB_EMPTY_NODE(&node->rb)) {
node             3573 drivers/gpu/drm/i915/gt/intel_lrc.c 				rb_erase_cached(&node->rb,
node             3575 drivers/gpu/drm/i915/gt/intel_lrc.c 				RB_CLEAR_NODE(&node->rb);
node             3583 drivers/gpu/drm/i915/gt/intel_lrc.c 		if (!RB_EMPTY_NODE(&node->rb)) {
node             3589 drivers/gpu/drm/i915/gt/intel_lrc.c 				&node->rb;
node             3590 drivers/gpu/drm/i915/gt/intel_lrc.c 			if (prio == node->prio || (prio > node->prio && first))
node             3593 drivers/gpu/drm/i915/gt/intel_lrc.c 			rb_erase_cached(&node->rb, &sibling->execlists.virtual);
node             3612 drivers/gpu/drm/i915/gt/intel_lrc.c 		rb_link_node(&node->rb, rb, parent);
node             3613 drivers/gpu/drm/i915/gt/intel_lrc.c 		rb_insert_color_cached(&node->rb,
node             3618 drivers/gpu/drm/i915/gt/intel_lrc.c 		GEM_BUG_ON(RB_EMPTY_NODE(&node->rb));
node             3619 drivers/gpu/drm/i915/gt/intel_lrc.c 		node->prio = prio;
node             3952 drivers/gpu/drm/i915/gt/intel_lrc.c 		struct i915_priolist *p = rb_entry(rb, typeof(*p), node);
node               96 drivers/gpu/drm/i915/gt/intel_renderstate.c 			u64 r = s + so->vma->node.start;
node              628 drivers/gpu/drm/i915/gt/intel_reset.c 		struct drm_vma_offset_node *node;
node              640 drivers/gpu/drm/i915/gt/intel_reset.c 		node = &vma->obj->base.vma_node;
node              643 drivers/gpu/drm/i915/gt/intel_reset.c 				    drm_vma_node_offset_addr(node) + vma_offset,
node              111 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	return hws->node.start + offset_in_page(sizeof(u32)*rq->fence.context);
node              202 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = lower_32_bits(vma->node.start);
node              203 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = upper_32_bits(vma->node.start);
node              216 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = lower_32_bits(vma->node.start);
node              229 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = lower_32_bits(vma->node.start);
node              241 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = lower_32_bits(vma->node.start);
node              256 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags);
node             1131 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	struct drm_mm_node evict = arg->vma->node;
node             1467 drivers/gpu/drm/i915/gt/selftest_lrc.c 							vma->node.start,
node              483 drivers/gpu/drm/i915/gt/selftest_workarounds.c 		u64 addr = scratch->node.start;
node              569 drivers/gpu/drm/i915/gt/selftest_workarounds.c 					    batch->node.start, PAGE_SIZE,
node              786 drivers/gpu/drm/i915/gt/selftest_workarounds.c 		u64 offset = results->node.start + sizeof(u32) * i;
node              854 drivers/gpu/drm/i915/gt/selftest_workarounds.c 	err = engine->emit_bb_start(rq, batch->node.start, 0, 0);
node               77 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	return rb_entry(rb, struct i915_priolist, node);
node              583 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		rb_erase_cached(&p->node, &execlists->queue);
node              727 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		rb_erase_cached(&p->node, &execlists->queue);
node              401 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	struct drm_mm_node *node = &ggtt->uc_fw;
node              403 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	GEM_BUG_ON(!node->allocated);
node              404 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	GEM_BUG_ON(upper_32_bits(node->start));
node              405 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	GEM_BUG_ON(upper_32_bits(node->start + node->size - 1));
node              407 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	return lower_32_bits(node->start);
node              416 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 		.node.start = uc_fw_ggtt_offset(uc_fw, ggtt),
node              417 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 		.node.size = obj->base.size,
node              423 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	GEM_BUG_ON(dummy.node.size > ggtt->uc_fw.size);
node               47 drivers/gpu/drm/i915/gvt/aperture_gm.c 	struct drm_mm_node *node;
node               51 drivers/gpu/drm/i915/gvt/aperture_gm.c 		node = &vgpu->gm.high_gm_node;
node               57 drivers/gpu/drm/i915/gvt/aperture_gm.c 		node = &vgpu->gm.low_gm_node;
node               66 drivers/gpu/drm/i915/gvt/aperture_gm.c 	ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node,
node               36 drivers/gpu/drm/i915/gvt/debugfs.c 	struct list_head node;
node               49 drivers/gpu/drm/i915/gvt/debugfs.c 	ma = container_of(a, struct diff_mmio, node);
node               50 drivers/gpu/drm/i915/gvt/debugfs.c 	mb = container_of(b, struct diff_mmio, node);
node               63 drivers/gpu/drm/i915/gvt/debugfs.c 	struct diff_mmio *node;
node               70 drivers/gpu/drm/i915/gvt/debugfs.c 		node = kmalloc(sizeof(*node), GFP_KERNEL);
node               71 drivers/gpu/drm/i915/gvt/debugfs.c 		if (!node)
node               74 drivers/gpu/drm/i915/gvt/debugfs.c 		node->offset = offset;
node               75 drivers/gpu/drm/i915/gvt/debugfs.c 		node->preg = preg;
node               76 drivers/gpu/drm/i915/gvt/debugfs.c 		node->vreg = vreg;
node               77 drivers/gpu/drm/i915/gvt/debugfs.c 		list_add(&node->node, &param->diff_mmio_list);
node               94 drivers/gpu/drm/i915/gvt/debugfs.c 	struct diff_mmio *node, *next;
node              113 drivers/gpu/drm/i915/gvt/debugfs.c 	list_for_each_entry_safe(node, next, &param.diff_mmio_list, node) {
node              114 drivers/gpu/drm/i915/gvt/debugfs.c 		u32 diff = node->preg ^ node->vreg;
node              117 drivers/gpu/drm/i915/gvt/debugfs.c 			   node->offset, node->preg, node->vreg,
node              119 drivers/gpu/drm/i915/gvt/debugfs.c 		list_del(&node->node);
node              120 drivers/gpu/drm/i915/gvt/debugfs.c 		kfree(node);
node               89 drivers/gpu/drm/i915/gvt/handlers.c 	hash_for_each_possible(gvt->mmio.mmio_info_table, e, node, offset) {
node              137 drivers/gpu/drm/i915/gvt/handlers.c 		INIT_HLIST_NODE(&info->node);
node              138 drivers/gpu/drm/i915/gvt/handlers.c 		hash_add(gvt->mmio.mmio_info_table, &info->node, info->offset);
node             3319 drivers/gpu/drm/i915/gvt/handlers.c 	hash_for_each_safe(gvt->mmio.mmio_info_table, i, tmp, e, node)
node             3413 drivers/gpu/drm/i915/gvt/handlers.c 	hash_for_each(gvt->mmio.mmio_info_table, i, e, node) {
node              222 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct rb_node *node = vgpu->vdev.dma_addr_cache.rb_node;
node              225 drivers/gpu/drm/i915/gvt/kvmgt.c 	while (node) {
node              226 drivers/gpu/drm/i915/gvt/kvmgt.c 		itr = rb_entry(node, struct gvt_dma, dma_addr_node);
node              229 drivers/gpu/drm/i915/gvt/kvmgt.c 			node = node->rb_left;
node              231 drivers/gpu/drm/i915/gvt/kvmgt.c 			node = node->rb_right;
node              240 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct rb_node *node = vgpu->vdev.gfn_cache.rb_node;
node              243 drivers/gpu/drm/i915/gvt/kvmgt.c 	while (node) {
node              244 drivers/gpu/drm/i915/gvt/kvmgt.c 		itr = rb_entry(node, struct gvt_dma, gfn_node);
node              247 drivers/gpu/drm/i915/gvt/kvmgt.c 			node = node->rb_left;
node              249 drivers/gpu/drm/i915/gvt/kvmgt.c 			node = node->rb_right;
node              317 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct rb_node *node = NULL;
node              321 drivers/gpu/drm/i915/gvt/kvmgt.c 		node = rb_first(&vgpu->vdev.gfn_cache);
node              322 drivers/gpu/drm/i915/gvt/kvmgt.c 		if (!node) {
node              326 drivers/gpu/drm/i915/gvt/kvmgt.c 		dma = rb_entry(node, struct gvt_dma, gfn_node);
node             1704 drivers/gpu/drm/i915/gvt/kvmgt.c 		struct kvm_page_track_notifier_node *node)
node             1706 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct kvmgt_guest_info *info = container_of(node,
node             1716 drivers/gpu/drm/i915/gvt/kvmgt.c 		struct kvm_page_track_notifier_node *node)
node             1720 drivers/gpu/drm/i915/gvt/kvmgt.c 	struct kvmgt_guest_info *info = container_of(node,
node               67 drivers/gpu/drm/i915/gvt/mmio.h 	struct hlist_node node;
node               32 drivers/gpu/drm/i915/i915_active.c 	struct rb_node node;
node               49 drivers/gpu/drm/i915/i915_active.c static inline struct llist_node *barrier_to_ll(struct active_node *node)
node               51 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(!is_barrier(&node->base));
node               52 drivers/gpu/drm/i915/i915_active.c 	return (struct llist_node *)&node->base.link;
node               56 drivers/gpu/drm/i915/i915_active.c __barrier_to_engine(struct active_node *node)
node               58 drivers/gpu/drm/i915/i915_active.c 	return (struct intel_engine_cs *)READ_ONCE(node->base.link.prev);
node               62 drivers/gpu/drm/i915/i915_active.c barrier_to_engine(struct active_node *node)
node               64 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(!is_barrier(&node->base));
node               65 drivers/gpu/drm/i915/i915_active.c 	return __barrier_to_engine(node);
node              146 drivers/gpu/drm/i915/i915_active.c 	rbtree_postorder_for_each_entry_safe(it, n, &root, node) {
node              184 drivers/gpu/drm/i915/i915_active.c 	struct active_node *node, *prealloc;
node              195 drivers/gpu/drm/i915/i915_active.c 	node = READ_ONCE(ref->cache);
node              196 drivers/gpu/drm/i915/i915_active.c 	if (node && node->timeline == idx)
node              197 drivers/gpu/drm/i915/i915_active.c 		return &node->base;
node              212 drivers/gpu/drm/i915/i915_active.c 		node = rb_entry(parent, struct active_node, node);
node              213 drivers/gpu/drm/i915/i915_active.c 		if (node->timeline == idx) {
node              218 drivers/gpu/drm/i915/i915_active.c 		if (node->timeline < idx)
node              224 drivers/gpu/drm/i915/i915_active.c 	node = prealloc;
node              225 drivers/gpu/drm/i915/i915_active.c 	i915_active_request_init(&node->base, &tl->mutex, NULL, node_retire);
node              226 drivers/gpu/drm/i915/i915_active.c 	node->ref = ref;
node              227 drivers/gpu/drm/i915/i915_active.c 	node->timeline = idx;
node              229 drivers/gpu/drm/i915/i915_active.c 	rb_link_node(&node->node, parent, p);
node              230 drivers/gpu/drm/i915/i915_active.c 	rb_insert_color(&node->node, &ref->tree);
node              233 drivers/gpu/drm/i915/i915_active.c 	ref->cache = node;
node              236 drivers/gpu/drm/i915/i915_active.c 	BUILD_BUG_ON(offsetof(typeof(*node), base));
node              237 drivers/gpu/drm/i915/i915_active.c 	return &node->base;
node              260 drivers/gpu/drm/i915/i915_active.c 				   struct active_node *node,
node              267 drivers/gpu/drm/i915/i915_active.c 	GEM_BUG_ON(node->timeline != engine->kernel_context->timeline->fence_context);
node              285 drivers/gpu/drm/i915/i915_active.c 		if (node == barrier_from_ll(pos)) {
node              286 drivers/gpu/drm/i915/i915_active.c 			node = NULL;
node              298 drivers/gpu/drm/i915/i915_active.c 	return !node;
node              302 drivers/gpu/drm/i915/i915_active.c __active_del_barrier(struct i915_active *ref, struct active_node *node)
node              304 drivers/gpu/drm/i915/i915_active.c 	return ____active_del_barrier(ref, node, barrier_to_engine(node));
node              426 drivers/gpu/drm/i915/i915_active.c 	rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
node              474 drivers/gpu/drm/i915/i915_active.c 	rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
node              495 drivers/gpu/drm/i915/i915_active.c static inline bool is_idle_barrier(struct active_node *node, u64 idx)
node              497 drivers/gpu/drm/i915/i915_active.c 	return node->timeline == idx && !i915_active_request_isset(&node->base);
node              518 drivers/gpu/drm/i915/i915_active.c 		p = &ref->cache->node;
node              525 drivers/gpu/drm/i915/i915_active.c 		struct active_node *node =
node              526 drivers/gpu/drm/i915/i915_active.c 			rb_entry(p, struct active_node, node);
node              528 drivers/gpu/drm/i915/i915_active.c 		if (is_idle_barrier(node, idx))
node              532 drivers/gpu/drm/i915/i915_active.c 		if (node->timeline < idx)
node              545 drivers/gpu/drm/i915/i915_active.c 		struct active_node *node =
node              546 drivers/gpu/drm/i915/i915_active.c 			rb_entry(p, struct active_node, node);
node              549 drivers/gpu/drm/i915/i915_active.c 		if (node->timeline > idx)
node              552 drivers/gpu/drm/i915/i915_active.c 		if (node->timeline < idx)
node              555 drivers/gpu/drm/i915/i915_active.c 		if (is_idle_barrier(node, idx))
node              565 drivers/gpu/drm/i915/i915_active.c 		engine = __barrier_to_engine(node);
node              567 drivers/gpu/drm/i915/i915_active.c 		if (is_barrier(&node->base) &&
node              568 drivers/gpu/drm/i915/i915_active.c 		    ____active_del_barrier(ref, node, engine))
node              578 drivers/gpu/drm/i915/i915_active.c 	if (p == &ref->cache->node)
node              582 drivers/gpu/drm/i915/i915_active.c 	return rb_entry(p, struct active_node, node);
node              603 drivers/gpu/drm/i915/i915_active.c 		struct active_node *node;
node              605 drivers/gpu/drm/i915/i915_active.c 		node = reuse_idle_barrier(ref, idx);
node              606 drivers/gpu/drm/i915/i915_active.c 		if (!node) {
node              607 drivers/gpu/drm/i915/i915_active.c 			node = kmem_cache_alloc(global.slab_cache, GFP_KERNEL);
node              608 drivers/gpu/drm/i915/i915_active.c 			if (!node) {
node              614 drivers/gpu/drm/i915/i915_active.c 			node->base.lock =
node              617 drivers/gpu/drm/i915/i915_active.c 			RCU_INIT_POINTER(node->base.request, NULL);
node              618 drivers/gpu/drm/i915/i915_active.c 			node->base.retire = node_retire;
node              619 drivers/gpu/drm/i915/i915_active.c 			node->timeline = idx;
node              620 drivers/gpu/drm/i915/i915_active.c 			node->ref = ref;
node              623 drivers/gpu/drm/i915/i915_active.c 		if (!i915_active_request_isset(&node->base)) {
node              633 drivers/gpu/drm/i915/i915_active.c 			RCU_INIT_POINTER(node->base.request, ERR_PTR(-EAGAIN));
node              634 drivers/gpu/drm/i915/i915_active.c 			node->base.link.prev = (void *)engine;
node              638 drivers/gpu/drm/i915/i915_active.c 		GEM_BUG_ON(barrier_to_engine(node) != engine);
node              639 drivers/gpu/drm/i915/i915_active.c 		llist_add(barrier_to_ll(node), &ref->preallocated_barriers);
node              647 drivers/gpu/drm/i915/i915_active.c 		struct active_node *node = barrier_from_ll(pos);
node              650 drivers/gpu/drm/i915/i915_active.c 		intel_engine_pm_put(barrier_to_engine(node));
node              652 drivers/gpu/drm/i915/i915_active.c 		kmem_cache_free(global.slab_cache, node);
node              671 drivers/gpu/drm/i915/i915_active.c 		struct active_node *node = barrier_from_ll(pos);
node              672 drivers/gpu/drm/i915/i915_active.c 		struct intel_engine_cs *engine = barrier_to_engine(node);
node              682 drivers/gpu/drm/i915/i915_active.c 			it = rb_entry(parent, struct active_node, node);
node              683 drivers/gpu/drm/i915/i915_active.c 			if (it->timeline < node->timeline)
node              688 drivers/gpu/drm/i915/i915_active.c 		rb_link_node(&node->node, parent, p);
node              689 drivers/gpu/drm/i915/i915_active.c 		rb_insert_color(&node->node, &ref->tree);
node              691 drivers/gpu/drm/i915/i915_active.c 		llist_add(barrier_to_ll(node), &engine->barrier_tasks);
node              700 drivers/gpu/drm/i915/i915_active.c 	struct llist_node *node, *next;
node              710 drivers/gpu/drm/i915/i915_active.c 	llist_for_each_safe(node, next, llist_del_all(&engine->barrier_tasks)) {
node              711 drivers/gpu/drm/i915/i915_active.c 		RCU_INIT_POINTER(barrier_from_ll(node)->base.request, rq);
node              713 drivers/gpu/drm/i915/i915_active.c 		list_add_tail((struct list_head *)node, &rq->active_list);
node              861 drivers/gpu/drm/i915/i915_cmd_parser.c 	struct hlist_node node;
node              908 drivers/gpu/drm/i915/i915_cmd_parser.c 			hash_add(engine->cmd_hash, &desc_node->node,
node              922 drivers/gpu/drm/i915/i915_cmd_parser.c 	hash_for_each_safe(engine->cmd_hash, i, tmp, desc_node, node) {
node              923 drivers/gpu/drm/i915/i915_cmd_parser.c 		hash_del(&desc_node->node);
node             1053 drivers/gpu/drm/i915/i915_cmd_parser.c 	hash_for_each_possible(engine->cmd_hash, desc_node, node,
node               54 drivers/gpu/drm/i915/i915_debugfs.c static inline struct drm_i915_private *node_to_i915(struct drm_info_node *node)
node               56 drivers/gpu/drm/i915/i915_debugfs.c 	return to_i915(node->minor->dev);
node              160 drivers/gpu/drm/i915/i915_debugfs.c 		if (!drm_mm_node_allocated(&vma->node))
node              170 drivers/gpu/drm/i915/i915_debugfs.c 			   vma->node.start, vma->node.size,
node              254 drivers/gpu/drm/i915/i915_debugfs.c 			if (!drm_mm_node_allocated(&vma->node))
node              258 drivers/gpu/drm/i915/i915_debugfs.c 				stats->active += vma->node.size;
node              260 drivers/gpu/drm/i915/i915_debugfs.c 				stats->inactive += vma->node.size;
node              263 drivers/gpu/drm/i915/i915_debugfs.c 				stats->closed += vma->node.size;
node              274 drivers/gpu/drm/i915/i915_debugfs.c 				if (drm_mm_node_allocated(&vma->node)) {
node              276 drivers/gpu/drm/i915/i915_debugfs.c 						stats->active += vma->node.size;
node              278 drivers/gpu/drm/i915/i915_debugfs.c 						stats->inactive += vma->node.size;
node              281 drivers/gpu/drm/i915/i915_debugfs.c 						stats->closed += vma->node.size;
node             1967 drivers/gpu/drm/i915/i915_debugfs.c 	struct drm_info_node *node = m->private;
node             1968 drivers/gpu/drm/i915/i915_debugfs.c 	struct drm_i915_private *dev_priv = node_to_i915(node);
node             1969 drivers/gpu/drm/i915/i915_debugfs.c 	bool dump_load_err = !!node->info_ent->data;
node             2386 drivers/gpu/drm/i915/i915_drv.h 					 struct drm_mm_node *node,
node               66 drivers/gpu/drm/i915/i915_gem.c                      struct drm_mm_node *node, u32 size)
node               68 drivers/gpu/drm/i915/i915_gem.c 	memset(node, 0, sizeof(*node));
node               69 drivers/gpu/drm/i915/i915_gem.c 	return drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
node               76 drivers/gpu/drm/i915/i915_gem.c remove_mappable_node(struct drm_mm_node *node)
node               78 drivers/gpu/drm/i915/i915_gem.c 	drm_mm_remove_node(node);
node               95 drivers/gpu/drm/i915/i915_gem.c 			pinned += vma->node.size;
node              334 drivers/gpu/drm/i915/i915_gem.c 	struct drm_mm_node node;
node              353 drivers/gpu/drm/i915/i915_gem.c 		node.start = i915_ggtt_offset(vma);
node              354 drivers/gpu/drm/i915/i915_gem.c 		node.allocated = false;
node              356 drivers/gpu/drm/i915/i915_gem.c 		ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
node              359 drivers/gpu/drm/i915/i915_gem.c 		GEM_BUG_ON(!node.allocated);
node              392 drivers/gpu/drm/i915/i915_gem.c 		u32 page_base = node.start;
node              396 drivers/gpu/drm/i915/i915_gem.c 		if (node.allocated) {
node              399 drivers/gpu/drm/i915/i915_gem.c 					     node.start, I915_CACHE_NONE, 0);
node              418 drivers/gpu/drm/i915/i915_gem.c 	if (node.allocated) {
node              419 drivers/gpu/drm/i915/i915_gem.c 		ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
node              420 drivers/gpu/drm/i915/i915_gem.c 		remove_mappable_node(&node);
node              527 drivers/gpu/drm/i915/i915_gem.c 	struct drm_mm_node node;
node              563 drivers/gpu/drm/i915/i915_gem.c 		node.start = i915_ggtt_offset(vma);
node              564 drivers/gpu/drm/i915/i915_gem.c 		node.allocated = false;
node              566 drivers/gpu/drm/i915/i915_gem.c 		ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
node              569 drivers/gpu/drm/i915/i915_gem.c 		GEM_BUG_ON(!node.allocated);
node              603 drivers/gpu/drm/i915/i915_gem.c 		u32 page_base = node.start;
node              607 drivers/gpu/drm/i915/i915_gem.c 		if (node.allocated) {
node              612 drivers/gpu/drm/i915/i915_gem.c 					     node.start, I915_CACHE_NONE, 0);
node              639 drivers/gpu/drm/i915/i915_gem.c 	if (node.allocated) {
node              640 drivers/gpu/drm/i915/i915_gem.c 		ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
node              641 drivers/gpu/drm/i915/i915_gem.c 		remove_mappable_node(&node);
node               65 drivers/gpu/drm/i915/i915_gem_evict.c 	return drm_mm_scan_add_block(scan, &vma->node);
node              102 drivers/gpu/drm/i915/i915_gem_evict.c 	struct drm_mm_node *node;
node              181 drivers/gpu/drm/i915/i915_gem_evict.c 		ret = drm_mm_scan_remove_block(&scan, &vma->node);
node              227 drivers/gpu/drm/i915/i915_gem_evict.c 		if (drm_mm_scan_remove_block(&scan, &vma->node))
node              241 drivers/gpu/drm/i915/i915_gem_evict.c 	while (ret == 0 && (node = drm_mm_scan_color_evict(&scan))) {
node              242 drivers/gpu/drm/i915/i915_gem_evict.c 		vma = container_of(node, struct i915_vma, node);
node              265 drivers/gpu/drm/i915/i915_gem_evict.c 	struct drm_mm_node *node;
node              297 drivers/gpu/drm/i915/i915_gem_evict.c 	drm_mm_for_each_node_in_range(node, &vm->mm, start, end) {
node              299 drivers/gpu/drm/i915/i915_gem_evict.c 		if (node->color == I915_COLOR_UNEVICTABLE) {
node              304 drivers/gpu/drm/i915/i915_gem_evict.c 		GEM_BUG_ON(!node->allocated);
node              305 drivers/gpu/drm/i915/i915_gem_evict.c 		vma = container_of(node, typeof(*vma), node);
node              314 drivers/gpu/drm/i915/i915_gem_evict.c 			if (node->start + node->size == target->start) {
node              315 drivers/gpu/drm/i915/i915_gem_evict.c 				if (node->color == target->color)
node              318 drivers/gpu/drm/i915/i915_gem_evict.c 			if (node->start == target->start + target->size) {
node              319 drivers/gpu/drm/i915/i915_gem_evict.c 				if (node->color == target->color)
node               85 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(!IS_ALIGNED(vma->node.start, I965_FENCE_PAGE));
node               89 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		val = (vma->node.start + vma->fence_size - I965_FENCE_PAGE) << 32;
node               90 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		val |= vma->node.start;
node              131 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(vma->node.start & ~I915_FENCE_START_MASK);
node              133 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(!IS_ALIGNED(vma->node.start, vma->fence_size));
node              141 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		val = vma->node.start;
node              169 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(vma->node.start & ~I830_FENCE_START_MASK);
node              172 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		GEM_BUG_ON(!IS_ALIGNED(vma->node.start, vma->fence_size));
node              174 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		val = vma->node.start;
node              154 drivers/gpu/drm/i915/i915_gem_gtt.c 						 vma->node.start, vma->size);
node              171 drivers/gpu/drm/i915/i915_gem_gtt.c 	vma->vm->clear_range(vma->vm, vma->node.start, vma->size);
node             1220 drivers/gpu/drm/i915/i915_gem_gtt.c 	u64 start = vma->node.start;
node             1300 drivers/gpu/drm/i915/i915_gem_gtt.c 		      !iter->sg && IS_ALIGNED(vma->node.start +
node             1301 drivers/gpu/drm/i915/i915_gem_gtt.c 					      vma->node.size,
node             1345 drivers/gpu/drm/i915/i915_gem_gtt.c 		u64 idx = vma->node.start >> GEN8_PTE_SHIFT;
node             1650 drivers/gpu/drm/i915/i915_gem_gtt.c 	unsigned first_entry = vma->node.start / I915_GTT_PAGE_SIZE;
node             2206 drivers/gpu/drm/i915/i915_gem_gtt.c 	gtt_entries += vma->node.start / I915_GTT_PAGE_SIZE;
node             2245 drivers/gpu/drm/i915/i915_gem_gtt.c 	unsigned int i = vma->node.start / I915_GTT_PAGE_SIZE;
node             2419 drivers/gpu/drm/i915/i915_gem_gtt.c 	intel_gtt_insert_sg_entries(vma->pages, vma->node.start >> PAGE_SHIFT,
node             2464 drivers/gpu/drm/i915/i915_gem_gtt.c 		vma->vm->clear_range(vma->vm, vma->node.start, vma->size);
node             2485 drivers/gpu/drm/i915/i915_gem_gtt.c 							  vma->node.start,
node             2516 drivers/gpu/drm/i915/i915_gem_gtt.c 			vm->clear_range(vm, vma->node.start, vma->size);
node             2523 drivers/gpu/drm/i915/i915_gem_gtt.c 		vm->clear_range(vm, vma->node.start, vma->size);
node             2560 drivers/gpu/drm/i915/i915_gem_gtt.c static void i915_gtt_color_adjust(const struct drm_mm_node *node,
node             2565 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (node->allocated && node->color != color)
node             2573 drivers/gpu/drm/i915/i915_gem_gtt.c 	node = list_next_entry(node, node_list);
node             2574 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (node->color != color)
node             3632 drivers/gpu/drm/i915/i915_gem_gtt.c 			 struct drm_mm_node *node,
node             3643 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(drm_mm_node_allocated(node));
node             3645 drivers/gpu/drm/i915/i915_gem_gtt.c 	node->size = size;
node             3646 drivers/gpu/drm/i915/i915_gem_gtt.c 	node->start = offset;
node             3647 drivers/gpu/drm/i915/i915_gem_gtt.c 	node->color = color;
node             3649 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = drm_mm_reserve_node(&vm->mm, node);
node             3656 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = i915_gem_evict_for_node(vm, node, flags);
node             3658 drivers/gpu/drm/i915/i915_gem_gtt.c 		err = drm_mm_reserve_node(&vm->mm, node);
node             3723 drivers/gpu/drm/i915/i915_gem_gtt.c 			struct drm_mm_node *node,
node             3740 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(drm_mm_node_allocated(node));
node             3764 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = drm_mm_insert_node_in_range(&vm->mm, node,
node             3771 drivers/gpu/drm/i915/i915_gem_gtt.c 		err = drm_mm_insert_node_in_range(&vm->mm, node,
node             3807 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = i915_gem_gtt_reserve(vm, node, size, offset, color, flags);
node             3820 drivers/gpu/drm/i915/i915_gem_gtt.c 	return drm_mm_insert_node_in_range(&vm->mm, node,
node              593 drivers/gpu/drm/i915/i915_gem_gtt.h 			 struct drm_mm_node *node,
node              598 drivers/gpu/drm/i915/i915_gem_gtt.h 			struct drm_mm_node *node,
node              985 drivers/gpu/drm/i915/i915_gpu_error.c 	dst->gtt_offset = vma->node.start;
node              986 drivers/gpu/drm/i915/i915_gpu_error.c 	dst->gtt_size = vma->node.size;
node             1350 drivers/gpu/drm/i915/i915_gpu_error.c 			.node = { .start = U64_MAX, .size = obj->base.size },
node              986 drivers/gpu/drm/i915/i915_pmu.c static int i915_pmu_cpu_online(unsigned int cpu, struct hlist_node *node)
node              988 drivers/gpu/drm/i915/i915_pmu.c 	struct i915_pmu *pmu = hlist_entry_safe(node, typeof(*pmu), node);
node              999 drivers/gpu/drm/i915/i915_pmu.c static int i915_pmu_cpu_offline(unsigned int cpu, struct hlist_node *node)
node             1001 drivers/gpu/drm/i915/i915_pmu.c 	struct i915_pmu *pmu = hlist_entry_safe(node, typeof(*pmu), node);
node             1033 drivers/gpu/drm/i915/i915_pmu.c 	ret = cpuhp_state_add_instance(slot, &pmu->node);
node             1046 drivers/gpu/drm/i915/i915_pmu.c 	WARN_ON(cpuhp_state_remove_instance(cpuhp_slot, &pmu->node));
node               44 drivers/gpu/drm/i915/i915_pmu.h 	struct hlist_node node;
node               47 drivers/gpu/drm/i915/i915_priolist_types.h 	struct rb_node node;
node               23 drivers/gpu/drm/i915/i915_scheduler.c node_to_request(const struct i915_sched_node *node)
node               25 drivers/gpu/drm/i915/i915_scheduler.c 	return container_of(node, const struct i915_request, sched);
node               28 drivers/gpu/drm/i915/i915_scheduler.c static inline bool node_started(const struct i915_sched_node *node)
node               30 drivers/gpu/drm/i915/i915_scheduler.c 	return i915_request_started(node_to_request(node));
node               33 drivers/gpu/drm/i915/i915_scheduler.c static inline bool node_signaled(const struct i915_sched_node *node)
node               35 drivers/gpu/drm/i915/i915_scheduler.c 	return i915_request_completed(node_to_request(node));
node               40 drivers/gpu/drm/i915/i915_scheduler.c 	return rb_entry(rb, struct i915_priolist, node);
node              130 drivers/gpu/drm/i915/i915_scheduler.c 	rb_link_node(&p->node, rb, parent);
node              131 drivers/gpu/drm/i915/i915_scheduler.c 	rb_insert_color_cached(&p->node, &execlists->queue, first);
node              149 drivers/gpu/drm/i915/i915_scheduler.c sched_lock_engine(const struct i915_sched_node *node,
node              153 drivers/gpu/drm/i915/i915_scheduler.c 	const struct i915_request *rq = node_to_request(node);
node              230 drivers/gpu/drm/i915/i915_scheduler.c static void __i915_schedule(struct i915_sched_node *node,
node              244 drivers/gpu/drm/i915/i915_scheduler.c 	if (prio <= READ_ONCE(node->attr.priority))
node              247 drivers/gpu/drm/i915/i915_scheduler.c 	if (node_signaled(node))
node              250 drivers/gpu/drm/i915/i915_scheduler.c 	stack.signaler = node;
node              272 drivers/gpu/drm/i915/i915_scheduler.c 		struct i915_sched_node *node = dep->signaler;
node              275 drivers/gpu/drm/i915/i915_scheduler.c 		if (node_started(node))
node              284 drivers/gpu/drm/i915/i915_scheduler.c 		list_for_each_entry(p, &node->signalers_list, signal_link) {
node              301 drivers/gpu/drm/i915/i915_scheduler.c 	if (node->attr.priority == I915_PRIORITY_INVALID) {
node              302 drivers/gpu/drm/i915/i915_scheduler.c 		GEM_BUG_ON(!list_empty(&node->link));
node              303 drivers/gpu/drm/i915/i915_scheduler.c 		node->attr = *attr;
node              312 drivers/gpu/drm/i915/i915_scheduler.c 	engine = node_to_request(node)->engine;
node              316 drivers/gpu/drm/i915/i915_scheduler.c 	engine = sched_lock_engine(node, engine, &cache);
node              320 drivers/gpu/drm/i915/i915_scheduler.c 		node = dep->signaler;
node              321 drivers/gpu/drm/i915/i915_scheduler.c 		engine = sched_lock_engine(node, engine, &cache);
node              325 drivers/gpu/drm/i915/i915_scheduler.c 		if (prio <= node->attr.priority || node_signaled(node))
node              328 drivers/gpu/drm/i915/i915_scheduler.c 		GEM_BUG_ON(node_to_request(node)->engine != engine);
node              330 drivers/gpu/drm/i915/i915_scheduler.c 		node->attr.priority = prio;
node              332 drivers/gpu/drm/i915/i915_scheduler.c 		if (list_empty(&node->link)) {
node              345 drivers/gpu/drm/i915/i915_scheduler.c 		    !i915_request_is_active(node_to_request(node))) {
node              350 drivers/gpu/drm/i915/i915_scheduler.c 			list_move_tail(&node->link, cache.priolist);
node              354 drivers/gpu/drm/i915/i915_scheduler.c 		kick_submission(engine, node_to_request(node), prio);
node              367 drivers/gpu/drm/i915/i915_scheduler.c static void __bump_priority(struct i915_sched_node *node, unsigned int bump)
node              369 drivers/gpu/drm/i915/i915_scheduler.c 	struct i915_sched_attr attr = node->attr;
node              372 drivers/gpu/drm/i915/i915_scheduler.c 	__i915_schedule(node, &attr);
node              388 drivers/gpu/drm/i915/i915_scheduler.c void i915_sched_node_init(struct i915_sched_node *node)
node              390 drivers/gpu/drm/i915/i915_scheduler.c 	INIT_LIST_HEAD(&node->signalers_list);
node              391 drivers/gpu/drm/i915/i915_scheduler.c 	INIT_LIST_HEAD(&node->waiters_list);
node              392 drivers/gpu/drm/i915/i915_scheduler.c 	INIT_LIST_HEAD(&node->link);
node              393 drivers/gpu/drm/i915/i915_scheduler.c 	node->attr.priority = I915_PRIORITY_INVALID;
node              394 drivers/gpu/drm/i915/i915_scheduler.c 	node->semaphores = 0;
node              395 drivers/gpu/drm/i915/i915_scheduler.c 	node->flags = 0;
node              410 drivers/gpu/drm/i915/i915_scheduler.c bool __i915_sched_node_add_dependency(struct i915_sched_node *node,
node              422 drivers/gpu/drm/i915/i915_scheduler.c 		dep->waiter = node;
node              428 drivers/gpu/drm/i915/i915_scheduler.c 			node->flags |= I915_SCHED_HAS_SEMAPHORE_CHAIN;
node              431 drivers/gpu/drm/i915/i915_scheduler.c 		list_add(&dep->signal_link, &node->signalers_list);
node              452 drivers/gpu/drm/i915/i915_scheduler.c int i915_sched_node_add_dependency(struct i915_sched_node *node,
node              461 drivers/gpu/drm/i915/i915_scheduler.c 	if (!__i915_sched_node_add_dependency(node, signal, dep,
node              469 drivers/gpu/drm/i915/i915_scheduler.c void i915_sched_node_fini(struct i915_sched_node *node)
node              481 drivers/gpu/drm/i915/i915_scheduler.c 	list_for_each_entry_safe(dep, tmp, &node->signalers_list, signal_link) {
node              491 drivers/gpu/drm/i915/i915_scheduler.c 	list_for_each_entry_safe(dep, tmp, &node->waiters_list, wait_link) {
node              492 drivers/gpu/drm/i915/i915_scheduler.c 		GEM_BUG_ON(dep->signaler != node);
node               28 drivers/gpu/drm/i915/i915_scheduler.h void i915_sched_node_init(struct i915_sched_node *node);
node               30 drivers/gpu/drm/i915/i915_scheduler.h bool __i915_sched_node_add_dependency(struct i915_sched_node *node,
node               35 drivers/gpu/drm/i915/i915_scheduler.h int i915_sched_node_add_dependency(struct i915_sched_node *node,
node               38 drivers/gpu/drm/i915/i915_scheduler.h void i915_sched_node_fini(struct i915_sched_node *node);
node              469 drivers/gpu/drm/i915/i915_trace.h 			   __entry->offset = vma->node.start;
node              470 drivers/gpu/drm/i915/i915_trace.h 			   __entry->size = vma->node.size;
node              494 drivers/gpu/drm/i915/i915_trace.h 			   __entry->offset = vma->node.start;
node              495 drivers/gpu/drm/i915/i915_trace.h 			   __entry->size = vma->node.size;
node              618 drivers/gpu/drm/i915/i915_trace.h 	    TP_PROTO(struct i915_address_space *vm, struct drm_mm_node *node, unsigned int flags),
node              619 drivers/gpu/drm/i915/i915_trace.h 	    TP_ARGS(vm, node, flags),
node              633 drivers/gpu/drm/i915/i915_trace.h 			   __entry->start = node->start;
node              634 drivers/gpu/drm/i915/i915_trace.h 			   __entry->size = node->size;
node              635 drivers/gpu/drm/i915/i915_trace.h 			   __entry->color = node->color;
node              121 drivers/gpu/drm/i915/i915_vgpu.c 				struct drm_mm_node *node)
node              123 drivers/gpu/drm/i915/i915_vgpu.c 	if (!drm_mm_node_allocated(node))
node              127 drivers/gpu/drm/i915/i915_vgpu.c 			 node->start,
node              128 drivers/gpu/drm/i915/i915_vgpu.c 			 node->start + node->size,
node              129 drivers/gpu/drm/i915/i915_vgpu.c 			 node->size / 1024);
node              131 drivers/gpu/drm/i915/i915_vgpu.c 	ggtt->vm.reserved -= node->size;
node              132 drivers/gpu/drm/i915/i915_vgpu.c 	drm_mm_remove_node(node);
node              156 drivers/gpu/drm/i915/i915_vgpu.c 			     struct drm_mm_node *node,
node              167 drivers/gpu/drm/i915/i915_vgpu.c 	ret = i915_gem_gtt_reserve(&ggtt->vm, node,
node               63 drivers/gpu/drm/i915/i915_vma.c 	if (!vma->node.stack) {
node               65 drivers/gpu/drm/i915/i915_vma.c 				 vma->node.start, vma->node.size, reason);
node               69 drivers/gpu/drm/i915/i915_vma.c 	nr_entries = stack_depot_fetch(vma->node.stack, &entries);
node               72 drivers/gpu/drm/i915/i915_vma.c 			 vma->node.start, vma->node.size, reason, buf);
node              311 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              312 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(vma->size > vma->node.size);
node              314 drivers/gpu/drm/i915/i915_vma.c 	if (GEM_DEBUG_WARN_ON(range_overflows(vma->node.start,
node              315 drivers/gpu/drm/i915/i915_vma.c 					      vma->node.size,
node              367 drivers/gpu/drm/i915/i915_vma.c 					vma->node.start,
node              368 drivers/gpu/drm/i915/i915_vma.c 					vma->node.size);
node              438 drivers/gpu/drm/i915/i915_vma.c 	if (!drm_mm_node_allocated(&vma->node))
node              441 drivers/gpu/drm/i915/i915_vma.c 	if (vma->node.size < size)
node              445 drivers/gpu/drm/i915/i915_vma.c 	if (alignment && !IS_ALIGNED(vma->node.start, alignment))
node              452 drivers/gpu/drm/i915/i915_vma.c 	    vma->node.start < (flags & PIN_OFFSET_MASK))
node              456 drivers/gpu/drm/i915/i915_vma.c 	    vma->node.start != (flags & PIN_OFFSET_MASK))
node              469 drivers/gpu/drm/i915/i915_vma.c 	fenceable = (vma->node.size >= vma->fence_size &&
node              470 drivers/gpu/drm/i915/i915_vma.c 		     IS_ALIGNED(vma->node.start, vma->fence_alignment));
node              472 drivers/gpu/drm/i915/i915_vma.c 	mappable = vma->node.start + vma->fence_size <= i915_vm_to_ggtt(vma->vm)->mappable_end;
node              480 drivers/gpu/drm/i915/i915_vma.c static bool color_differs(struct drm_mm_node *node, unsigned long color)
node              482 drivers/gpu/drm/i915/i915_vma.c 	return node->allocated && node->color != color;
node              487 drivers/gpu/drm/i915/i915_vma.c 	struct drm_mm_node *node = &vma->node;
node              501 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(node));
node              502 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(list_empty(&node->node_list));
node              504 drivers/gpu/drm/i915/i915_vma.c 	other = list_prev_entry(node, node_list);
node              508 drivers/gpu/drm/i915/i915_vma.c 	other = list_next_entry(node, node_list);
node              509 drivers/gpu/drm/i915/i915_vma.c 	if (color_differs(other, cache_level) && !drm_mm_hole_follows(node))
node              551 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
node              610 drivers/gpu/drm/i915/i915_vma.c 		ret = i915_gem_gtt_reserve(vma->vm, &vma->node,
node              649 drivers/gpu/drm/i915/i915_vma.c 		ret = i915_gem_gtt_insert(vma->vm, &vma->node,
node              655 drivers/gpu/drm/i915/i915_vma.c 		GEM_BUG_ON(vma->node.start < start);
node              656 drivers/gpu/drm/i915/i915_vma.c 		GEM_BUG_ON(vma->node.start + vma->node.size > end);
node              658 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              683 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              689 drivers/gpu/drm/i915/i915_vma.c 	drm_mm_remove_node(&vma->node);
node              732 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              800 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(vma->node.allocated);
node              864 drivers/gpu/drm/i915/i915_vma.c 	struct drm_vma_offset_node *node = &vma->obj->base.vma_node;
node              877 drivers/gpu/drm/i915/i915_vma.c 			    drm_vma_node_offset_addr(node) + vma_offset,
node              895 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              971 drivers/gpu/drm/i915/i915_vma.c 	if (!drm_mm_node_allocated(&vma->node))
node               53 drivers/gpu/drm/i915/i915_vma.h 	struct drm_mm_node node;
node              217 drivers/gpu/drm/i915/i915_vma.h 	GEM_BUG_ON(!vma->node.allocated);
node              218 drivers/gpu/drm/i915/i915_vma.h 	GEM_BUG_ON(upper_32_bits(vma->node.start));
node              219 drivers/gpu/drm/i915/i915_vma.h 	GEM_BUG_ON(upper_32_bits(vma->node.start + vma->node.size - 1));
node              220 drivers/gpu/drm/i915/i915_vma.h 	return lower_32_bits(vma->node.start);
node              334 drivers/gpu/drm/i915/i915_vma.h 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              366 drivers/gpu/drm/i915/i915_vma.h 	GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              255 drivers/gpu/drm/i915/selftests/i915_gem_evict.c static void mock_color_adjust(const struct drm_mm_node *node,
node              389 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		struct drm_mm_node node;
node              434 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		if (i915_gem_gtt_insert(&i915->ggtt.vm, &r->node,
node              518 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		drm_mm_remove_node(&reserved->node);
node              293 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			mock_vma.node.size = BIT_ULL(size);
node              294 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			mock_vma.node.start = addr;
node              408 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					if (!drm_mm_node_allocated(&vma->node) ||
node              411 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
node              438 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					if (!drm_mm_node_allocated(&vma->node) ||
node              441 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size,
node              450 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size,
node              481 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					if (!drm_mm_node_allocated(&vma->node) ||
node              484 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
node              511 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					if (!drm_mm_node_allocated(&vma->node) ||
node              514 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
node              523 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 						       __func__, p->name, vma->node.start, vma->node.size,
node              599 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			if (!drm_mm_node_allocated(&vma->node) ||
node              614 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
node              683 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			if (!drm_mm_node_allocated(&vma->node) ||
node              787 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			if (!drm_mm_node_allocated(&vma->node) ||
node              865 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (!drm_mm_node_allocated(&vma->node) ||
node             1084 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct drm_mm_node *node;
node             1091 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) {
node             1096 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			ggtt->vm.mm.color_adjust(node, 0,
node             1333 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
node             1345 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1346 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (vma->node.start != total ||
node             1347 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		    vma->node.size != 2*I915_GTT_PAGE_SIZE) {
node             1349 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       vma->node.start, vma->node.size,
node             1383 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
node             1395 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1396 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (vma->node.start != total ||
node             1397 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		    vma->node.size != 2*I915_GTT_PAGE_SIZE) {
node             1399 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       vma->node.start, vma->node.size,
node             1427 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
node             1439 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1440 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (vma->node.start != offset ||
node             1441 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		    vma->node.size != 2*I915_GTT_PAGE_SIZE) {
node             1443 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       vma->node.start, vma->node.size,
node             1540 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
node             1557 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1569 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (!drm_mm_node_allocated(&vma->node)) {
node             1589 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1590 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		offset = vma->node.start;
node             1598 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
node             1609 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node             1610 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (vma->node.start != offset) {
node             1612 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       offset, vma->node.start);
node             1645 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
node             1656 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
node              669 drivers/gpu/drm/i915/selftests/i915_request.c 				    batch->node.start,
node              670 drivers/gpu/drm/i915/selftests/i915_request.c 				    batch->node.size,
node              793 drivers/gpu/drm/i915/selftests/i915_request.c 		*cmd++ = lower_32_bits(vma->node.start);
node              794 drivers/gpu/drm/i915/selftests/i915_request.c 		*cmd++ = upper_32_bits(vma->node.start);
node              797 drivers/gpu/drm/i915/selftests/i915_request.c 		*cmd++ = lower_32_bits(vma->node.start);
node              800 drivers/gpu/drm/i915/selftests/i915_request.c 		*cmd++ = lower_32_bits(vma->node.start);
node              872 drivers/gpu/drm/i915/selftests/i915_request.c 					    batch->node.start,
node              873 drivers/gpu/drm/i915/selftests/i915_request.c 					    batch->node.size,
node              990 drivers/gpu/drm/i915/selftests/i915_request.c 					    batch->node.start,
node              991 drivers/gpu/drm/i915/selftests/i915_request.c 					    batch->node.size,
node              567 drivers/gpu/drm/i915/selftests/i915_vma.c 					if (vma->node.size < vma->size) {
node              569 drivers/gpu/drm/i915/selftests/i915_vma.c 						       vma->size, vma->node.size);
node              661 drivers/gpu/drm/i915/selftests/i915_vma.c 	if (vma->node.size < vma->size) {
node              663 drivers/gpu/drm/i915/selftests/i915_vma.c 		       name, vma->size, vma->node.size);
node               71 drivers/gpu/drm/i915/selftests/igt_spinner.c 	return hws->node.start + seqno_offset(rq->fence.context);
node              143 drivers/gpu/drm/i915/selftests/igt_spinner.c 	*batch++ = lower_32_bits(vma->node.start);
node              144 drivers/gpu/drm/i915/selftests/igt_spinner.c 	*batch++ = upper_32_bits(vma->node.start);
node              156 drivers/gpu/drm/i915/selftests/igt_spinner.c 	err = engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, 0);
node               16 drivers/gpu/drm/lima/lima_vm.c 	struct drm_mm_node node;
node              131 drivers/gpu/drm/lima/lima_vm.c 	err = drm_mm_insert_node(&vm->mm, &bo_va->node, bo->gem.size);
node              135 drivers/gpu/drm/lima/lima_vm.c 	err = lima_vm_map_page_table(vm, bo->pages_dma_addr, bo_va->node.start,
node              136 drivers/gpu/drm/lima/lima_vm.c 				     bo_va->node.start + bo_va->node.size - 1);
node              148 drivers/gpu/drm/lima/lima_vm.c 	drm_mm_remove_node(&bo_va->node);
node              171 drivers/gpu/drm/lima/lima_vm.c 	lima_vm_unmap_page_table(vm, bo_va->node.start,
node              172 drivers/gpu/drm/lima/lima_vm.c 				 bo_va->node.start + bo_va->node.size - 1);
node              174 drivers/gpu/drm/lima/lima_vm.c 	drm_mm_remove_node(&bo_va->node);
node              193 drivers/gpu/drm/lima/lima_vm.c 	ret = bo_va->node.start;
node              555 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 		struct device_node *node;
node              557 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 		node = priv->comp_node[comp_id];
node              558 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 		if (!node) {
node              588 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 		struct device_node *node;
node              590 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 		node = priv->comp_node[comp_id];
node              593 drivers/gpu/drm/mediatek/mtk_drm_crtc.c 			dev_err(dev, "Component %pOF not initialized\n", node);
node              240 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c int mtk_ddp_comp_get_id(struct device_node *node,
node              243 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 	int id = of_alias_get_id(node, mtk_ddp_comp_stem[comp_type]);
node              255 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c int mtk_ddp_comp_init(struct device *dev, struct device_node *node,
node              285 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 	comp->regs = of_iomap(node, 0);
node              286 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 	comp->irq = of_irq_get(node, 0);
node              287 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 	comp->clk = of_clk_get(node, 0);
node              298 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 	larb_node = of_parse_phandle(node, "mediatek,larb", 0);
node              301 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.c 			"Missing mediadek,larb phandle in %pOF node\n", node);
node              161 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h int mtk_ddp_comp_get_id(struct device_node *node,
node              489 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	struct device_node *node;
node              512 drivers/gpu/drm/mediatek/mtk_drm_drv.c 	for_each_child_of_node(dev->of_node->parent, node) {
node              517 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		of_id = of_match_node(mtk_ddp_comp_dt_ids, node);
node              521 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		if (!of_device_is_available(node)) {
node              523 drivers/gpu/drm/mediatek/mtk_drm_drv.c 				node);
node              530 drivers/gpu/drm/mediatek/mtk_drm_drv.c 			private->mutex_node = of_node_get(node);
node              534 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		comp_id = mtk_ddp_comp_get_id(node, comp_type);
node              537 drivers/gpu/drm/mediatek/mtk_drm_drv.c 				 node);
node              541 drivers/gpu/drm/mediatek/mtk_drm_drv.c 		private->comp_node[comp_id] = of_node_get(node);
node              554 drivers/gpu/drm/mediatek/mtk_drm_drv.c 				 node);
node              556 drivers/gpu/drm/mediatek/mtk_drm_drv.c 						   node);
node              563 drivers/gpu/drm/mediatek/mtk_drm_drv.c 				of_node_put(node);
node              567 drivers/gpu/drm/mediatek/mtk_drm_drv.c 			ret = mtk_ddp_comp_init(dev, node, comp, comp_id, NULL);
node              569 drivers/gpu/drm/mediatek/mtk_drm_drv.c 				of_node_put(node);
node               80 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               81 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c 	struct drm_device *dev = node->minor->dev;
node               85 drivers/gpu/drm/msm/adreno/a5xx_debugfs.c 		node->info_ent->data;
node             1260 drivers/gpu/drm/msm/adreno/a6xx_gmu.c int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node)
node             1263 drivers/gpu/drm/msm/adreno/a6xx_gmu.c 	struct platform_device *pdev = of_find_device_by_node(node);
node             1271 drivers/gpu/drm/msm/adreno/a6xx_gmu.c 	of_dma_configure(gmu->dev, node, true);
node              851 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	struct device_node *node;
node              874 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	node = of_parse_phandle(pdev->dev.of_node, "qcom,gmu", 0);
node              877 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	BUG_ON(!node);
node              879 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	ret = a6xx_gmu_init(a6xx_gpu, node);
node               56 drivers/gpu/drm/msm/adreno/a6xx_gpu.h int a6xx_gmu_init(struct a6xx_gpu *a6xx_gpu, struct device_node *node);
node               79 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c 	struct list_head node;
node               91 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c 	list_add_tail(&obj->node, &a6xx_state->objs);
node              913 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c 	list_for_each_entry_safe(obj, tmp, &a6xx_state->objs, node)
node              274 drivers/gpu/drm/msm/adreno/adreno_device.c 	struct device_node *node = dev->of_node;
node              280 drivers/gpu/drm/msm/adreno/adreno_device.c 	ret = of_property_read_string_index(node, "compatible", 0, &compat);
node              298 drivers/gpu/drm/msm/adreno/adreno_device.c 	ret = of_property_read_u32(node, "qcom,chipid", &chipid);
node              823 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	struct device_node *child, *node;
node              826 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	node = of_get_compatible_child(dev->of_node, "qcom,gpu-pwrlevels");
node              827 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	if (!node) {
node              832 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	for_each_child_of_node(node, child) {
node              847 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	of_node_put(node);
node              245 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              246 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c 	struct drm_device *dev = node->minor->dev;
node               28 drivers/gpu/drm/msm/disp/mdp_kms.c 	list_for_each_entry(irq, &mdp_kms->irq_list, node)
node               53 drivers/gpu/drm/msm/disp/mdp_kms.c 	list_for_each_entry_safe(handler, n, &mdp_kms->irq_list, node) {
node              111 drivers/gpu/drm/msm/disp/mdp_kms.c 		list_add(&irq->node, &mdp_kms->irq_list);
node              130 drivers/gpu/drm/msm/disp/mdp_kms.c 		list_del(&irq->node);
node               58 drivers/gpu/drm/msm/disp/mdp_kms.h 	struct list_head node;
node              162 drivers/gpu/drm/msm/msm_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              163 drivers/gpu/drm/msm/msm_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              165 drivers/gpu/drm/msm/msm_debugfs.c 			node->info_ent->data;
node              312 drivers/gpu/drm/msm/msm_drv.c 	struct device_node *node;
node              333 drivers/gpu/drm/msm/msm_drv.c 	node = of_parse_phandle(dev->dev->of_node, "memory-region", 0);
node              334 drivers/gpu/drm/msm/msm_drv.c 	if (node) {
node              336 drivers/gpu/drm/msm/msm_drv.c 		ret = of_address_to_resource(node, 0, &r);
node              337 drivers/gpu/drm/msm/msm_drv.c 		of_node_put(node);
node             1065 drivers/gpu/drm/msm/msm_gem.c 		to_msm_bo(obj)->vram_node = &vma->node;
node               29 drivers/gpu/drm/msm/msm_gem.h 	struct drm_mm_node node;
node              134 drivers/gpu/drm/msm/msm_gem.h 	struct list_head node;   /* node in ring submit list */
node               55 drivers/gpu/drm/msm/msm_gem_submit.c 	INIT_LIST_HEAD(&submit->node);
node               65 drivers/gpu/drm/msm/msm_gem_submit.c 	list_del(&submit->node);
node               34 drivers/gpu/drm/msm/msm_gem_vma.c 	unsigned size = vma->node.size << PAGE_SHIFT;
node               96 drivers/gpu/drm/msm/msm_gem_vma.c 		drm_mm_remove_node(&vma->node);
node              114 drivers/gpu/drm/msm/msm_gem_vma.c 	ret = drm_mm_insert_node(&aspace->mm, &vma->node, npages);
node              120 drivers/gpu/drm/msm/msm_gem_vma.c 	vma->iova = vma->node.start << PAGE_SHIFT;
node              394 drivers/gpu/drm/msm/msm_gpu.c 	list_for_each_entry(submit, &ring->submits, node) {
node              410 drivers/gpu/drm/msm/msm_gpu.c 	list_for_each_entry(submit, &ring->submits, node)
node              501 drivers/gpu/drm/msm/msm_gpu.c 			list_for_each_entry(submit, &ring->submits, node)
node              699 drivers/gpu/drm/msm/msm_gpu.c 		list_for_each_entry_safe(submit, tmp, &ring->submits, node) {
node              745 drivers/gpu/drm/msm/msm_gpu.c 	list_add_tail(&submit->node, &ring->submits);
node              176 drivers/gpu/drm/msm/msm_gpu.h 	struct list_head node;
node               28 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry(entry, &ctx->submitqueues, node) {
node               52 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry_safe(entry, tmp, &ctx->submitqueues, node)
node               87 drivers/gpu/drm/msm/msm_submitqueue.c 	list_add_tail(&queue->node, &ctx->submitqueues);
node              173 drivers/gpu/drm/msm/msm_submitqueue.c 	list_for_each_entry(entry, &ctx->submitqueues, node) {
node              175 drivers/gpu/drm/msm/msm_submitqueue.c 			list_del(&entry->node);
node               17 drivers/gpu/drm/nouveau/include/nvkm/core/gpuobj.h 	struct nvkm_mm_node *node;
node               46 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	struct nvkm_mm_node *node;
node               48 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	list_for_each_entry(node, &mm->nodes, nl_entry) {
node               49 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 		if (node->heap == heap)
node               50 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 			size += node->length;
node               56 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_contiguous(struct nvkm_mm_node *node)
node               58 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	return !node->next;
node               62 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_addr(struct nvkm_mm_node *node)
node               64 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	if (WARN_ON(!nvkm_mm_contiguous(node)))
node               66 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	return node->offset;
node               70 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h nvkm_mm_size(struct nvkm_mm_node *node)
node               74 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 		size += node->length;
node               75 drivers/gpu/drm/nouveau/include/nvkm/core/mm.h 	} while ((node = node->next));
node               20 drivers/gpu/drm/nouveau/include/nvkm/core/object.h 	struct rb_node node;
node              118 drivers/gpu/drm/nouveau/nouveau_abi16.c 	nvkm_mm_free(&chan->heap, &ntfy->node);
node              546 drivers/gpu/drm/nouveau/nouveau_abi16.c 			   &ntfy->node);
node              550 drivers/gpu/drm/nouveau/nouveau_abi16.c 	args.start = ntfy->node->offset;
node              551 drivers/gpu/drm/nouveau/nouveau_abi16.c 	args.limit = ntfy->node->offset + ntfy->node->length - 1;
node              580 drivers/gpu/drm/nouveau/nouveau_abi16.c 	info->offset = ntfy->node->offset;
node               18 drivers/gpu/drm/nouveau/nouveau_abi16.h 	struct nvkm_mm_node *node;
node               40 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               41 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	struct nouveau_drm *drm = nouveau_drm(node->minor->dev);
node               52 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	struct drm_info_node *node = m->private;
node               53 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	struct nouveau_drm *drm = nouveau_drm(node->minor->dev);
node              112 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 	return nvkm_memory_map(gpuobj->parent, gpuobj->node->offset + offset,
node              119 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 	return nvkm_ro32(gpuobj->parent, gpuobj->node->offset + offset);
node              125 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 	nvkm_wo32(gpuobj->parent, gpuobj->node->offset + offset, data);
node              157 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 		gpuobj->map  = (u8 *)gpuobj->map + gpuobj->node->offset;
node              181 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 					   max(align, 1), &gpuobj->node);
node              184 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 					   -align, &gpuobj->node);
node              191 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 		gpuobj->addr = parent->addr + gpuobj->node->offset;
node              192 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 		gpuobj->size = gpuobj->node->length;
node              220 drivers/gpu/drm/nouveau/nvkm/core/gpuobj.c 			nvkm_mm_free(&gpuobj->parent->heap, &gpuobj->node);
node               32 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *node;
node               36 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_for_each_entry(node, &mm->nodes, nl_entry) {
node               38 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		       node->offset, node->length, node->type);
node               41 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_for_each_entry(node, &mm->free, fl_entry) {
node               43 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		       node->offset, node->length, node->type);
node               53 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		struct nvkm_mm_node *prev = node(this, prev);
node               54 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		struct nvkm_mm_node *next = node(this, next);
node              129 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = node(this, prev);
node              133 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = node(this, next);
node              203 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = node(this, prev);
node              207 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		next = node(this, next);
node              242 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *node, *prev;
node              246 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry);
node              250 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			if (!(node = kzalloc(sizeof(*node), GFP_KERNEL)))
node              252 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			node->type   = NVKM_MM_TYPE_HOLE;
node              253 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			node->offset = next;
node              254 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			node->length = offset - next;
node              255 drivers/gpu/drm/nouveau/nvkm/core/mm.c 			list_add_tail(&node->nl_entry, &mm->nodes);
node              265 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              266 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	if (!node)
node              270 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		node->offset  = roundup(offset, mm->block_size);
node              271 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		node->length  = rounddown(offset + length, mm->block_size);
node              272 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		node->length -= node->offset;
node              275 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_add_tail(&node->nl_entry, &mm->nodes);
node              276 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_add_tail(&node->fl_entry, &mm->free);
node              277 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	node->heap = heap;
node              285 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	struct nvkm_mm_node *node, *temp;
node              291 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_for_each_entry(node, &mm->nodes, nl_entry) {
node              292 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		if (node->type != NVKM_MM_TYPE_HOLE) {
node              300 drivers/gpu/drm/nouveau/nvkm/core/mm.c 	list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) {
node              301 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		list_del(&node->nl_entry);
node              302 drivers/gpu/drm/nouveau/nvkm/core/mm.c 		kfree(node);
node               35 drivers/gpu/drm/nouveau/nvkm/core/object.c 		struct rb_node *node = client->objroot.rb_node;
node               36 drivers/gpu/drm/nouveau/nvkm/core/object.c 		while (node) {
node               37 drivers/gpu/drm/nouveau/nvkm/core/object.c 			object = rb_entry(node, typeof(*object), node);
node               39 drivers/gpu/drm/nouveau/nvkm/core/object.c 				node = node->rb_left;
node               42 drivers/gpu/drm/nouveau/nvkm/core/object.c 				node = node->rb_right;
node               60 drivers/gpu/drm/nouveau/nvkm/core/object.c 	if (!RB_EMPTY_NODE(&object->node))
node               61 drivers/gpu/drm/nouveau/nvkm/core/object.c 		rb_erase(&object->node, &object->client->objroot);
node               71 drivers/gpu/drm/nouveau/nvkm/core/object.c 		struct nvkm_object *this = rb_entry(*ptr, typeof(*this), node);
node               82 drivers/gpu/drm/nouveau/nvkm/core/object.c 	rb_link_node(&object->node, parent, ptr);
node               83 drivers/gpu/drm/nouveau/nvkm/core/object.c 	rb_insert_color(&object->node, &object->client->objroot);
node              307 drivers/gpu/drm/nouveau/nvkm/core/object.c 	RB_CLEAR_NODE(&object->node);
node               83 drivers/gpu/drm/nouveau/nvkm/core/ramht.c 				inst = data->inst->node->offset;
node               74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c 	nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4);
node               81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c 				     (chan->ramht->gpuobj->node->offset >> 4));
node               74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c 	nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4);
node               81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c 				     (chan->ramht->gpuobj->node->offset >> 4));
node               74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c 	nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4);
node               82 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c 				     (chan->ramht->gpuobj->node->offset >> 4));
node               74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c 	nvkm_wo32(chan->ramfc, 0x48, chan->base.push->node->offset >> 4);
node               82 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c 				     (chan->ramht->gpuobj->node->offset >> 4));
node               69 drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c 	nvkm_wr32(device, 0x001708, 0x80000000 | bar->bar1->node->offset >> 4);
node               91 drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c 	nvkm_wr32(device, 0x00170c, 0x80000000 | bar->bar2->node->offset >> 4);
node               83 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	struct nvkm_mm_node *node;
node               85 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	while ((node = next)) {
node               86 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		next = node->next;
node               87 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		nvkm_mm_free(&vram->ram->vram, &node);
node              109 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	struct nvkm_mm_node **node, *r;
node              130 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 	node = &vram->mn;
node              142 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		*node = r;
node              143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c 		node = &r->next;
node              176 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = gk20a_instobj(memory);
node              177 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->imem;
node              182 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	return node->vaddr;
node              188 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
node              189 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->base.imem;
node              197 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (node->base.vaddr) {
node              198 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		if (!node->use_cpt) {
node              200 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 			list_del(&node->vaddr_node);
node              209 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP,
node              211 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (!node->base.vaddr) {
node              222 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->use_cpt++;
node              225 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	return node->base.vaddr;
node              231 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = gk20a_instobj(memory);
node              232 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->imem;
node              243 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
node              244 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->base.imem;
node              250 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (WARN_ON(node->use_cpt == 0))
node              254 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (--node->use_cpt == 0)
node              255 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		list_add_tail(&node->vaddr_node, &imem->vaddr_lru);
node              267 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = gk20a_instobj(memory);
node              269 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	return node->vaddr[offset / 4];
node              275 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = gk20a_instobj(memory);
node              277 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->vaddr[offset / 4] = data;
node              284 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = gk20a_instobj(memory);
node              286 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		.memory = &node->memory,
node              288 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		.mem = node->mn,
node              297 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_dma *node = gk20a_instobj_dma(memory);
node              298 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->base.imem;
node              301 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (unlikely(!node->base.vaddr))
node              304 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT,
node              305 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		       node->base.vaddr, node->handle, imem->attrs);
node              308 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	return node;
node              314 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory);
node              315 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instmem *imem = node->base.imem;
node              317 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct nvkm_mm_node *r = node->base.mn;
node              326 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (node->base.vaddr)
node              327 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		gk20a_instobj_iommu_recycle_vaddr(node);
node              335 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	for (i = 0; i < node->base.mn->length; i++) {
node              338 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		dma_unmap_page(dev, node->dma_addrs[i], PAGE_SIZE,
node              340 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		__free_page(node->pages[i]);
node              349 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	return node;
node              386 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_dma *node;
node              390 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (!(node = kzalloc(sizeof(*node), GFP_KERNEL)))
node              392 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	*_node = &node->base;
node              394 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	nvkm_memory_ctor(&gk20a_instobj_func_dma, &node->base.memory);
node              395 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.memory.ptrs = &gk20a_instobj_ptrs;
node              397 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.vaddr = dma_alloc_attrs(dev, npages << PAGE_SHIFT,
node              398 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 					   &node->handle, GFP_KERNEL,
node              400 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (!node->base.vaddr) {
node              406 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (unlikely(node->handle & (align - 1)))
node              409 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 			  &node->handle, align);
node              412 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->r.type = 12;
node              413 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->r.offset = node->handle >> 12;
node              414 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->r.length = (npages << PAGE_SHIFT) >> 12;
node              416 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.mn = &node->r;
node              424 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj_iommu *node;
node              435 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	if (!(node = kzalloc(sizeof(*node) + ((sizeof(node->pages[0]) +
node              436 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 			     sizeof(*node->dma_addrs)) * npages), GFP_KERNEL)))
node              438 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	*_node = &node->base;
node              439 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->dma_addrs = (void *)(node->pages + npages);
node              441 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	nvkm_memory_ctor(&gk20a_instobj_func_iommu, &node->base.memory);
node              442 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.memory.ptrs = &gk20a_instobj_ptrs;
node              453 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		node->pages[i] = p;
node              460 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		node->dma_addrs[i] = dma_adr;
node              477 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		ret = iommu_map(imem->domain, offset, node->dma_addrs[i],
node              493 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->base.mn = r;
node              502 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	for (i = 0; i < npages && node->pages[i] != NULL; i++) {
node              503 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		dma_addr_t dma_addr = node->dma_addrs[i];
node              507 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		__free_page(node->pages[i]);
node              519 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	struct gk20a_instobj *node = NULL;
node              531 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 					       align, &node);
node              534 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 					     align, &node);
node              535 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	*pmemory = node ? &node->memory : NULL;
node              539 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 	node->imem = imem;
node              542 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c 		   size, align, (u64)node->mn->offset << 12);
node               42 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	struct nvkm_mm_node *node;
node               50 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	nvkm_wr32(device, 0x700000 + iobj->node->offset + offset, data);
node               58 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	return nvkm_rd32(device, 0x700000 + iobj->node->offset + offset);
node               77 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	return device->pri + 0x700000 + iobj->node->offset;
node               83 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	return nv04_instobj(memory)->node->length;
node               89 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	return nv04_instobj(memory)->node->offset;
node              103 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 	nvkm_mm_free(&iobj->imem->heap, &iobj->node);
node              137 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c 			   align ? align : 1, &iobj->node);
node               44 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	struct nvkm_mm_node *node;
node               51 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset);
node               58 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset);
node               77 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	return iobj->imem->iomem + iobj->node->offset;
node               83 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	return nv40_instobj(memory)->node->length;
node               89 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	return nv40_instobj(memory)->node->offset;
node              103 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 	nvkm_mm_free(&iobj->imem->heap, &iobj->node);
node              137 drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c 			   align ? align : 1, &iobj->node);
node              865 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct rb_node *node = vmm->root.rb_node;
node              866 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	while (node) {
node              867 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
node              869 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			node = node->rb_left;
node              872 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			node = node->rb_right;
node              981 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct rb_node *node;
node              986 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	while ((node = rb_first(&vmm->root))) {
node              987 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
node             1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->addr == addr && vma->part && (prev = node(vma, prev))) {
node             1162 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->addr + vma->size == addr + size && (next = node(vma, next))) {
node             1199 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((vma = node(vma, next)) && (start = vma->addr) < limit);
node             1314 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			vma = node(vma, next);
node             1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->part && (prev = node(vma, prev)) && prev->mapped)
node             1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((next = node(vma, next)) && (!next->part || next->mapped))
node             1526 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((prev = node(vma, prev)) && !prev->used) {
node             1532 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((next = node(vma, next)) && !next->used) {
node             1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			while ((next = node(next, next)) && next->part &&
node             1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((next = node(vma, next)) && next->part);
node             1640 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	struct rb_node *node = NULL, *temp;
node             1693 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			node = temp;
node             1698 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (unlikely(!node))
node             1705 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *this = rb_entry(node, typeof(*this), tree);
node             1706 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *prev = node(this, prev);
node             1707 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		struct nvkm_vma *next = node(this, next);
node             1724 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((node = rb_next(node)));
node              404 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c 	list_for_each_entry(img, imgs, base.node) {
node              424 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c 	list_for_each_entry(_img, imgs, node) {
node              437 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c 	list_for_each_entry(_img, imgs, node) {
node              512 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c 		list_add_tail(&img->node, &imgs);
node              584 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r352.c 	list_for_each_entry_safe(img, t, &imgs, node) {
node              260 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c 	list_for_each_entry(img, imgs, base.node) {
node              276 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c 	list_for_each_entry(_img, imgs, node) {
node              289 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/acr_r367.c 	list_for_each_entry(_img, imgs, node) {
node               96 drivers/gpu/drm/nouveau/nvkm/subdev/secboot/ls_ucode.h 	struct list_head node;
node             1165 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c 	struct device_node *node = pdev->dev.of_node;
node             1186 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c 	err = of_get_display_timing(node, "panel-timing", &timing);
node             1198 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c 	of_property_read_u32(node, "width-mm", &ddata->width_mm);
node             1201 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c 	of_property_read_u32(node, "height-mm", &ddata->height_mm);
node             1219 drivers/gpu/drm/omapdrm/displays/panel-dsi-cm.c 	backlight = of_parse_phandle(node, "backlight", 0);
node               69 drivers/gpu/drm/omapdrm/dss/base.c static bool omapdss_device_is_registered(struct device_node *node)
node               77 drivers/gpu/drm/omapdrm/dss/base.c 		if (dssdev->dev->of_node == node) {
node              108 drivers/gpu/drm/omapdrm/dss/base.c struct omap_dss_device *omapdss_find_device_by_node(struct device_node *node)
node              113 drivers/gpu/drm/omapdrm/dss/base.c 		if (dssdev->dev->of_node == node)
node              296 drivers/gpu/drm/omapdrm/dss/base.c 	struct device_node *node;
node              301 drivers/gpu/drm/omapdrm/dss/base.c static bool omapdss_list_contains(const struct device_node *node)
node              306 drivers/gpu/drm/omapdrm/dss/base.c 		if (comp->node == node)
node              313 drivers/gpu/drm/omapdrm/dss/base.c static void omapdss_walk_device(struct device *dev, struct device_node *node,
node              321 drivers/gpu/drm/omapdrm/dss/base.c 	ret = of_property_read_string(node, "compatible", &compat);
node              327 drivers/gpu/drm/omapdrm/dss/base.c 		comp->node = node;
node              337 drivers/gpu/drm/omapdrm/dss/base.c 	n = of_get_child_by_name(node, "ports");
node              339 drivers/gpu/drm/omapdrm/dss/base.c 		n = of_get_child_by_name(node, "port");
node              346 drivers/gpu/drm/omapdrm/dss/base.c 	while ((n = of_graph_get_next_endpoint(node, n)) != NULL) {
node              380 drivers/gpu/drm/omapdrm/dss/base.c 	if (omapdss_device_is_registered(comp->node))
node             5144 drivers/gpu/drm/omapdrm/dss/dsi.c 	struct device_node *node = dsi->dev->of_node;
node             5152 drivers/gpu/drm/omapdrm/dss/dsi.c 	ep = of_graph_get_endpoint_by_regs(node, 0, 0);
node               14 drivers/gpu/drm/omapdrm/dss/dss-of.c omapdss_of_find_connected_device(struct device_node *node, unsigned int port)
node               19 drivers/gpu/drm/omapdrm/dss/dss-of.c 	remote_node = of_graph_get_remote_node(node, port, 0);
node              699 drivers/gpu/drm/omapdrm/dss/hdmi4.c 	struct device_node *node = pdev->dev.of_node;
node              703 drivers/gpu/drm/omapdrm/dss/hdmi4.c 	ep = of_graph_get_endpoint_by_regs(node, 0, 0);
node              683 drivers/gpu/drm/omapdrm/dss/hdmi5.c 	struct device_node *node = pdev->dev.of_node;
node              687 drivers/gpu/drm/omapdrm/dss/hdmi5.c 	ep = of_graph_get_endpoint_by_regs(node, 0, 0);
node               28 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	struct device_node *node;
node               44 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_update_prop(struct device_node *node, char *compat,
node               57 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	of_update_property(node, prop);
node               80 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_omapify_node(struct device_node *node)
node               87 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	prop = of_find_property(node, "compatible", NULL);
node              106 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	omapdss_update_prop(node, new_compat, new_len);
node              109 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_add_to_list(struct device_node *node, bool root)
node              113 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 		n->node = node;
node              119 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static bool __init omapdss_list_contains(const struct device_node *node)
node              124 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 		if (n->node == node)
node              131 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c static void __init omapdss_walk_device(struct device_node *node, bool root)
node              135 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	omapdss_add_to_list(node, root);
node              141 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	n = of_get_child_by_name(node, "ports");
node              143 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 		n = of_get_child_by_name(node, "port");
node              150 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 	while ((n = of_graph_get_next_endpoint(node, n)) != NULL) {
node              212 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 		if (of_match_node(omapdss_of_fixups_whitelist, n->node))
node              213 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 			omapdss_omapify_node(n->node);
node              216 drivers/gpu/drm/omapdrm/dss/omapdss-boot-init.c 		of_node_put(n->node);
node              472 drivers/gpu/drm/omapdrm/dss/omapdss.h struct omap_dss_device *omapdss_find_device_by_node(struct device_node *node);
node              506 drivers/gpu/drm/omapdrm/dss/omapdss.h omapdss_of_find_connected_device(struct device_node *node, unsigned int port);
node              777 drivers/gpu/drm/omapdrm/dss/venc.c 	struct device_node *node = venc->pdev->dev.of_node;
node              782 drivers/gpu/drm/omapdrm/dss/venc.c 	ep = of_graph_get_endpoint_by_regs(node, 0, 0);
node               21 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               22 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_device *dev = node->minor->dev;
node               35 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               36 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_device *dev = node->minor->dev;
node               47 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               48 drivers/gpu/drm/omapdrm/omap_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              208 drivers/gpu/drm/omapdrm/omap_drv.c 	struct device_node *node = NULL;
node              214 drivers/gpu/drm/omapdrm/omap_drv.c 		node = display->dev->of_node;
node              222 drivers/gpu/drm/omapdrm/omap_drv.c 		node = bridge->of_node;
node              224 drivers/gpu/drm/omapdrm/omap_drv.c 		node = output->panel->dev->of_node;
node              227 drivers/gpu/drm/omapdrm/omap_drv.c 	return node ? of_alias_get_id(node, "display") : -ENODEV;
node               12 drivers/gpu/drm/omapdrm/omap_irq.c 	struct list_head node;
node               27 drivers/gpu/drm/omapdrm/omap_irq.c 	list_for_each_entry(wait, &priv->wait_list, node)
node               53 drivers/gpu/drm/omapdrm/omap_irq.c 	list_add(&wait->node, &priv->wait_list);
node               70 drivers/gpu/drm/omapdrm/omap_irq.c 	list_del(&wait->node);
node              242 drivers/gpu/drm/omapdrm/omap_irq.c 	list_for_each_entry_safe(wait, n, &priv->wait_list, node) {
node              371 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c 		.node = NULL,
node              416 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c 	info.node = of_graph_get_remote_port(endpoint);
node              417 drivers/gpu/drm/panel/panel-raspberrypi-touchscreen.c 	if (!info.node)
node              567 drivers/gpu/drm/panel/panel-truly-nt35597.c 		.node = NULL,
node              410 drivers/gpu/drm/panfrost/panfrost_drv.c 					 node);
node              457 drivers/gpu/drm/panfrost/panfrost_drv.c static void panfrost_drm_mm_color_adjust(const struct drm_mm_node *node,
node               62 drivers/gpu/drm/panfrost/panfrost_gem.c 	list_for_each_entry(iter, &bo->mappings.list, node) {
node              113 drivers/gpu/drm/panfrost/panfrost_gem.c 	list_for_each_entry(mapping, &bo->mappings.list, node)
node              132 drivers/gpu/drm/panfrost/panfrost_gem.c 	INIT_LIST_HEAD(&mapping->node);
node              164 drivers/gpu/drm/panfrost/panfrost_gem.c 	list_add_tail(&mapping->node, &bo->mappings.list);
node              180 drivers/gpu/drm/panfrost/panfrost_gem.c 	list_for_each_entry(iter, &bo->mappings.list, node) {
node              183 drivers/gpu/drm/panfrost/panfrost_gem.c 			list_del(&iter->node);
node               44 drivers/gpu/drm/panfrost/panfrost_gem.h 	struct list_head node;
node               59 drivers/gpu/drm/panfrost/panfrost_gem.h drm_mm_node_to_panfrost_mapping(struct drm_mm_node *node)
node               61 drivers/gpu/drm/panfrost/panfrost_gem.h 	return container_of(node, struct panfrost_gem_mapping, mmnode);
node              412 drivers/gpu/drm/panfrost/panfrost_mmu.c 	struct drm_mm_node *node;
node              428 drivers/gpu/drm/panfrost/panfrost_mmu.c 	drm_mm_for_each_node(node, &priv->mm) {
node              429 drivers/gpu/drm/panfrost/panfrost_mmu.c 		if (offset >= node->start &&
node              430 drivers/gpu/drm/panfrost/panfrost_mmu.c 		    offset < (node->start + node->size)) {
node              431 drivers/gpu/drm/panfrost/panfrost_mmu.c 			mapping = drm_mm_node_to_panfrost_mapping(node);
node               36 drivers/gpu/drm/pl111/pl111_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node               37 drivers/gpu/drm/pl111/pl111_debugfs.c 	struct drm_device *dev = node->minor->dev;
node               41 drivers/gpu/drm/qxl/qxl_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               42 drivers/gpu/drm/qxl/qxl_debugfs.c 	struct qxl_device *qdev = node->minor->dev->dev_private;
node               55 drivers/gpu/drm/qxl/qxl_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               56 drivers/gpu/drm/qxl/qxl_debugfs.c 	struct qxl_device *qdev = node->minor->dev->dev_private;
node              369 drivers/gpu/drm/qxl/qxl_ttm.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              370 drivers/gpu/drm/qxl/qxl_ttm.c 	struct drm_mm *mm = (struct drm_mm *)node->info_ent->data;
node              371 drivers/gpu/drm/qxl/qxl_ttm.c 	struct drm_device *dev = node->minor->dev;
node             2925 drivers/gpu/drm/radeon/r100.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             2926 drivers/gpu/drm/radeon/r100.c 	struct drm_device *dev = node->minor->dev;
node             2946 drivers/gpu/drm/radeon/r100.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             2947 drivers/gpu/drm/radeon/r100.c 	struct drm_device *dev = node->minor->dev;
node             2974 drivers/gpu/drm/radeon/r100.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             2975 drivers/gpu/drm/radeon/r100.c 	struct drm_device *dev = node->minor->dev;
node             3024 drivers/gpu/drm/radeon/r100.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             3025 drivers/gpu/drm/radeon/r100.c 	struct drm_device *dev = node->minor->dev;
node              595 drivers/gpu/drm/radeon/r300.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              596 drivers/gpu/drm/radeon/r300.c 	struct drm_device *dev = node->minor->dev;
node              485 drivers/gpu/drm/radeon/r420.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              486 drivers/gpu/drm/radeon/r420.c 	struct drm_device *dev = node->minor->dev;
node             4351 drivers/gpu/drm/radeon/r600.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             4352 drivers/gpu/drm/radeon/r600.c 	struct drm_device *dev = node->minor->dev;
node              751 drivers/gpu/drm/radeon/radeon_dp_mst.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              752 drivers/gpu/drm/radeon/radeon_dp_mst.c 	struct drm_device *dev = node->minor->dev;
node              978 drivers/gpu/drm/radeon/radeon_fence.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              979 drivers/gpu/drm/radeon/radeon_fence.c 	struct drm_device *dev = node->minor->dev;
node             1011 drivers/gpu/drm/radeon/radeon_fence.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             1012 drivers/gpu/drm/radeon/radeon_fence.c 	struct drm_device *dev = node->minor->dev;
node              782 drivers/gpu/drm/radeon/radeon_gem.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              783 drivers/gpu/drm/radeon/radeon_gem.c 	struct drm_device *dev = node->minor->dev;
node              299 drivers/gpu/drm/radeon/radeon_ib.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              300 drivers/gpu/drm/radeon/radeon_ib.c 	struct drm_device *dev = node->minor->dev;
node               85 drivers/gpu/drm/radeon/radeon_mn.c 		struct radeon_mn_node *node;
node               94 drivers/gpu/drm/radeon/radeon_mn.c 		node = container_of(it, struct radeon_mn_node, it);
node               97 drivers/gpu/drm/radeon/radeon_mn.c 		list_for_each_entry(bo, &node->bos, mn_list) {
node              180 drivers/gpu/drm/radeon/radeon_mn.c 	struct radeon_mn_node *node = NULL;
node              194 drivers/gpu/drm/radeon/radeon_mn.c 		kfree(node);
node              195 drivers/gpu/drm/radeon/radeon_mn.c 		node = container_of(it, struct radeon_mn_node, it);
node              196 drivers/gpu/drm/radeon/radeon_mn.c 		interval_tree_remove(&node->it, &rmn->objects);
node              199 drivers/gpu/drm/radeon/radeon_mn.c 		list_splice(&node->bos, &bos);
node              202 drivers/gpu/drm/radeon/radeon_mn.c 	if (!node) {
node              203 drivers/gpu/drm/radeon/radeon_mn.c 		node = kmalloc(sizeof(struct radeon_mn_node), GFP_KERNEL);
node              204 drivers/gpu/drm/radeon/radeon_mn.c 		if (!node) {
node              212 drivers/gpu/drm/radeon/radeon_mn.c 	node->it.start = addr;
node              213 drivers/gpu/drm/radeon/radeon_mn.c 	node->it.last = end;
node              214 drivers/gpu/drm/radeon/radeon_mn.c 	INIT_LIST_HEAD(&node->bos);
node              215 drivers/gpu/drm/radeon/radeon_mn.c 	list_splice(&bos, &node->bos);
node              216 drivers/gpu/drm/radeon/radeon_mn.c 	list_add(&bo->mn_list, &node->bos);
node              218 drivers/gpu/drm/radeon/radeon_mn.c 	interval_tree_insert(&node->it, &rmn->objects);
node              247 drivers/gpu/drm/radeon/radeon_mn.c 		struct radeon_mn_node *node;
node              248 drivers/gpu/drm/radeon/radeon_mn.c 		node = container_of(head, struct radeon_mn_node, bos);
node              249 drivers/gpu/drm/radeon/radeon_mn.c 		interval_tree_remove(&node->it, &rmn->objects);
node              250 drivers/gpu/drm/radeon/radeon_mn.c 		kfree(node);
node             1866 drivers/gpu/drm/radeon/radeon_pm.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node             1867 drivers/gpu/drm/radeon/radeon_pm.c 	struct drm_device *dev = node->minor->dev;
node              466 drivers/gpu/drm/radeon/radeon_ring.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              467 drivers/gpu/drm/radeon/radeon_ring.c 	struct drm_device *dev = node->minor->dev;
node              469 drivers/gpu/drm/radeon/radeon_ring.c 	int ridx = *(int*)node->info_ent->data;
node              929 drivers/gpu/drm/radeon/radeon_ttm.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              930 drivers/gpu/drm/radeon/radeon_ttm.c 	unsigned ttm_pl = *(int*)node->info_ent->data;
node              931 drivers/gpu/drm/radeon/radeon_ttm.c 	struct drm_device *dev = node->minor->dev;
node              312 drivers/gpu/drm/radeon/rs400.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              313 drivers/gpu/drm/radeon/rs400.c 	struct drm_device *dev = node->minor->dev;
node              240 drivers/gpu/drm/radeon/rv515.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              241 drivers/gpu/drm/radeon/rv515.c 	struct drm_device *dev = node->minor->dev;
node              258 drivers/gpu/drm/radeon/rv515.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              259 drivers/gpu/drm/radeon/rv515.c 	struct drm_device *dev = node->minor->dev;
node               32 drivers/gpu/drm/rcar-du/rcar_du_encoder.c static unsigned int rcar_du_encoder_count_ports(struct device_node *node)
node               38 drivers/gpu/drm/rcar-du/rcar_du_encoder.c 	ports = of_get_child_by_name(node, "ports");
node               40 drivers/gpu/drm/rcar-du/rcar_du_encoder.c 		ports = of_node_get(node);
node              710 drivers/gpu/drm/rcar-du/rcar_lvds.c 	struct device_node *node;
node              741 drivers/gpu/drm/rcar-du/rcar_lvds.c 	for_each_endpoint_of_node(remote, node) {
node              742 drivers/gpu/drm/rcar-du/rcar_lvds.c 		if (node != remote_input) {
node              748 drivers/gpu/drm/rcar-du/rcar_lvds.c 			of_node_put(node);
node              694 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 	struct device_node *node = NULL, *local;
node              702 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 	while ((node = of_find_compatible_node(node, NULL,
node              707 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 		if (node == dsi->dev->of_node)
node              710 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 		remote = of_graph_get_remote_node(node, 1, 0);
node              719 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 			pdev = of_find_device_by_node(node);
node              727 drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c 			of_node_put(node);
node              274 drivers/gpu/drm/rockchip/rockchip_drm_drv.c 	struct device_node *node = of_graph_get_remote_port_parent(ep);
node              279 drivers/gpu/drm/rockchip/rockchip_drm_drv.c 	if (!node)
node              283 drivers/gpu/drm/rockchip/rockchip_drm_drv.c 	pdev = of_find_device_by_node(node);
node              284 drivers/gpu/drm/rockchip/rockchip_drm_drv.c 	of_node_put(node);
node              275 drivers/gpu/drm/scheduler/sched_main.c 	list_add_tail(&s_job->node, &sched->ring_mirror_list);
node              288 drivers/gpu/drm/scheduler/sched_main.c 				       struct drm_sched_job, node);
node              378 drivers/gpu/drm/scheduler/sched_main.c 	list_for_each_entry_safe_reverse(s_job, tmp, &sched->ring_mirror_list, node) {
node              389 drivers/gpu/drm/scheduler/sched_main.c 			list_del_init(&s_job->node);
node              441 drivers/gpu/drm/scheduler/sched_main.c 	list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) {
node              484 drivers/gpu/drm/scheduler/sched_main.c 	list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) {
node              542 drivers/gpu/drm/scheduler/sched_main.c 	INIT_LIST_HEAD(&job->node);
node              658 drivers/gpu/drm/scheduler/sched_main.c 				       struct drm_sched_job, node);
node              662 drivers/gpu/drm/scheduler/sched_main.c 		list_del_init(&job->node);
node              108 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *check, *found;
node              117 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node(node, mm) {
node              118 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->start != addr) {
node              120 drivers/gpu/drm/selftests/test-drm_mm.c 			       n, addr, node->start);
node              124 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->size != size) {
node              126 drivers/gpu/drm/selftests/test-drm_mm.c 			       n, size, node->size);
node              130 drivers/gpu/drm/selftests/test-drm_mm.c 		if (drm_mm_hole_follows(node)) {
node              137 drivers/gpu/drm/selftests/test-drm_mm.c 			if (node != check) {
node              139 drivers/gpu/drm/selftests/test-drm_mm.c 				       node->start, check->start);
node              157 drivers/gpu/drm/selftests/test-drm_mm.c static u64 misalignment(struct drm_mm_node *node, u64 alignment)
node              164 drivers/gpu/drm/selftests/test-drm_mm.c 	div64_u64_rem(node->start, alignment, &rem);
node              168 drivers/gpu/drm/selftests/test-drm_mm.c static bool assert_node(struct drm_mm_node *node, struct drm_mm *mm,
node              173 drivers/gpu/drm/selftests/test-drm_mm.c 	if (!drm_mm_node_allocated(node) || node->mm != mm) {
node              178 drivers/gpu/drm/selftests/test-drm_mm.c 	if (node->size != size) {
node              180 drivers/gpu/drm/selftests/test-drm_mm.c 		       node->size, size);
node              184 drivers/gpu/drm/selftests/test-drm_mm.c 	if (misalignment(node, alignment)) {
node              186 drivers/gpu/drm/selftests/test-drm_mm.c 		       node->start, misalignment(node, alignment), alignment);
node              190 drivers/gpu/drm/selftests/test-drm_mm.c 	if (node->color != color) {
node              192 drivers/gpu/drm/selftests/test-drm_mm.c 		       node->color, color);
node              299 drivers/gpu/drm/selftests/test-drm_mm.c static struct drm_mm_node *set_node(struct drm_mm_node *node,
node              302 drivers/gpu/drm/selftests/test-drm_mm.c 	node->start = start;
node              303 drivers/gpu/drm/selftests/test-drm_mm.c 	node->size = size;
node              304 drivers/gpu/drm/selftests/test-drm_mm.c 	return node;
node              307 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_reserve_fail(struct drm_mm *mm, struct drm_mm_node *node)
node              311 drivers/gpu/drm/selftests/test-drm_mm.c 	err = drm_mm_reserve_node(mm, node);
node              317 drivers/gpu/drm/selftests/test-drm_mm.c 		       node->start, node->size);
node              318 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              321 drivers/gpu/drm/selftests/test-drm_mm.c 		       err, -ENOSPC, node->start, node->size);
node              375 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node tmp, *nodes, *node, *next;
node              466 drivers/gpu/drm/selftests/test-drm_mm.c 			node = &nodes[order[(o + m) % count]];
node              467 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node              471 drivers/gpu/drm/selftests/test-drm_mm.c 			node = &nodes[order[(o + m) % count]];
node              472 drivers/gpu/drm/selftests/test-drm_mm.c 			err = drm_mm_reserve_node(&mm, node);
node              475 drivers/gpu/drm/selftests/test-drm_mm.c 				       m, n, node->start);
node              489 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node              490 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              525 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_insert(struct drm_mm *mm, struct drm_mm_node *node,
node              531 drivers/gpu/drm/selftests/test-drm_mm.c 	err = drm_mm_insert_node_generic(mm, node,
node              540 drivers/gpu/drm/selftests/test-drm_mm.c 	if (!assert_node(node, mm, size, alignment, color)) {
node              541 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              573 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
node              598 drivers/gpu/drm/selftests/test-drm_mm.c 			node = replace ? &tmp : &nodes[n];
node              599 drivers/gpu/drm/selftests/test-drm_mm.c 			memset(node, 0, sizeof(*node));
node              600 drivers/gpu/drm/selftests/test-drm_mm.c 			if (!expect_insert(&mm, node, size, 0, n, mode)) {
node              661 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node              662 drivers/gpu/drm/selftests/test-drm_mm.c 				drm_mm_remove_node(node);
node              666 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node              667 drivers/gpu/drm/selftests/test-drm_mm.c 				if (!expect_insert(&mm, node, size, 0, n, mode)) {
node              683 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
node              684 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node              692 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node              693 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              762 drivers/gpu/drm/selftests/test-drm_mm.c static bool expect_insert_in_range(struct drm_mm *mm, struct drm_mm_node *node,
node              769 drivers/gpu/drm/selftests/test-drm_mm.c 	err = drm_mm_insert_node_in_range(mm, node,
node              780 drivers/gpu/drm/selftests/test-drm_mm.c 	if (!assert_node(node, mm, size, alignment, color)) {
node              781 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              820 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node;
node              827 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node(node, mm) {
node              828 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->start < start || node->start + node->size > end) {
node              830 drivers/gpu/drm/selftests/test-drm_mm.c 			       n, node->start, node->start + node->size, start, end);
node              834 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->start != n * size) {
node              836 drivers/gpu/drm/selftests/test-drm_mm.c 			       n, n * size, node->start);
node              840 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->size != size) {
node              842 drivers/gpu/drm/selftests/test-drm_mm.c 			       n, size, node->size);
node              846 drivers/gpu/drm/selftests/test-drm_mm.c 		if (drm_mm_hole_follows(node) &&
node              847 drivers/gpu/drm/selftests/test-drm_mm.c 		    drm_mm_hole_node_end(node) < end) {
node              856 drivers/gpu/drm/selftests/test-drm_mm.c 		node = __drm_mm_interval_first(mm, 0, start - 1);
node              857 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->allocated) {
node              859 drivers/gpu/drm/selftests/test-drm_mm.c 			       node->start, node->size, start);
node              865 drivers/gpu/drm/selftests/test-drm_mm.c 		node = __drm_mm_interval_first(mm, end, U64_MAX);
node              866 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->allocated) {
node              868 drivers/gpu/drm/selftests/test-drm_mm.c 			       node->start, node->size, end);
node              880 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
node              947 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
node              948 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node              956 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node              957 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1041 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
node             1073 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
node             1074 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node             1082 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             1083 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1093 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
node             1104 drivers/gpu/drm/selftests/test-drm_mm.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1105 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!node) {
node             1112 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!expect_insert(&mm, node,
node             1125 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm) {
node             1126 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1127 drivers/gpu/drm/selftests/test-drm_mm.c 		kfree(node);
node             1183 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node node;
node             1200 drivers/gpu/drm/selftests/test-drm_mm.c 		if (drm_mm_scan_add_block(scan, &e->node))
node             1204 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!drm_mm_scan_remove_block(scan, &e->node))
node             1214 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(&e->node);
node             1217 drivers/gpu/drm/selftests/test-drm_mm.c 		struct drm_mm_node *node;
node             1219 drivers/gpu/drm/selftests/test-drm_mm.c 		while ((node = drm_mm_scan_color_evict(scan))) {
node             1220 drivers/gpu/drm/selftests/test-drm_mm.c 			e = container_of(node, typeof(*e), node);
node             1221 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(&e->node);
node             1241 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node;
node             1248 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_scan_add_block(&scan, &e->node);
node             1251 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_scan_remove_block(&scan, &e->node);
node             1256 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!drm_mm_node_allocated(&e->node)) {
node             1264 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node(node, mm) {
node             1265 drivers/gpu/drm/selftests/test-drm_mm.c 		e = container_of(node, typeof(*e), node);
node             1278 drivers/gpu/drm/selftests/test-drm_mm.c 	return assert_continuous(mm, nodes[0].node.size);
node             1295 drivers/gpu/drm/selftests/test-drm_mm.c 		if (drm_mm_scan_add_block(&scan, &e->node))
node             1301 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!drm_mm_scan_remove_block(&scan, &e->node)) {
node             1304 drivers/gpu/drm/selftests/test-drm_mm.c 				       e->node.start);
node             1313 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(&e->node);
node             1319 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_reserve_node(mm, &e->node);
node             1322 drivers/gpu/drm/selftests/test-drm_mm.c 			       e->node.start);
node             1327 drivers/gpu/drm/selftests/test-drm_mm.c 	return assert_continuous(mm, nodes[0].node.size);
node             1385 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_reserve_node(mm, &e->node);
node             1388 drivers/gpu/drm/selftests/test-drm_mm.c 			       e->node.start);
node             1393 drivers/gpu/drm/selftests/test-drm_mm.c 	if (!assert_continuous(mm, nodes[0].node.size)) {
node             1408 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
node             1431 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_insert_node(&mm, &nodes[n].node, 1);
node             1501 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             1502 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1521 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
node             1541 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_insert_node(&mm, &nodes[n].node, 1);
node             1598 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             1599 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1608 drivers/gpu/drm/selftests/test-drm_mm.c static unsigned int node_index(const struct drm_mm_node *node)
node             1610 drivers/gpu/drm/selftests/test-drm_mm.c 	return div64_u64(node->start, node->size);
node             1621 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
node             1670 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node             1671 drivers/gpu/drm/selftests/test-drm_mm.c 				drm_mm_remove_node(node);
node             1672 drivers/gpu/drm/selftests/test-drm_mm.c 				__set_bit(node_index(node), bitmap);
node             1678 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node             1679 drivers/gpu/drm/selftests/test-drm_mm.c 				if (!expect_insert(&mm, node,
node             1686 drivers/gpu/drm/selftests/test-drm_mm.c 				if (drm_mm_hole_follows(node)) {
node             1688 drivers/gpu/drm/selftests/test-drm_mm.c 					       m, n, node->start);
node             1693 drivers/gpu/drm/selftests/test-drm_mm.c 				if (node_index(node) != last) {
node             1695 drivers/gpu/drm/selftests/test-drm_mm.c 					       m, n, size, last, node_index(node));
node             1707 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
node             1708 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node             1715 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             1716 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1735 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *nodes, *node, *next;
node             1777 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node             1778 drivers/gpu/drm/selftests/test-drm_mm.c 				drm_mm_remove_node(node);
node             1779 drivers/gpu/drm/selftests/test-drm_mm.c 				__set_bit(node_index(node), bitmap);
node             1785 drivers/gpu/drm/selftests/test-drm_mm.c 				node = &nodes[order[(o + m) % count]];
node             1786 drivers/gpu/drm/selftests/test-drm_mm.c 				if (!expect_insert(&mm, node,
node             1794 drivers/gpu/drm/selftests/test-drm_mm.c 				if (node_index(node) != first) {
node             1796 drivers/gpu/drm/selftests/test-drm_mm.c 					       m, n, first, node_index(node));
node             1807 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, next, &mm)
node             1808 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node             1815 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             1816 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1830 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node rsvd_lo, rsvd_hi, node;
node             1859 drivers/gpu/drm/selftests/test-drm_mm.c 	memset(&node, 0, sizeof(node));
node             1860 drivers/gpu/drm/selftests/test-drm_mm.c 	err = drm_mm_insert_node_generic(&mm, &node,
node             1865 drivers/gpu/drm/selftests/test-drm_mm.c 		       node.start);
node             1870 drivers/gpu/drm/selftests/test-drm_mm.c 	err = drm_mm_insert_node_generic(&mm, &node, 2, 0, 0, mode);
node             1878 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_remove_node(&node);
node             1898 drivers/gpu/drm/selftests/test-drm_mm.c static void separate_adjacent_colors(const struct drm_mm_node *node,
node             1903 drivers/gpu/drm/selftests/test-drm_mm.c 	if (node->allocated && node->color != color)
node             1906 drivers/gpu/drm/selftests/test-drm_mm.c 	node = list_next_entry(node, node_list);
node             1907 drivers/gpu/drm/selftests/test-drm_mm.c 	if (node->allocated && node->color != color)
node             1911 drivers/gpu/drm/selftests/test-drm_mm.c static bool colors_abutt(const struct drm_mm_node *node)
node             1913 drivers/gpu/drm/selftests/test-drm_mm.c 	if (!drm_mm_hole_follows(node) &&
node             1914 drivers/gpu/drm/selftests/test-drm_mm.c 	    list_next_entry(node, node_list)->allocated) {
node             1916 drivers/gpu/drm/selftests/test-drm_mm.c 		       node->color, node->start, node->size,
node             1917 drivers/gpu/drm/selftests/test-drm_mm.c 		       list_next_entry(node, node_list)->color,
node             1918 drivers/gpu/drm/selftests/test-drm_mm.c 		       list_next_entry(node, node_list)->start,
node             1919 drivers/gpu/drm/selftests/test-drm_mm.c 		       list_next_entry(node, node_list)->size);
node             1931 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *nn;
node             1945 drivers/gpu/drm/selftests/test-drm_mm.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1946 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!node) {
node             1951 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!expect_insert(&mm, node,
node             1955 drivers/gpu/drm/selftests/test-drm_mm.c 			kfree(node);
node             1960 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, nn, &mm) {
node             1961 drivers/gpu/drm/selftests/test-drm_mm.c 		if (node->color != node->size) {
node             1963 drivers/gpu/drm/selftests/test-drm_mm.c 			       node->size, node->color);
node             1968 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             1969 drivers/gpu/drm/selftests/test-drm_mm.c 		kfree(node);
node             1977 drivers/gpu/drm/selftests/test-drm_mm.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1978 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!node) {
node             1983 drivers/gpu/drm/selftests/test-drm_mm.c 		node->size = 1 + 2*count;
node             1984 drivers/gpu/drm/selftests/test-drm_mm.c 		node->color = node->size;
node             1986 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_reserve_node(&mm, node);
node             1993 drivers/gpu/drm/selftests/test-drm_mm.c 		last = node->start + node->size;
node             1998 drivers/gpu/drm/selftests/test-drm_mm.c 			node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1999 drivers/gpu/drm/selftests/test-drm_mm.c 			if (!node) {
node             2004 drivers/gpu/drm/selftests/test-drm_mm.c 			node->start = last;
node             2005 drivers/gpu/drm/selftests/test-drm_mm.c 			node->size = n + count;
node             2006 drivers/gpu/drm/selftests/test-drm_mm.c 			node->color = node->size;
node             2008 drivers/gpu/drm/selftests/test-drm_mm.c 			err = drm_mm_reserve_node(&mm, node);
node             2015 drivers/gpu/drm/selftests/test-drm_mm.c 			node->start += n + 1;
node             2016 drivers/gpu/drm/selftests/test-drm_mm.c 			rem = misalignment(node, n + count);
node             2017 drivers/gpu/drm/selftests/test-drm_mm.c 			node->start += n + count - rem;
node             2019 drivers/gpu/drm/selftests/test-drm_mm.c 			err = drm_mm_reserve_node(&mm, node);
node             2026 drivers/gpu/drm/selftests/test-drm_mm.c 			last = node->start + node->size;
node             2030 drivers/gpu/drm/selftests/test-drm_mm.c 			node = kzalloc(sizeof(*node), GFP_KERNEL);
node             2031 drivers/gpu/drm/selftests/test-drm_mm.c 			if (!node) {
node             2036 drivers/gpu/drm/selftests/test-drm_mm.c 			if (!expect_insert(&mm, node,
node             2041 drivers/gpu/drm/selftests/test-drm_mm.c 				kfree(node);
node             2046 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_for_each_node_safe(node, nn, &mm) {
node             2049 drivers/gpu/drm/selftests/test-drm_mm.c 			if (node->color != node->size) {
node             2051 drivers/gpu/drm/selftests/test-drm_mm.c 				       mode->name, node->size, node->color);
node             2056 drivers/gpu/drm/selftests/test-drm_mm.c 			if (colors_abutt(node))
node             2059 drivers/gpu/drm/selftests/test-drm_mm.c 			div64_u64_rem(node->start, node->size, &rem);
node             2062 drivers/gpu/drm/selftests/test-drm_mm.c 				       mode->name, node->start, node->size, rem);
node             2066 drivers/gpu/drm/selftests/test-drm_mm.c 			drm_mm_remove_node(node);
node             2067 drivers/gpu/drm/selftests/test-drm_mm.c 			kfree(node);
node             2075 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, nn, &mm) {
node             2076 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             2077 drivers/gpu/drm/selftests/test-drm_mm.c 		kfree(node);
node             2140 drivers/gpu/drm/selftests/test-drm_mm.c 		err = drm_mm_reserve_node(mm, &e->node);
node             2143 drivers/gpu/drm/selftests/test-drm_mm.c 			       e->node.start);
node             2160 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
node             2183 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!expect_insert(&mm, &nodes[n].node,
node             2242 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             2243 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node             2263 drivers/gpu/drm/selftests/test-drm_mm.c 	struct drm_mm_node *node, *next;
node             2284 drivers/gpu/drm/selftests/test-drm_mm.c 		if (!expect_insert(&mm, &nodes[n].node,
node             2343 drivers/gpu/drm/selftests/test-drm_mm.c 	drm_mm_for_each_node_safe(node, next, &mm)
node             2344 drivers/gpu/drm/selftests/test-drm_mm.c 		drm_mm_remove_node(node);
node              108 drivers/gpu/drm/sti/sti_cursor.c 	struct drm_info_node *node = s->private;
node              109 drivers/gpu/drm/sti/sti_cursor.c 	struct sti_cursor *cursor = (struct sti_cursor *)node->info_ent->data;
node               77 drivers/gpu/drm/sti/sti_drv.c 	struct drm_info_node *node = s->private;
node               78 drivers/gpu/drm/sti/sti_drv.c 	struct drm_device *dev = node->minor->dev;
node              258 drivers/gpu/drm/sti/sti_drv.c 	struct device_node *node = dev->of_node;
node              266 drivers/gpu/drm/sti/sti_drv.c 	child_np = of_get_next_available_child(node, NULL);
node              271 drivers/gpu/drm/sti/sti_drv.c 		child_np = of_get_next_available_child(node, child_np);
node              180 drivers/gpu/drm/sti/sti_dvo.c 	struct drm_info_node *node = s->private;
node              181 drivers/gpu/drm/sti/sti_dvo.c 	struct sti_dvo *dvo = (struct sti_dvo *)node->info_ent->data;
node              215 drivers/gpu/drm/sti/sti_gdp.c 	struct drm_info_node *node = s->private;
node              216 drivers/gpu/drm/sti/sti_gdp.c 	struct sti_gdp *gdp = (struct sti_gdp *)node->info_ent->data;
node              258 drivers/gpu/drm/sti/sti_gdp.c static void gdp_node_dump_node(struct seq_file *s, struct sti_gdp_node *node)
node              260 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\t@:0x%p", node);
node              261 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tCTL  0x%08X", node->gam_gdp_ctl);
node              262 drivers/gpu/drm/sti/sti_gdp.c 	gdp_dbg_ctl(s, node->gam_gdp_ctl);
node              263 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tAGC  0x%08X", node->gam_gdp_agc);
node              264 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tVPO  0x%08X", node->gam_gdp_vpo);
node              265 drivers/gpu/drm/sti/sti_gdp.c 	gdp_dbg_vpo(s, node->gam_gdp_vpo);
node              266 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tVPS  0x%08X", node->gam_gdp_vps);
node              267 drivers/gpu/drm/sti/sti_gdp.c 	gdp_dbg_vps(s, node->gam_gdp_vps);
node              268 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tPML  0x%08X", node->gam_gdp_pml);
node              269 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tPMP  0x%08X", node->gam_gdp_pmp);
node              270 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tSIZE 0x%08X", node->gam_gdp_size);
node              271 drivers/gpu/drm/sti/sti_gdp.c 	gdp_dbg_size(s, node->gam_gdp_size);
node              272 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tNVN  0x%08X", node->gam_gdp_nvn);
node              273 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tKEY1 0x%08X", node->gam_gdp_key1);
node              274 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tKEY2 0x%08X", node->gam_gdp_key2);
node              275 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tPPT  0x%08X", node->gam_gdp_ppt);
node              276 drivers/gpu/drm/sti/sti_gdp.c 	gdp_dbg_ppt(s, node->gam_gdp_ppt);
node              277 drivers/gpu/drm/sti/sti_gdp.c 	seq_printf(s, "\n\tCML  0x%08X\n", node->gam_gdp_cml);
node              282 drivers/gpu/drm/sti/sti_gdp.c 	struct drm_info_node *node = s->private;
node              283 drivers/gpu/drm/sti/sti_gdp.c 	struct sti_gdp *gdp = (struct sti_gdp *)node->info_ent->data;
node              346 drivers/gpu/drm/sti/sti_hda.c 	struct drm_info_node *node = s->private;
node              347 drivers/gpu/drm/sti/sti_hda.c 	struct sti_hda *hda = (struct sti_hda *)node->info_ent->data;
node              663 drivers/gpu/drm/sti/sti_hdmi.c 	struct drm_info_node *node = s->private;
node              664 drivers/gpu/drm/sti/sti_hdmi.c 	struct sti_hdmi *hdmi = (struct sti_hdmi *)node->info_ent->data;
node              565 drivers/gpu/drm/sti/sti_hqvdp.c 	struct drm_info_node *node = s->private;
node              566 drivers/gpu/drm/sti/sti_hqvdp.c 	struct sti_hqvdp *hqvdp = (struct sti_hqvdp *)node->info_ent->data;
node              150 drivers/gpu/drm/sti/sti_mixer.c 	struct drm_info_node *node = s->private;
node              151 drivers/gpu/drm/sti/sti_mixer.c 	struct sti_mixer *mixer = (struct sti_mixer *)node->info_ent->data;
node              499 drivers/gpu/drm/sti/sti_tvout.c 	struct drm_info_node *node = s->private;
node              500 drivers/gpu/drm/sti/sti_tvout.c 	struct sti_tvout *tvout = (struct sti_tvout *)node->info_ent->data;
node              842 drivers/gpu/drm/sti/sti_tvout.c 	struct device_node *node = dev->of_node;
node              848 drivers/gpu/drm/sti/sti_tvout.c 	if (!node)
node               95 drivers/gpu/drm/sti/sti_vid.c 	struct drm_info_node *node = s->private;
node               96 drivers/gpu/drm/sti/sti_vid.c 	struct sti_vid *vid = (struct sti_vid *)node->info_ent->data;
node              712 drivers/gpu/drm/sun4i/sun4i_backend.c static int sun4i_backend_of_get_id(struct device_node *node)
node              718 drivers/gpu/drm/sun4i/sun4i_backend.c 	ep = of_graph_get_endpoint_by_regs(node, 0, -1);
node              734 drivers/gpu/drm/sun4i/sun4i_backend.c 							  struct device_node *node)
node              739 drivers/gpu/drm/sun4i/sun4i_backend.c 	port = of_graph_get_port_by_id(node, 0);
node              751 drivers/gpu/drm/sun4i/sun4i_backend.c 			if (remote == frontend->node) {
node              818 drivers/gpu/drm/sun4i/sun4i_backend.c 	backend->engine.node = dev->of_node;
node              148 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_connector(struct device_node *node)
node              150 drivers/gpu/drm/sun4i/sun4i_drv.c 	return of_device_is_compatible(node, "hdmi-connector");
node              153 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_frontend(struct device_node *node)
node              155 drivers/gpu/drm/sun4i/sun4i_drv.c 	return of_device_is_compatible(node, "allwinner,sun4i-a10-display-frontend") ||
node              156 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun5i-a13-display-frontend") ||
node              157 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun6i-a31-display-frontend") ||
node              158 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun7i-a20-display-frontend") ||
node              159 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun8i-a23-display-frontend") ||
node              160 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun8i-a33-display-frontend") ||
node              161 drivers/gpu/drm/sun4i/sun4i_drv.c 		of_device_is_compatible(node, "allwinner,sun9i-a80-display-frontend");
node              164 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_deu(struct device_node *node)
node              166 drivers/gpu/drm/sun4i/sun4i_drv.c 	return of_device_is_compatible(node, "allwinner,sun9i-a80-deu");
node              169 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_supported_frontend(struct device_node *node)
node              172 drivers/gpu/drm/sun4i/sun4i_drv.c 		return !!of_match_node(sun4i_frontend_of_table, node);
node              177 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon(struct device_node *node)
node              179 drivers/gpu/drm/sun4i/sun4i_drv.c 	return !!of_match_node(sun4i_tcon_of_table, node);
node              182 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon_with_ch0(struct device_node *node)
node              186 drivers/gpu/drm/sun4i/sun4i_drv.c 	match = of_match_node(sun4i_tcon_of_table, node);
node              198 drivers/gpu/drm/sun4i/sun4i_drv.c static bool sun4i_drv_node_is_tcon_top(struct device_node *node)
node              201 drivers/gpu/drm/sun4i/sun4i_drv.c 		!!of_match_node(sun8i_tcon_top_of_table, node);
node              239 drivers/gpu/drm/sun4i/sun4i_drv.c 					 struct device_node *node,
node              244 drivers/gpu/drm/sun4i/sun4i_drv.c 	port = of_graph_get_port_by_id(node, port_id);
node              257 drivers/gpu/drm/sun4i/sun4i_drv.c 		if (sun4i_drv_node_is_tcon(node)) {
node              275 drivers/gpu/drm/sun4i/sun4i_drv.c 			if (sun4i_drv_node_is_tcon_with_ch0(node)) {
node              299 drivers/gpu/drm/sun4i/sun4i_drv.c 				   struct device_node *node)
node              310 drivers/gpu/drm/sun4i/sun4i_drv.c 	if (!sun4i_drv_node_is_frontend(node) &&
node              311 drivers/gpu/drm/sun4i/sun4i_drv.c 	    !of_device_is_available(node))
node              318 drivers/gpu/drm/sun4i/sun4i_drv.c 	if (sun4i_drv_node_is_connector(node))
node              326 drivers/gpu/drm/sun4i/sun4i_drv.c 	if (!(sun4i_drv_node_is_frontend(node) ||
node              327 drivers/gpu/drm/sun4i/sun4i_drv.c 	      sun4i_drv_node_is_deu(node)) ||
node              328 drivers/gpu/drm/sun4i/sun4i_drv.c 	    (sun4i_drv_node_is_supported_frontend(node) &&
node              329 drivers/gpu/drm/sun4i/sun4i_drv.c 	     of_device_is_available(node))) {
node              331 drivers/gpu/drm/sun4i/sun4i_drv.c 		DRM_DEBUG_DRIVER("Adding component %pOF\n", node);
node              332 drivers/gpu/drm/sun4i/sun4i_drv.c 		drm_of_component_match_add(dev, match, compare_of, node);
node              337 drivers/gpu/drm/sun4i/sun4i_drv.c 	sun4i_drv_traverse_endpoints(list, node, 1);
node              340 drivers/gpu/drm/sun4i/sun4i_drv.c 	if (sun4i_drv_node_is_tcon_top(node)) {
node              341 drivers/gpu/drm/sun4i/sun4i_drv.c 		sun4i_drv_traverse_endpoints(list, node, 3);
node              342 drivers/gpu/drm/sun4i/sun4i_drv.c 		sun4i_drv_traverse_endpoints(list, node, 5);
node              571 drivers/gpu/drm/sun4i/sun4i_frontend.c 	frontend->node = dev->of_node;
node              128 drivers/gpu/drm/sun4i/sun4i_frontend.h 	struct device_node	*node;
node              859 drivers/gpu/drm/sun4i/sun4i_tcon.c 				struct device_node *node,
node              866 drivers/gpu/drm/sun4i/sun4i_tcon.c 	port = of_graph_get_port_by_id(node, port_id);
node              893 drivers/gpu/drm/sun4i/sun4i_tcon.c 		if (remote == engine->node)
node              973 drivers/gpu/drm/sun4i/sun4i_tcon.c static bool sun4i_tcon_connected_to_tcon_top(struct device_node *node)
node              978 drivers/gpu/drm/sun4i/sun4i_tcon.c 	remote = of_graph_get_remote_node(node, 0, -1);
node             1038 drivers/gpu/drm/sun4i/sun4i_tcon.c 						   struct device_node *node)
node             1043 drivers/gpu/drm/sun4i/sun4i_tcon.c 	port = of_graph_get_port_by_id(node, 0);
node             1066 drivers/gpu/drm/sun4i/sun4i_tcon.c 		if (sun4i_tcon_connected_to_tcon_top(node))
node             1080 drivers/gpu/drm/sun4i/sun4i_tcon.c 	return sun4i_tcon_find_engine_traverse(drv, node, 0);
node             1293 drivers/gpu/drm/sun4i/sun4i_tcon.c 	struct device_node *node = pdev->dev.of_node;
node             1303 drivers/gpu/drm/sun4i/sun4i_tcon.c 		ret = drm_of_find_panel_or_bridge(node, 1, 0, &panel, &bridge);
node              269 drivers/gpu/drm/sun4i/sun4i_tcon.h struct drm_bridge *sun4i_tcon_find_bridge(struct device_node *node);
node              270 drivers/gpu/drm/sun4i/sun4i_tcon.h struct drm_panel *sun4i_tcon_find_panel(struct device_node *node);
node               62 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c static bool sun8i_dw_hdmi_node_is_tcon_top(struct device_node *node)
node               65 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c 		!!of_match_node(sun8i_tcon_top_of_table, node);
node               69 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c 					     struct device_node *node)
node               74 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c 	remote = of_graph_get_remote_node(node, 0, -1);
node               91 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.c 		crtcs = drm_of_find_possible_crtcs(drm, node);
node              203 drivers/gpu/drm/sun4i/sun8i_dw_hdmi.h int sun8i_hdmi_phy_probe(struct sun8i_dw_hdmi *hdmi, struct device_node *node);
node              607 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c int sun8i_hdmi_phy_probe(struct sun8i_dw_hdmi *hdmi, struct device_node *node)
node              616 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 	match = of_match_node(sun8i_hdmi_phy_of_table, node);
node              628 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 	ret = of_address_to_resource(node, 0, &res);
node              647 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 	phy->clk_bus = of_clk_get_by_name(node, "bus");
node              653 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 	phy->clk_mod = of_clk_get_by_name(node, "mod");
node              661 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 		phy->clk_pll0 = of_clk_get_by_name(node, "pll-0");
node              669 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 			phy->clk_pll1 = of_clk_get_by_name(node, "pll-1");
node              678 drivers/gpu/drm/sun4i/sun8i_hdmi_phy.c 	phy->rst_phy = of_reset_control_get_shared(node, "phy");
node              402 drivers/gpu/drm/sun4i/sun8i_mixer.c static int sun8i_mixer_of_get_id(struct device_node *node)
node              408 drivers/gpu/drm/sun4i/sun8i_mixer.c 	ep = of_graph_get_endpoint_by_regs(node, 1, -1);
node              453 drivers/gpu/drm/sun4i/sun8i_mixer.c 	mixer->engine.node = dev->of_node;
node               23 drivers/gpu/drm/sun4i/sun8i_tcon_top.c static bool sun8i_tcon_top_node_is_tcon_top(struct device_node *node)
node               25 drivers/gpu/drm/sun4i/sun8i_tcon_top.c 	return !!of_match_node(sun8i_tcon_top_of_table, node);
node              123 drivers/gpu/drm/sun4i/sunxi_engine.h 	struct device_node		*node;
node               96 drivers/gpu/drm/tegra/dc.c 		if (it.node == dev->of_node)
node             1417 drivers/gpu/drm/tegra/dc.c 	struct drm_info_node *node = s->private;
node             1418 drivers/gpu/drm/tegra/dc.c 	struct tegra_dc *dc = node->info_ent->data;
node             1443 drivers/gpu/drm/tegra/dc.c 	struct drm_info_node *node = s->private;
node             1444 drivers/gpu/drm/tegra/dc.c 	struct tegra_dc *dc = node->info_ent->data;
node             1474 drivers/gpu/drm/tegra/dc.c 	struct drm_info_node *node = s->private;
node             1475 drivers/gpu/drm/tegra/dc.c 	struct tegra_dc *dc = node->info_ent->data;
node              966 drivers/gpu/drm/tegra/drm.c 	struct drm_info_node *node = (struct drm_info_node *)s->private;
node              967 drivers/gpu/drm/tegra/drm.c 	struct drm_device *drm = node->minor->dev;
node              987 drivers/gpu/drm/tegra/drm.c 	struct drm_info_node *node = (struct drm_info_node *)s->private;
node              988 drivers/gpu/drm/tegra/drm.c 	struct drm_device *drm = node->minor->dev;
node              200 drivers/gpu/drm/tegra/dsi.c 	struct drm_info_node *node = s->private;
node              201 drivers/gpu/drm/tegra/dsi.c 	struct tegra_dsi *dsi = node->info_ent->data;
node              203 drivers/gpu/drm/tegra/dsi.c 	struct drm_device *drm = node->minor->dev;
node             1030 drivers/gpu/drm/tegra/hdmi.c 	struct drm_info_node *node = s->private;
node             1031 drivers/gpu/drm/tegra/hdmi.c 	struct tegra_hdmi *hdmi = node->info_ent->data;
node             1033 drivers/gpu/drm/tegra/hdmi.c 	struct drm_device *drm = node->minor->dev;
node             1265 drivers/gpu/drm/tegra/sor.c 	struct drm_info_node *node = s->private;
node             1266 drivers/gpu/drm/tegra/sor.c 	struct tegra_sor *sor = node->info_ent->data;
node             1268 drivers/gpu/drm/tegra/sor.c 	struct drm_device *drm = node->minor->dev;
node             1427 drivers/gpu/drm/tegra/sor.c 	struct drm_info_node *node = s->private;
node             1428 drivers/gpu/drm/tegra/sor.c 	struct tegra_sor *sor = node->info_ent->data;
node             1430 drivers/gpu/drm/tegra/sor.c 	struct drm_device *drm = node->minor->dev;
node              225 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	struct device_node *node = dev->of_node;
node              273 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	if (of_property_read_u32(node, "max-bandwidth", &priv->max_bandwidth))
node              278 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	if (of_property_read_u32(node, "max-width", &priv->max_width))
node              283 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	if (of_property_read_u32(node, "max-pixelclock",
node              319 drivers/gpu/drm/tilcdc/tilcdc_drv.c 		of_property_read_string(node, "blue-and-red-wiring", &str);
node              455 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              456 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	struct drm_device *dev = node->minor->dev;
node              476 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node              477 drivers/gpu/drm/tilcdc/tilcdc_drv.c 	struct drm_device *dev = node->minor->dev;
node              169 drivers/gpu/drm/tilcdc/tilcdc_external.c 	struct device_node *node;
node              171 drivers/gpu/drm/tilcdc/tilcdc_external.c 	node = of_graph_get_remote_node(dev->of_node, 0, 0);
node              173 drivers/gpu/drm/tilcdc/tilcdc_external.c 	if (!of_device_is_compatible(node, "nxp,tda998x")) {
node              174 drivers/gpu/drm/tilcdc/tilcdc_external.c 		of_node_put(node);
node              179 drivers/gpu/drm/tilcdc/tilcdc_external.c 		drm_of_component_match_add(dev, match, dev_match_of, node);
node              180 drivers/gpu/drm/tilcdc/tilcdc_external.c 	of_node_put(node);
node              306 drivers/gpu/drm/tilcdc/tilcdc_panel.c 	struct device_node *bl_node, *node = pdev->dev.of_node;
node              313 drivers/gpu/drm/tilcdc/tilcdc_panel.c 	if (!node) {
node              322 drivers/gpu/drm/tilcdc/tilcdc_panel.c 	bl_node = of_parse_phandle(node, "backlight", 0);
node              353 drivers/gpu/drm/tilcdc/tilcdc_panel.c 	panel_mod->timings = of_get_display_timings(node);
node              360 drivers/gpu/drm/tilcdc/tilcdc_panel.c 	panel_mod->info = of_get_panel_info(node);
node              274 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c 	struct device_node *node = pdev->dev.of_node;
node              283 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c 	if (!node) {
node              301 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c 	if (of_property_read_u32(node, "i2c", &i2c_phandle)) {
node              321 drivers/gpu/drm/tilcdc/tilcdc_tfp410.c 	tfp410_mod->gpio = of_get_named_gpio_flags(node, "powerdn-gpio",
node               55 drivers/gpu/drm/ttm/ttm_agp_backend.c 	struct drm_mm_node *node = bo_mem->mm_node;
node               78 drivers/gpu/drm/ttm/ttm_agp_backend.c 	ret = agp_bind_memory(mem, node->start);
node               58 drivers/gpu/drm/ttm/ttm_bo_manager.c 	struct drm_mm_node *node;
node               67 drivers/gpu/drm/ttm/ttm_bo_manager.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node               68 drivers/gpu/drm/ttm/ttm_bo_manager.c 	if (!node)
node               76 drivers/gpu/drm/ttm/ttm_bo_manager.c 	ret = drm_mm_insert_node_in_range(mm, node,
node               83 drivers/gpu/drm/ttm/ttm_bo_manager.c 		kfree(node);
node               85 drivers/gpu/drm/ttm/ttm_bo_manager.c 		mem->mm_node = node;
node               86 drivers/gpu/drm/ttm/ttm_bo_manager.c 		mem->start = node->start;
node              407 drivers/gpu/drm/ttm/ttm_bo_vm.c 	struct drm_vma_offset_node *node;
node              412 drivers/gpu/drm/ttm/ttm_bo_vm.c 	node = drm_vma_offset_lookup_locked(&bdev->vma_manager, offset, pages);
node              413 drivers/gpu/drm/ttm/ttm_bo_vm.c 	if (likely(node)) {
node              414 drivers/gpu/drm/ttm/ttm_bo_vm.c 		bo = container_of(node, struct ttm_buffer_object,
node              236 drivers/gpu/drm/udl/udl_fb.c 		  info->node, user, info, ufbdev->fb_count);
node              261 drivers/gpu/drm/udl/udl_fb.c 		info->node, user, ufbdev->fb_count);
node              170 drivers/gpu/drm/udl/udl_main.c 	struct list_head *node;
node              182 drivers/gpu/drm/udl/udl_main.c 		node = udl->urbs.list.next; /* have reserved one with sem */
node              183 drivers/gpu/drm/udl/udl_main.c 		list_del_init(node);
node              187 drivers/gpu/drm/udl/udl_main.c 		unode = list_entry(node, struct urb_node, entry);
node              194 drivers/gpu/drm/udl/udl_main.c 		kfree(node);
node               44 drivers/gpu/drm/v3d/v3d_bo.c 	drm_mm_remove_node(&bo->node);
node              107 drivers/gpu/drm/v3d/v3d_bo.c 	ret = drm_mm_insert_node_generic(&v3d->mm, &bo->node,
node              185 drivers/gpu/drm/v3d/v3d_bo.c 	args->offset = bo->node.start << PAGE_SHIFT;
node              230 drivers/gpu/drm/v3d/v3d_bo.c 	args->offset = bo->node.start << PAGE_SHIFT;
node               82 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node               83 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              129 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              130 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              198 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              199 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              214 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              215 drivers/gpu/drm/v3d/v3d_debugfs.c 	struct drm_device *dev = node->minor->dev;
node              143 drivers/gpu/drm/v3d/v3d_drv.h 	struct drm_mm_node node;
node               71 drivers/gpu/drm/v3d/v3d_irq.c 	V3D_CORE_WRITE(0, V3D_PTB_BPOA, bo->node.start << PAGE_SHIFT);
node               91 drivers/gpu/drm/v3d/v3d_mmu.c 	u32 page = bo->node.start;
node              108 drivers/gpu/drm/v3d/v3d_mmu.c 	WARN_ON_ONCE(page - bo->node.start !=
node              121 drivers/gpu/drm/v3d/v3d_mmu.c 	for (page = bo->node.start; page < bo->node.start + npages; page++)
node               68 drivers/gpu/drm/vc4/vc4_bo.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node               69 drivers/gpu/drm/vc4/vc4_bo.c 	struct drm_device *dev = node->minor->dev;
node               45 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node               46 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct debugfs_regset32 *regset = node->info_ent->data;
node              257 drivers/gpu/drm/vc4/vc4_drv.c 	struct device_node *node;
node              267 drivers/gpu/drm/vc4/vc4_drv.c 	node = of_find_matching_node_and_match(NULL, vc4_v3d_dt_match, NULL);
node              268 drivers/gpu/drm/vc4/vc4_drv.c 	if (!node || !of_device_is_available(node))
node              270 drivers/gpu/drm/vc4/vc4_drv.c 	of_node_put(node);
node              184 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              185 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct drm_device *dev = node->minor->dev;
node               85 drivers/gpu/drm/vc4/vc4_hvs.c 	struct drm_info_node *node = m->private;
node               86 drivers/gpu/drm/vc4/vc4_hvs.c 	struct drm_device *dev = node->minor->dev;
node              101 drivers/gpu/drm/vc4/vc4_v3d.c 	struct drm_info_node *node = (struct drm_info_node *)m->private;
node              102 drivers/gpu/drm/vc4/vc4_v3d.c 	struct drm_device *dev = node->minor->dev;
node               45 drivers/gpu/drm/virtio/virtgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               46 drivers/gpu/drm/virtio/virtgpu_debugfs.c 	struct virtio_gpu_device *vgdev = node->minor->dev->dev_private;
node               58 drivers/gpu/drm/virtio/virtgpu_debugfs.c 	struct drm_info_node *node = (struct drm_info_node *) m->private;
node               59 drivers/gpu/drm/virtio/virtgpu_debugfs.c 	struct virtio_gpu_device *vgdev = node->minor->dev->dev_private;
node              104 drivers/gpu/drm/virtio/virtgpu_drv.h 	struct list_head node;
node               98 drivers/gpu/drm/virtio/virtgpu_fence.c 	list_add_tail(&fence->node, &drv->fences);
node              116 drivers/gpu/drm/virtio/virtgpu_fence.c 	list_for_each_entry_safe(fence, tmp, &drv->fences, node) {
node              120 drivers/gpu/drm/virtio/virtgpu_fence.c 		list_del(&fence->node);
node               88 drivers/gpu/drm/vkms/vkms_crtc.c 	*vblank_time = output->vblank_hrtimer.node.expires;
node              158 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct drm_mm_node node;
node              187 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct drm_mm_node *node;
node              267 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	drm_mm_remove_node(&header->node);
node              770 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	memset(info->node, 0, sizeof(*info->node));
node              772 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	ret = drm_mm_insert_node(&man->mm, info->node, info->page_size);
node              775 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		ret = drm_mm_insert_node(&man->mm, info->node, info->page_size);
node              797 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 				  struct drm_mm_node *node,
node              804 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	info.node = node;
node              872 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	ret = vmw_cmdbuf_alloc_space(man, &header->node,  size, interruptible);
node              884 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	header->size = header->node.size << PAGE_SHIFT;
node              886 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	offset = header->node.start << PAGE_SHIFT;
node              900 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	drm_mm_remove_node(&header->node);
node              211 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 				   struct vmw_ctx_validation_info *node)
node              229 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		node->staged = vmw_binding_state_alloc(dev_priv);
node              230 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		if (IS_ERR(node->staged)) {
node              231 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			ret = PTR_ERR(node->staged);
node              232 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			node->staged = NULL;
node              236 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		node->staged = sw_context->staged_bindings;
node              240 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	node->ctx = res;
node              241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	node->cur = vmw_context_binding_state(res);
node              242 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	list_add_tail(&node->head, &sw_context->ctx_list);
node              899 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		struct vmw_ctx_validation_info *node;
node              901 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		node = vmw_execbuf_info_from_res(sw_context, ctx);
node              902 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		if (!node)
node              909 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		vmw_binding_add(node->staged, &binding.bi, 0, binding.slot);
node             1660 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			struct vmw_ctx_validation_info *node;
node             1662 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			node = vmw_execbuf_info_from_res(sw_context, ctx);
node             1663 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			if (!node)
node             1670 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			vmw_binding_add(node->staged, &binding.bi, 0,
node              198 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	list_add(&cb.base.node, &f->cb_list);
node              234 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	if (!list_empty(&cb.base.node))
node              235 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		list_del(&cb.base.node);
node              317 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	struct vmw_validation_res_node *node;
node              320 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	node = vmw_validation_find_res_dup(ctx, res);
node              321 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	if (node) {
node              322 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		node->first_usage = 0;
node              326 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	node = vmw_validation_mem_alloc(ctx, sizeof(*node) + priv_size);
node              327 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	if (!node) {
node              333 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		node->hash.key = (unsigned long) res;
node              334 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		ret = drm_ht_insert_item(ctx->ht, &node->hash);
node              341 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	node->res = vmw_resource_reference_unless_doomed(res);
node              342 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	if (!node->res)
node              345 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 	node->first_usage = 1;
node              347 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		list_add_tail(&node->head, &ctx->resource_list);
node              352 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			list_add(&node->head, &ctx->resource_ctx_list);
node              355 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			list_add_tail(&node->head, &ctx->resource_ctx_list);
node              358 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 			list_add_tail(&node->head, &ctx->resource_list);
node              365 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		node->dirty_set = 1;
node              367 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
node              370 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		*first_usage = node->first_usage;
node              372 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c 		*p_node = &node->private;
node              657 drivers/hid/hid-debug.c 	list_for_each_entry(list, &hdev->debug_list, node)
node             1084 drivers/hid/hid-debug.c 	list_add_tail(&list->node, &list->hdev->debug_list);
node             1169 drivers/hid/hid-debug.c 	list_del(&list->node);
node             1016 drivers/hid/hid-quirks.c 	struct list_head node;
node             1040 drivers/hid/hid-quirks.c 	list_for_each_entry(q, &dquirks_list, node) {
node             1094 drivers/hid/hid-quirks.c 	list_for_each_entry(q, &dquirks_list, node) {
node             1098 drivers/hid/hid-quirks.c 			list_replace(&q->node, &q_new->node);
node             1108 drivers/hid/hid-quirks.c 		list_add_tail(&q_new->node, &dquirks_list);
node             1131 drivers/hid/hid-quirks.c 	list_for_each_entry_safe(q, temp, &dquirks_list, node) {
node             1133 drivers/hid/hid-quirks.c 			list_del(&q->node);
node               61 drivers/hid/hid-roccat.c 	struct list_head node;
node              191 drivers/hid/hid-roccat.c 	list_add_tail(&reader->node, &device->readers);
node              219 drivers/hid/hid-roccat.c 	list_del(&reader->node);
node              268 drivers/hid/hid-roccat.c 	list_for_each_entry(reader, &device->readers, node) {
node              300 drivers/hid/hidraw.c 	list_add_tail(&list->node, &hidraw_table[minor]->list);
node              353 drivers/hid/hidraw.c 	list_del(&list->node);
node              486 drivers/hid/hidraw.c 	list_for_each_entry(list, &dev->list, node) {
node               45 drivers/hid/usbhid/hiddev.c 	struct list_head node;
node              148 drivers/hid/usbhid/hiddev.c 	list_for_each_entry(list, &hiddev->list, node) {
node              222 drivers/hid/usbhid/hiddev.c 	list_del(&list->node);
node              269 drivers/hid/usbhid/hiddev.c 	list_add_tail(&list->node, &hiddev->list);
node              436 drivers/hsi/controllers/omap_ssi_port.c 	struct list_head *node, *tmp;
node              439 drivers/hsi/controllers/omap_ssi_port.c 	list_for_each_safe(node, tmp, queue) {
node              440 drivers/hsi/controllers/omap_ssi_port.c 		msg = list_entry(node, struct hsi_msg, link);
node              443 drivers/hsi/controllers/omap_ssi_port.c 		list_del(node);
node               40 drivers/hv/vmbus_drv.c 	struct list_head node;
node              724 drivers/hv/vmbus_drv.c 	list_for_each_entry(dynid, &drv->dynids.list, node) {
node              775 drivers/hv/vmbus_drv.c 	list_add_tail(&dynid->node, &drv->dynids.list);
node              786 drivers/hv/vmbus_drv.c 	list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) {
node              787 drivers/hv/vmbus_drv.c 		list_del(&dynid->node);
node              838 drivers/hv/vmbus_drv.c 	list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) {
node              842 drivers/hv/vmbus_drv.c 			list_del(&dynid->node);
node             1082 drivers/hwmon/applesmc.c 	struct applesmc_dev_attr *node;
node             1085 drivers/hwmon/applesmc.c 		for (node = grp->nodes; node->sda.dev_attr.attr.name; node++)
node             1087 drivers/hwmon/applesmc.c 					  &node->sda.dev_attr.attr);
node             1099 drivers/hwmon/applesmc.c 	struct applesmc_dev_attr *node;
node             1104 drivers/hwmon/applesmc.c 		grp->nodes = kcalloc(num + 1, sizeof(*node), GFP_KERNEL);
node             1110 drivers/hwmon/applesmc.c 			node = &grp->nodes[i];
node             1111 drivers/hwmon/applesmc.c 			scnprintf(node->name, sizeof(node->name), grp->format,
node             1113 drivers/hwmon/applesmc.c 			node->sda.index = (grp->option << 16) | (i & 0xffff);
node             1114 drivers/hwmon/applesmc.c 			node->sda.dev_attr.show = grp->show;
node             1115 drivers/hwmon/applesmc.c 			node->sda.dev_attr.store = grp->store;
node             1116 drivers/hwmon/applesmc.c 			attr = &node->sda.dev_attr.attr;
node             1118 drivers/hwmon/applesmc.c 			attr->name = node->name;
node              388 drivers/hwmon/ibmpowernv.c 						  struct device_node *node,
node              412 drivers/hwmon/ibmpowernv.c 			if (it.phandle == node->phandle) {
node              413 drivers/hwmon/ibmpowernv.c 				of_node_put(it.node);
node              468 drivers/hwmon/max6697.c static void max6697_get_config_of(struct device_node *node,
node              475 drivers/hwmon/max6697.c 		of_property_read_bool(node, "smbus-timeout-disable");
node              477 drivers/hwmon/max6697.c 		of_property_read_bool(node, "extended-range-enable");
node              479 drivers/hwmon/max6697.c 		of_property_read_bool(node, "beta-compensation-enable");
node              481 drivers/hwmon/max6697.c 	prop = of_get_property(node, "alert-mask", &len);
node              484 drivers/hwmon/max6697.c 	prop = of_get_property(node, "over-temperature-mask", &len);
node              487 drivers/hwmon/max6697.c 	prop = of_get_property(node, "resistance-cancellation", &len);
node              494 drivers/hwmon/max6697.c 	prop = of_get_property(node, "transistor-ideality", &len);
node               76 drivers/hwspinlock/omap_hwspinlock.c 	struct device_node *node = pdev->dev.of_node;
node               85 drivers/hwspinlock/omap_hwspinlock.c 	if (!node)
node              260 drivers/hwtracing/coresight/coresight-catu.c catu_init_sg_table(struct device *catu_dev, int node,
node              271 drivers/hwtracing/coresight/coresight-catu.c 	catu_table = tmc_alloc_sg_table(catu_dev, node, nr_tpages,
node              330 drivers/hwtracing/coresight/coresight-catu.c 			      struct etr_buf *etr_buf, int node, void **pages)
node              343 drivers/hwtracing/coresight/coresight-catu.c 	catu_table = catu_init_sg_table(&csdev->dev, node,
node              377 drivers/hwtracing/coresight/coresight-etb10.c 	int node;
node              380 drivers/hwtracing/coresight/coresight-etb10.c 	node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
node              382 drivers/hwtracing/coresight/coresight-etb10.c 	buf = kzalloc_node(sizeof(struct cs_buffers), GFP_KERNEL, node);
node               84 drivers/hwtracing/coresight/coresight-etm-perf.c 	int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu);
node               86 drivers/hwtracing/coresight/coresight-etm-perf.c 	filters = kzalloc_node(sizeof(struct etm_filters), GFP_KERNEL, node);
node               66 drivers/hwtracing/coresight/coresight-platform.c static void of_coresight_get_ports_legacy(const struct device_node *node,
node               73 drivers/hwtracing/coresight/coresight-platform.c 		ep = of_graph_get_next_endpoint(node, ep);
node              104 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_input_ports_node(const struct device_node *node)
node              106 drivers/hwtracing/coresight/coresight-platform.c 	return of_get_child_by_name(node, "in-ports");
node              110 drivers/hwtracing/coresight/coresight-platform.c of_coresight_get_output_ports_node(const struct device_node *node)
node              112 drivers/hwtracing/coresight/coresight-platform.c 	return of_get_child_by_name(node, "out-ports");
node              126 drivers/hwtracing/coresight/coresight-platform.c static void of_coresight_get_ports(const struct device_node *node,
node              131 drivers/hwtracing/coresight/coresight-platform.c 	input_ports = of_coresight_get_input_ports_node(node);
node              132 drivers/hwtracing/coresight/coresight-platform.c 	output_ports = of_coresight_get_output_ports_node(node);
node              145 drivers/hwtracing/coresight/coresight-platform.c 		of_coresight_get_ports_legacy(node, nr_inport, nr_outport);
node              245 drivers/hwtracing/coresight/coresight-platform.c 	struct device_node *node = dev->of_node;
node              248 drivers/hwtracing/coresight/coresight-platform.c 	of_coresight_get_ports(node, &pdata->nr_inport, &pdata->nr_outport);
node              258 drivers/hwtracing/coresight/coresight-platform.c 	parent = of_coresight_get_output_ports_node(node);
node              266 drivers/hwtracing/coresight/coresight-platform.c 		parent = node;
node              743 drivers/hwtracing/coresight/coresight-stm.c 	list_for_each_entry(rent, &res_list, node) {
node              392 drivers/hwtracing/coresight/coresight-tmc-etf.c 	int node;
node              395 drivers/hwtracing/coresight/coresight-tmc-etf.c 	node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
node              398 drivers/hwtracing/coresight/coresight-tmc-etf.c 	buf = kzalloc_node(sizeof(struct cs_buffers), GFP_KERNEL, node);
node              191 drivers/hwtracing/coresight/coresight-tmc-etr.c 			   struct device *dev, int node,
node              218 drivers/hwtracing/coresight/coresight-tmc-etr.c 			page = alloc_pages_node(node,
node              291 drivers/hwtracing/coresight/coresight-tmc-etr.c 			     sg_table->dev, sg_table->node,
node              316 drivers/hwtracing/coresight/coresight-tmc-etr.c 					int node,
node              329 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sg_table->node = node;
node              556 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_init_etr_sg_table(struct device *dev, int node,
node              570 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sg_table = tmc_alloc_sg_table(dev, node, nr_tpages, nr_dpages, pages);
node              591 drivers/hwtracing/coresight/coresight-tmc-etr.c 				  struct etr_buf *etr_buf, int node,
node              671 drivers/hwtracing/coresight/coresight-tmc-etr.c 				struct etr_buf *etr_buf, int node,
node              677 drivers/hwtracing/coresight/coresight-tmc-etr.c 	etr_table = tmc_init_etr_sg_table(dev, node,
node              797 drivers/hwtracing/coresight/coresight-tmc-etr.c 					 struct etr_buf *etr_buf, int node,
node              808 drivers/hwtracing/coresight/coresight-tmc-etr.c 						      node, pages);
node              827 drivers/hwtracing/coresight/coresight-tmc-etr.c 					 int node, void **pages)
node              862 drivers/hwtracing/coresight/coresight-tmc-etr.c 					    etr_buf, node, pages);
node              865 drivers/hwtracing/coresight/coresight-tmc-etr.c 					    etr_buf, node, pages);
node              868 drivers/hwtracing/coresight/coresight-tmc-etr.c 					    etr_buf, node, pages);
node             1206 drivers/hwtracing/coresight/coresight-tmc-etr.c 	int node;
node             1210 drivers/hwtracing/coresight/coresight-tmc-etr.c 	node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
node             1217 drivers/hwtracing/coresight/coresight-tmc-etr.c 					    0, node, NULL);
node             1228 drivers/hwtracing/coresight/coresight-tmc-etr.c 		etr_buf = tmc_alloc_etr_buf(drvdata, size, 0, node, NULL);
node             1332 drivers/hwtracing/coresight/coresight-tmc-etr.c 	int node;
node             1336 drivers/hwtracing/coresight/coresight-tmc-etr.c 	node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
node             1338 drivers/hwtracing/coresight/coresight-tmc-etr.c 	etr_perf = kzalloc_node(sizeof(*etr_perf), GFP_KERNEL, node);
node              214 drivers/hwtracing/coresight/coresight-tmc.h 		     int node, void **pages);
node              248 drivers/hwtracing/coresight/coresight-tmc.h 	int node;
node              310 drivers/hwtracing/coresight/coresight-tmc.h 					int node,
node              626 drivers/hwtracing/coresight/coresight.c 	struct coresight_node *node;
node              654 drivers/hwtracing/coresight/coresight.c 	node = kzalloc(sizeof(struct coresight_node), GFP_KERNEL);
node              655 drivers/hwtracing/coresight/coresight.c 	if (!node)
node              659 drivers/hwtracing/coresight/coresight.c 	node->csdev = csdev;
node              660 drivers/hwtracing/coresight/coresight.c 	list_add(&node->link, path);
node              193 drivers/hwtracing/intel_th/core.c 	char *node;
node              196 drivers/hwtracing/intel_th/core.c 		node = kasprintf(GFP_KERNEL, "intel_th%d/%s%d", th->id,
node              199 drivers/hwtracing/intel_th/core.c 		node = kasprintf(GFP_KERNEL, "intel_th%d/%s", th->id,
node              202 drivers/hwtracing/intel_th/core.c 	return node;
node               86 drivers/hwtracing/stm/p_sys-t.c 	struct sys_t_policy_node	node;
node              107 drivers/hwtracing/stm/p_sys-t.c 	memcpy(&opriv->node, pn, sizeof(opriv->node));
node              240 drivers/hwtracing/stm/p_sys-t.c 	if (op->node.ts_interval &&
node              241 drivers/hwtracing/stm/p_sys-t.c 	    time_after(jiffies, op->ts_jiffies + op->node.ts_interval)) {
node              252 drivers/hwtracing/stm/p_sys-t.c 	if (op->node.clocksync_interval &&
node              254 drivers/hwtracing/stm/p_sys-t.c 		       op->clocksync_jiffies + op->node.clocksync_interval)) {
node              307 drivers/hwtracing/stm/p_sys-t.c 	if (op->node.do_len)
node              325 drivers/hwtracing/stm/p_sys-t.c 	sz = stm_data_write(data, m, c, false, op->node.uuid.b, UUID_SIZE);
node              330 drivers/hwtracing/stm/p_sys-t.c 	if (op->node.do_len) {
node               83 drivers/hwtracing/stm/policy.c 	struct stp_policy_node *node = to_stp_policy_node(item);
node               85 drivers/hwtracing/stm/policy.c 	return stp_policy_node_priv(node);
node              184 drivers/hwtracing/stm/policy.c 	struct stp_policy_node *node = to_stp_policy_node(item);
node              186 drivers/hwtracing/stm/policy.c 	kfree(node);
node             1331 drivers/i2c/busses/i2c-img-scb.c 	struct device_node *node = pdev->dev.of_node;
node             1375 drivers/i2c/busses/i2c-img-scb.c 	if (!of_property_read_u32(node, "clock-frequency", &val))
node             1380 drivers/i2c/busses/i2c-img-scb.c 	i2c->adap.dev.of_node = node;
node               81 drivers/i2c/busses/i2c-mpc.c 	void (*setup)(struct device_node *node, struct mpc_i2c *i2c, u32 clock);
node              201 drivers/i2c/busses/i2c-mpc.c static int mpc_i2c_get_fdr_52xx(struct device_node *node, u32 clock,
node              211 drivers/i2c/busses/i2c-mpc.c 		*real_clk = mpc5xxx_get_bus_frequency(node) / 2048;
node              216 drivers/i2c/busses/i2c-mpc.c 	divider = mpc5xxx_get_bus_frequency(node) / clock;
node              231 drivers/i2c/busses/i2c-mpc.c 	*real_clk = mpc5xxx_get_bus_frequency(node) / div->divider;
node              235 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_52xx(struct device_node *node,
node              247 drivers/i2c/busses/i2c-mpc.c 	ret = mpc_i2c_get_fdr_52xx(node, clock, &i2c->real_clk);
node              257 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_52xx(struct device_node *node,
node              265 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_512x(struct device_node *node,
node              281 drivers/i2c/busses/i2c-mpc.c 			pval = of_get_property(node, "reg", NULL);
node              290 drivers/i2c/busses/i2c-mpc.c 	mpc_i2c_setup_52xx(node, i2c, clock);
node              293 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_512x(struct device_node *node,
node              323 drivers/i2c/busses/i2c-mpc.c 	struct device_node *node;
node              327 drivers/i2c/busses/i2c-mpc.c 	node = of_find_node_by_name(NULL, "global-utilities");
node              328 drivers/i2c/busses/i2c-mpc.c 	if (node) {
node              329 drivers/i2c/busses/i2c-mpc.c 		const u32 *prop = of_get_property(node, "reg", NULL);
node              348 drivers/i2c/busses/i2c-mpc.c 	of_node_put(node);
node              388 drivers/i2c/busses/i2c-mpc.c static int mpc_i2c_get_fdr_8xxx(struct device_node *node, u32 clock,
node              421 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_8xxx(struct device_node *node,
node              434 drivers/i2c/busses/i2c-mpc.c 	ret = mpc_i2c_get_fdr_8xxx(node, clock, &i2c->real_clk);
node              446 drivers/i2c/busses/i2c-mpc.c static void mpc_i2c_setup_8xxx(struct device_node *node,
node              766 drivers/i2c/busses/i2c-mxs.c 	struct device_node *node = dev->of_node;
node              769 drivers/i2c/busses/i2c-mxs.c 	ret = of_property_read_u32(node, "clock-frequency", &speed);
node              137 drivers/i2c/busses/i2c-octeon-platdrv.c 	struct device_node *node = pdev->dev.of_node;
node              143 drivers/i2c/busses/i2c-octeon-platdrv.c 	cn78xx_style = of_device_is_compatible(node, "cavium,octeon-7890-twsi");
node              182 drivers/i2c/busses/i2c-octeon-platdrv.c 	if (of_property_read_u32(node, "clock-frequency", &i2c->twsi_freq) &&
node              183 drivers/i2c/busses/i2c-octeon-platdrv.c 	    of_property_read_u32(node, "clock-rate", &i2c->twsi_freq)) {
node              244 drivers/i2c/busses/i2c-octeon-platdrv.c 	i2c->adap.dev.of_node = node;
node             1361 drivers/i2c/busses/i2c-omap.c 	struct device_node	*node = pdev->dev.of_node;
node             1390 drivers/i2c/busses/i2c-omap.c 		of_property_read_u32(node, "clock-frequency", &freq);
node               38 drivers/i2c/busses/i2c-parport.c 	struct list_head node;
node              243 drivers/i2c/busses/i2c-parport.c 	list_add_tail(&adapter->node, &adapter_list);
node              260 drivers/i2c/busses/i2c-parport.c 	list_for_each_entry_safe(adapter, _n, &adapter_list, node) {
node              274 drivers/i2c/busses/i2c-parport.c 			list_del(&adapter->node);
node              208 drivers/i2c/busses/i2c-powermac.c 					   struct device_node *node)
node              214 drivers/i2c/busses/i2c-powermac.c 	prop = of_get_property(node, "reg", &len);
node              219 drivers/i2c/busses/i2c-powermac.c 	prop = of_get_property(node, "i2c-address", &len);
node              224 drivers/i2c/busses/i2c-powermac.c 	if (of_node_name_eq(node, "cereal"))
node              226 drivers/i2c/busses/i2c-powermac.c 	else if (of_node_name_eq(node, "deq"))
node              229 drivers/i2c/busses/i2c-powermac.c 	dev_warn(&adap->dev, "No i2c address for %pOF\n", node);
node              277 drivers/i2c/busses/i2c-powermac.c 					    struct device_node *node,
node              293 drivers/i2c/busses/i2c-powermac.c 	if (of_modalias_node(node, tmp, sizeof(tmp)) >= 0) {
node              299 drivers/i2c/busses/i2c-powermac.c 	if (of_node_name_eq(node, "deq")) {
node              310 drivers/i2c/busses/i2c-powermac.c 	dev_err(&adap->dev, "i2c-powermac: modalias failure on %pOF\n", node);
node              318 drivers/i2c/busses/i2c-powermac.c 	struct device_node *node;
node              329 drivers/i2c/busses/i2c-powermac.c 	for_each_child_of_node(adap->dev.of_node, node) {
node              334 drivers/i2c/busses/i2c-powermac.c 		addr = i2c_powermac_get_addr(adap, bus, node);
node              339 drivers/i2c/busses/i2c-powermac.c 		if (!pmac_i2c_match_adapter(node, adap))
node              342 drivers/i2c/busses/i2c-powermac.c 		dev_dbg(&adap->dev, "i2c-powermac: register %pOF\n", node);
node              348 drivers/i2c/busses/i2c-powermac.c 		if (of_device_is_compatible(node, "pcm3052"))
node              352 drivers/i2c/busses/i2c-powermac.c 		if (!i2c_powermac_get_type(adap, node, addr,
node              359 drivers/i2c/busses/i2c-powermac.c 		info.irq = irq_of_parse_and_map(node, 0);
node              360 drivers/i2c/busses/i2c-powermac.c 		info.of_node = of_node_get(node);
node              365 drivers/i2c/busses/i2c-powermac.c 				" %pOF\n", node);
node              366 drivers/i2c/busses/i2c-powermac.c 			of_node_put(node);
node              227 drivers/i2c/busses/i2c-stm32f7.c 	struct list_head node;
node              476 drivers/i2c/busses/i2c-stm32f7.c 					list_add_tail(&v->node,
node              508 drivers/i2c/busses/i2c-stm32f7.c 	list_for_each_entry(v, &solutions, node) {
node              564 drivers/i2c/busses/i2c-stm32f7.c 	list_for_each_entry_safe(v, _v, &solutions, node) {
node              565 drivers/i2c/busses/i2c-stm32f7.c 		list_del(&v->node);
node              119 drivers/i2c/busses/i2c-thunderx-pcidrv.c 				      struct device_node *node)
node              121 drivers/i2c/busses/i2c-thunderx-pcidrv.c 	if (!node)
node              124 drivers/i2c/busses/i2c-thunderx-pcidrv.c 	i2c->alert_data.irq = irq_of_parse_and_map(node, 0);
node              135 drivers/i2c/busses/i2c-thunderx-pcidrv.c 				   struct device_node *node)
node              141 drivers/i2c/busses/i2c-thunderx-pcidrv.c 	return thunder_i2c_smbus_setup_of(i2c, node);
node               22 drivers/i2c/i2c-core-of.c int of_i2c_get_board_info(struct device *dev, struct device_node *node,
node               30 drivers/i2c/i2c-core-of.c 	if (of_modalias_node(node, info->type, sizeof(info->type)) < 0) {
node               31 drivers/i2c/i2c-core-of.c 		dev_err(dev, "of_i2c: modalias failure on %pOF\n", node);
node               35 drivers/i2c/i2c-core-of.c 	ret = of_property_read_u32(node, "reg", &addr);
node               37 drivers/i2c/i2c-core-of.c 		dev_err(dev, "of_i2c: invalid reg on %pOF\n", node);
node               52 drivers/i2c/i2c-core-of.c 	info->of_node = node;
node               54 drivers/i2c/i2c-core-of.c 	if (of_property_read_bool(node, "host-notify"))
node               57 drivers/i2c/i2c-core-of.c 	if (of_get_property(node, "wakeup-source", NULL))
node               65 drivers/i2c/i2c-core-of.c 						 struct device_node *node)
node               71 drivers/i2c/i2c-core-of.c 	dev_dbg(&adap->dev, "of_i2c: register %pOF\n", node);
node               73 drivers/i2c/i2c-core-of.c 	ret = of_i2c_get_board_info(&adap->dev, node, &info);
node               79 drivers/i2c/i2c-core-of.c 		dev_err(&adap->dev, "of_i2c: Failure registering %pOF\n", node);
node               87 drivers/i2c/i2c-core-of.c 	struct device_node *bus, *node;
node              100 drivers/i2c/i2c-core-of.c 	for_each_available_child_of_node(bus, node) {
node              101 drivers/i2c/i2c-core-of.c 		if (of_node_test_and_set_flag(node, OF_POPULATED))
node              104 drivers/i2c/i2c-core-of.c 		client = of_i2c_register_device(adap, node);
node              108 drivers/i2c/i2c-core-of.c 				 node);
node              109 drivers/i2c/i2c-core-of.c 			of_node_clear_flag(node, OF_POPULATED);
node              128 drivers/i2c/i2c-core-of.c struct i2c_client *of_find_i2c_device_by_node(struct device_node *node)
node              133 drivers/i2c/i2c-core-of.c 	dev = bus_find_device_by_of_node(&i2c_bus_type, node);
node              146 drivers/i2c/i2c-core-of.c struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node)
node              151 drivers/i2c/i2c-core-of.c 	dev = bus_find_device(&i2c_bus_type, NULL, node,
node              165 drivers/i2c/i2c-core-of.c struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node)
node              169 drivers/i2c/i2c-core-of.c 	adapter = of_find_i2c_adapter_by_node(node);
node               64 drivers/i2c/i2c-stub.c 	struct list_head node;
node               96 drivers/i2c/i2c-stub.c 	list_for_each_entry(b, &chip->smbus_blocks, node) {
node              107 drivers/i2c/i2c-stub.c 		list_add(&rb->node, &chip->smbus_blocks);
node               41 drivers/i2c/muxes/i2c-mux-pinctrl.c 	list_for_each_entry(setting, &state->settings, node) {
node             1342 drivers/i3c/master.c 	list_add_tail(&dev->common.node, &master->bus.devs.i3c);
node             1384 drivers/i3c/master.c 	list_del(&dev->common.node);
node             1398 drivers/i3c/master.c 	list_add_tail(&dev->common.node, &master->bus.devs.i2c);
node             1407 drivers/i3c/master.c 	list_del(&dev->common.node);
node             1578 drivers/i3c/master.c 				 common.node) {
node             1590 drivers/i3c/master.c 				 common.node) {
node             1642 drivers/i3c/master.c 	list_for_each_entry(i2cboardinfo, &master->boardinfo.i2c, node) {
node             1666 drivers/i3c/master.c 	list_for_each_entry(i3cboardinfo, &master->boardinfo.i3c, node) {
node             1936 drivers/i3c/master.c 				struct device_node *node, u32 *reg)
node             1946 drivers/i3c/master.c 	ret = of_i2c_get_board_info(dev, node, &boardinfo->base);
node             1963 drivers/i3c/master.c 	list_add_tail(&boardinfo->node, &master->boardinfo.i2c);
node             1964 drivers/i3c/master.c 	of_node_get(node);
node             1971 drivers/i3c/master.c 				struct device_node *node, u32 *reg)
node             1994 drivers/i3c/master.c 	if (!of_property_read_u32(node, "assigned-address", &init_dyn_addr)) {
node             2011 drivers/i3c/master.c 	boardinfo->of_node = of_node_get(node);
node             2012 drivers/i3c/master.c 	list_add_tail(&boardinfo->node, &master->boardinfo.i3c);
node             2018 drivers/i3c/master.c 				 struct device_node *node)
node             2023 drivers/i3c/master.c 	if (!master || !node)
node             2026 drivers/i3c/master.c 	ret = of_property_read_u32_array(node, "reg", reg, ARRAY_SIZE(reg));
node             2035 drivers/i3c/master.c 		ret = of_i3c_master_add_i2c_boardinfo(master, node, reg);
node             2037 drivers/i3c/master.c 		ret = of_i3c_master_add_i3c_boardinfo(master, node, reg);
node             2046 drivers/i3c/master.c 	struct device_node *node;
node             2053 drivers/i3c/master.c 	for_each_available_child_of_node(i3cbus_np, node) {
node             2054 drivers/i3c/master.c 		ret = of_i3c_master_add_dev(master, node);
node             2056 drivers/i3c/master.c 			of_node_put(node);
node             2215 drivers/i3c/master.c 	struct list_head node;
node             2241 drivers/i3c/master.c 					struct i3c_generic_ibi_slot, node);
node             2242 drivers/i3c/master.c 		list_del(&slot->node);
node             2307 drivers/i3c/master.c 		list_add_tail(&slot->node, &pool->free_slots);
node             2337 drivers/i3c/master.c 					struct i3c_generic_ibi_slot, node);
node             2339 drivers/i3c/master.c 		list_del(&slot->node);
node             2365 drivers/i3c/master.c 	list_add_tail(&slot->node, &pool->free_slots);
node             2445 drivers/i3c/master.c 	list_for_each_entry(i2cbi, &master->boardinfo.i2c, node) {
node              220 drivers/i3c/master/dw-i3c-master.c 	struct list_head node;
node              366 drivers/i3c/master/dw-i3c-master.c 	INIT_LIST_HEAD(&xfer->node);
node              414 drivers/i3c/master/dw-i3c-master.c 		list_add_tail(&xfer->node, &master->xferqueue.list);
node              437 drivers/i3c/master/dw-i3c-master.c 		list_del_init(&xfer->node);
node              510 drivers/i3c/master/dw-i3c-master.c 					node);
node              512 drivers/i3c/master/dw-i3c-master.c 		list_del_init(&xfer->node);
node              384 drivers/i3c/master/i3c-master-cdns.c 	struct list_head node;
node              508 drivers/i3c/master/i3c-master-cdns.c 	INIT_LIST_HEAD(&xfer->node);
node              615 drivers/i3c/master/i3c-master-cdns.c 					struct cdns_i3c_xfer, node);
node              617 drivers/i3c/master/i3c-master-cdns.c 		list_del_init(&xfer->node);
node              631 drivers/i3c/master/i3c-master-cdns.c 		list_add_tail(&xfer->node, &master->xferqueue.list);
node              661 drivers/i3c/master/i3c-master-cdns.c 		list_del_init(&xfer->node);
node             1314 drivers/ide/ide-probe.c static int ide_port_alloc_devices(ide_hwif_t *hwif, int node)
node             1320 drivers/ide/ide-probe.c 		drive = kzalloc_node(sizeof(*drive), GFP_KERNEL, node);
node             1332 drivers/ide/ide-probe.c 		drive->id = kzalloc_node(SECTOR_SIZE, GFP_KERNEL, node);
node             1352 drivers/ide/ide-probe.c 	int node = dev ? dev_to_node(dev) : -1;
node             1355 drivers/ide/ide-probe.c 	host = kzalloc_node(sizeof(*host), GFP_KERNEL, node);
node             1366 drivers/ide/ide-probe.c 		hwif = kzalloc_node(sizeof(*hwif), GFP_KERNEL, node);
node             1370 drivers/ide/ide-probe.c 		if (ide_port_alloc_devices(hwif, node) < 0) {
node               41 drivers/ide/ide-scan-pci.c 	list_add_tail(&driver->node, &ide_pci_drivers);
node               62 drivers/ide/ide-scan-pci.c 		d = list_entry(l, struct pci_driver, node);
node              106 drivers/ide/ide-scan-pci.c 		d = list_entry(l, struct pci_driver, node);
node               57 drivers/ide/pmac.c 	struct device_node*		node;
node              874 drivers/ide/pmac.c 	ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, pmif->node, pmif->aapl_bus_id,
node              887 drivers/ide/pmac.c 		ppc_md.feature_call(PMAC_FTR_IDE_RESET, pmif->node, pmif->aapl_bus_id, 1);
node              888 drivers/ide/pmac.c 		ppc_md.feature_call(PMAC_FTR_IDE_ENABLE, pmif->node, pmif->aapl_bus_id, 1);
node              890 drivers/ide/pmac.c 		ppc_md.feature_call(PMAC_FTR_IDE_RESET, pmif->node, pmif->aapl_bus_id, 0);
node              913 drivers/ide/pmac.c 	struct device_node *np = pmif->node;
node             1019 drivers/ide/pmac.c 	struct device_node *np = pmif->node;
node             1174 drivers/ide/pmac.c 	pmif->node = mdev->ofdev.dev.of_node;
node             1287 drivers/ide/pmac.c 	pmif->node = np;
node              834 drivers/iio/adc/at91_adc.c static int at91_adc_probe_dt_ts(struct device_node *node,
node              840 drivers/iio/adc/at91_adc.c 	ret = of_property_read_u32(node, "atmel,adc-ts-wires", &prop);
node              859 drivers/iio/adc/at91_adc.c 	of_property_read_u32(node, "atmel,adc-ts-pressure-threshold", &prop);
node              873 drivers/iio/adc/at91_adc.c 	struct device_node *node = pdev->dev.of_node;
node              878 drivers/iio/adc/at91_adc.c 	if (!node)
node              884 drivers/iio/adc/at91_adc.c 	st->use_external = of_property_read_bool(node, "atmel,adc-use-external-triggers");
node              886 drivers/iio/adc/at91_adc.c 	if (of_property_read_u32(node, "atmel,adc-channels-used", &prop)) {
node              893 drivers/iio/adc/at91_adc.c 	st->sleep_mode = of_property_read_bool(node, "atmel,adc-sleep-mode");
node              895 drivers/iio/adc/at91_adc.c 	if (of_property_read_u32(node, "atmel,adc-startup-time", &prop)) {
node              903 drivers/iio/adc/at91_adc.c 	of_property_read_u32(node, "atmel,adc-sample-hold-time", &prop);
node              906 drivers/iio/adc/at91_adc.c 	if (of_property_read_u32(node, "atmel,adc-vref", &prop)) {
node              919 drivers/iio/adc/at91_adc.c 	st->trigger_number = of_get_child_count(node);
node              930 drivers/iio/adc/at91_adc.c 	for_each_child_of_node(node, trig_node) {
node              953 drivers/iio/adc/at91_adc.c 		return at91_adc_probe_dt_ts(node, st, &idev->dev);
node              310 drivers/iio/adc/cc10001_adc.c 	struct device_node *node = pdev->dev.of_node;
node              325 drivers/iio/adc/cc10001_adc.c 	if (!of_property_read_u32(node, "adc-reserved-channels", &ret)) {
node              509 drivers/iio/adc/qcom-spmi-adc5.c 				    struct device_node *node,
node              512 drivers/iio/adc/qcom-spmi-adc5.c 	const char *name = node->name, *channel_name;
node              517 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32(node, "reg", &chan);
node              532 drivers/iio/adc/qcom-spmi-adc5.c 	channel_name = of_get_property(node,
node              533 drivers/iio/adc/qcom-spmi-adc5.c 				"label", NULL) ? : node->name;
node              540 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32(node, "qcom,decimation", &value);
node              553 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32_array(node, "qcom,pre-scaling", varr, 2);
node              567 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32(node, "qcom,hw-settle-time", &value);
node              599 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32(node, "qcom,avg-samples", &value);
node              612 drivers/iio/adc/qcom-spmi-adc5.c 	if (of_property_read_bool(node, "qcom,ratiometric"))
node              669 drivers/iio/adc/qcom-spmi-adc5.c static int adc5_get_dt_data(struct adc5_chip *adc, struct device_node *node)
node              680 drivers/iio/adc/qcom-spmi-adc5.c 	adc->nchannels = of_get_available_child_count(node);
node              696 drivers/iio/adc/qcom-spmi-adc5.c 	id = of_match_node(adc5_match_table, node);
node              703 drivers/iio/adc/qcom-spmi-adc5.c 	for_each_available_child_of_node(node, child) {
node              731 drivers/iio/adc/qcom-spmi-adc5.c 	struct device_node *node = pdev->dev.of_node;
node              743 drivers/iio/adc/qcom-spmi-adc5.c 	ret = of_property_read_u32(node, "reg", &reg);
node              758 drivers/iio/adc/qcom-spmi-adc5.c 	ret = adc5_get_dt_data(adc, node);
node              777 drivers/iio/adc/qcom-spmi-adc5.c 	indio_dev->dev.of_node = node;
node              430 drivers/iio/adc/qcom-spmi-iadc.c static int iadc_rsense_read(struct iadc_chip *iadc, struct device_node *node)
node              435 drivers/iio/adc/qcom-spmi-iadc.c 	ret = of_property_read_u32(node, "qcom,external-resistor-micro-ohms",
node              487 drivers/iio/adc/qcom-spmi-iadc.c 	struct device_node *node = pdev->dev.of_node;
node              508 drivers/iio/adc/qcom-spmi-iadc.c 	ret = of_property_read_u32(node, "reg", &res);
node              518 drivers/iio/adc/qcom-spmi-iadc.c 	ret = iadc_rsense_read(iadc, node);
node              557 drivers/iio/adc/qcom-spmi-iadc.c 	indio_dev->dev.of_node = node;
node              653 drivers/iio/adc/qcom-spmi-vadc.c 				    struct device_node *node)
node              655 drivers/iio/adc/qcom-spmi-vadc.c 	const char *name = node->name;
node              659 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32(node, "reg", &chan);
node              673 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32(node, "qcom,decimation", &value);
node              686 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32_array(node, "qcom,pre-scaling", varr, 2);
node              699 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32(node, "qcom,hw-settle-time", &value);
node              712 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32(node, "qcom,avg-samples", &value);
node              725 drivers/iio/adc/qcom-spmi-vadc.c 	if (of_property_read_bool(node, "qcom,ratiometric"))
node              735 drivers/iio/adc/qcom-spmi-vadc.c static int vadc_get_dt_data(struct vadc_priv *vadc, struct device_node *node)
node              744 drivers/iio/adc/qcom-spmi-vadc.c 	vadc->nchannels = of_get_available_child_count(node);
node              760 drivers/iio/adc/qcom-spmi-vadc.c 	for_each_available_child_of_node(node, child) {
node              852 drivers/iio/adc/qcom-spmi-vadc.c 	struct device_node *node = pdev->dev.of_node;
node              864 drivers/iio/adc/qcom-spmi-vadc.c 	ret = of_property_read_u32(node, "reg", &reg);
node              884 drivers/iio/adc/qcom-spmi-vadc.c 	ret = vadc_get_dt_data(vadc, node);
node              911 drivers/iio/adc/qcom-spmi-vadc.c 	indio_dev->dev.of_node = node;
node             1579 drivers/iio/adc/stm32-adc.c 	struct device_node *node = indio_dev->dev.of_node;
node             1584 drivers/iio/adc/stm32-adc.c 	if (of_property_read_u32(node, "assigned-resolution-bits", &res))
node             1658 drivers/iio/adc/stm32-adc.c 	struct device_node *node = indio_dev->dev.of_node;
node             1668 drivers/iio/adc/stm32-adc.c 	ret = of_property_count_u32_elems(node, "st,adc-channels");
node             1676 drivers/iio/adc/stm32-adc.c 	ret = of_property_count_elems_of_size(node, "st,adc-diff-channels",
node             1686 drivers/iio/adc/stm32-adc.c 		ret = of_property_read_u32_array(node, "st,adc-diff-channels",
node             1698 drivers/iio/adc/stm32-adc.c 	ret = of_property_count_u32_elems(node, "st,min-sample-time-nsecs");
node             1709 drivers/iio/adc/stm32-adc.c 	of_property_for_each_u32(node, "st,adc-channels", prop, cur, val) {
node             1748 drivers/iio/adc/stm32-adc.c 		of_property_read_u32_index(node, "st,min-sample-time-nsecs",
node              220 drivers/iio/adc/stm32-dfsdm-core.c 	struct device_node *node = pdev->dev.of_node;
node              226 drivers/iio/adc/stm32-dfsdm-core.c 	if (!node)
node              845 drivers/iio/adc/ti-ads1015.c 	struct device_node *node;
node              851 drivers/iio/adc/ti-ads1015.c 	for_each_child_of_node(client->dev.of_node, node) {
node              857 drivers/iio/adc/ti-ads1015.c 		if (of_property_read_u32(node, "reg", &pval)) {
node              859 drivers/iio/adc/ti-ads1015.c 				node);
node              867 drivers/iio/adc/ti-ads1015.c 				channel, node);
node              871 drivers/iio/adc/ti-ads1015.c 		if (!of_property_read_u32(node, "ti,gain", &pval)) {
node              875 drivers/iio/adc/ti-ads1015.c 					node);
node              876 drivers/iio/adc/ti-ads1015.c 				of_node_put(node);
node              881 drivers/iio/adc/ti-ads1015.c 		if (!of_property_read_u32(node, "ti,datarate", &pval)) {
node              886 drivers/iio/adc/ti-ads1015.c 					node);
node              887 drivers/iio/adc/ti-ads1015.c 				of_node_put(node);
node              579 drivers/iio/adc/ti_am335x_adc.c 	struct device_node *node = pdev->dev.of_node;
node              585 drivers/iio/adc/ti_am335x_adc.c 	of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) {
node              596 drivers/iio/adc/ti_am335x_adc.c 	of_property_read_u32_array(node, "ti,chan-step-avg",
node              598 drivers/iio/adc/ti_am335x_adc.c 	of_property_read_u32_array(node, "ti,chan-step-opendelay",
node              600 drivers/iio/adc/ti_am335x_adc.c 	of_property_read_u32_array(node, "ti,chan-step-sampledelay",
node              611 drivers/iio/adc/ti_am335x_adc.c 	struct device_node	*node = pdev->dev.of_node;
node              614 drivers/iio/adc/ti_am335x_adc.c 	if (!node) {
node              446 drivers/iio/common/ssp_sensors/ssp_dev.c 	struct device_node *node = dev->of_node;
node              453 drivers/iio/common/ssp_sensors/ssp_dev.c 	data->mcu_ap_gpio = of_get_named_gpio(node, "mcu-ap-gpios", 0);
node              457 drivers/iio/common/ssp_sensors/ssp_dev.c 	data->ap_mcu_gpio = of_get_named_gpio(node, "ap-mcu-gpios", 0);
node              461 drivers/iio/common/ssp_sensors/ssp_dev.c 	data->mcu_reset_gpio = of_get_named_gpio(node, "mcu-reset-gpios", 0);
node              475 drivers/iio/common/ssp_sensors/ssp_dev.c 	match = of_match_node(ssp_of_match, node);
node              697 drivers/infiniband/core/cm.c 	struct rb_node *node = cm.listen_service_table.rb_node;
node              700 drivers/infiniband/core/cm.c 	while (node) {
node              701 drivers/infiniband/core/cm.c 		cm_id_priv = rb_entry(node, struct cm_id_private, service_node);
node              708 drivers/infiniband/core/cm.c 			node = node->rb_left;
node              710 drivers/infiniband/core/cm.c 			node = node->rb_right;
node              712 drivers/infiniband/core/cm.c 			node = node->rb_left;
node              714 drivers/infiniband/core/cm.c 			node = node->rb_right;
node              716 drivers/infiniband/core/cm.c 			node = node->rb_right;
node              754 drivers/infiniband/core/cm.c 	struct rb_node *node = cm.remote_id_table.rb_node;
node              757 drivers/infiniband/core/cm.c 	while (node) {
node              758 drivers/infiniband/core/cm.c 		timewait_info = rb_entry(node, struct cm_timewait_info,
node              761 drivers/infiniband/core/cm.c 			node = node->rb_left;
node              763 drivers/infiniband/core/cm.c 			node = node->rb_right;
node              765 drivers/infiniband/core/cm.c 			node = node->rb_left;
node              767 drivers/infiniband/core/cm.c 			node = node->rb_right;
node             1637 drivers/infiniband/core/cma.c 	hlist_for_each_entry(id_priv, &bind_list->owners, node) {
node             1794 drivers/infiniband/core/cma.c 	hlist_del(&id_priv->node);
node             3267 drivers/infiniband/core/cma.c 	hlist_add_head(&id_priv->node, &bind_list->owners);
node             3302 drivers/infiniband/core/cma.c 	hlist_for_each_entry(cur_id, &bind_list->owners, node) {
node             3389 drivers/infiniband/core/cma.c 	hlist_for_each_entry(cur_id, &bind_list->owners, node) {
node               57 drivers/infiniband/core/cma_priv.h 	struct hlist_node	node;
node               98 drivers/infiniband/core/multicast.c 	struct rb_node		node;
node              132 drivers/infiniband/core/multicast.c 	struct rb_node *node = port->table.rb_node;
node              136 drivers/infiniband/core/multicast.c 	while (node) {
node              137 drivers/infiniband/core/multicast.c 		group = rb_entry(node, struct mcast_group, node);
node              143 drivers/infiniband/core/multicast.c 			node = node->rb_left;
node              145 drivers/infiniband/core/multicast.c 			node = node->rb_right;
node              161 drivers/infiniband/core/multicast.c 		cur_group = rb_entry(parent, struct mcast_group, node);
node              174 drivers/infiniband/core/multicast.c 	rb_link_node(&group->node, parent, link);
node              175 drivers/infiniband/core/multicast.c 	rb_insert_color(&group->node, &port->table);
node              192 drivers/infiniband/core/multicast.c 		rb_erase(&group->node, &port->table);
node              535 drivers/infiniband/core/multicast.c 			rb_erase(&group->node, &group->port->table);
node              774 drivers/infiniband/core/multicast.c 	struct rb_node *node;
node              778 drivers/infiniband/core/multicast.c 	for (node = rb_first(&port->table); node; node = rb_next(node)) {
node              779 drivers/infiniband/core/multicast.c 		group = rb_entry(node, struct mcast_group, node);
node               82 drivers/infiniband/core/umem_odp.c 	struct rb_node *node;
node               88 drivers/infiniband/core/umem_odp.c 	for (node = rb_first_cached(&per_mm->umem_tree); node;
node               89 drivers/infiniband/core/umem_odp.c 	     node = rb_next(node)) {
node               91 drivers/infiniband/core/umem_odp.c 			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
node              776 drivers/infiniband/core/umem_odp.c 	struct interval_tree_node *node, *next;
node              782 drivers/infiniband/core/umem_odp.c 	for (node = interval_tree_iter_first(root, start, last - 1);
node              783 drivers/infiniband/core/umem_odp.c 			node; node = next) {
node              787 drivers/infiniband/core/umem_odp.c 		next = interval_tree_iter_next(node, start, last - 1);
node              788 drivers/infiniband/core/umem_odp.c 		umem = container_of(node, struct ib_umem_odp, interval_tree);
node              472 drivers/infiniband/core/uverbs_cmd.c 	struct rb_node  node;
node              494 drivers/infiniband/core/uverbs_cmd.c 		scan = rb_entry(parent, struct xrcd_table_entry, node);
node              506 drivers/infiniband/core/uverbs_cmd.c 	rb_link_node(&entry->node, parent, p);
node              507 drivers/infiniband/core/uverbs_cmd.c 	rb_insert_color(&entry->node, &dev->xrcd_tree);
node              519 drivers/infiniband/core/uverbs_cmd.c 		entry = rb_entry(p, struct xrcd_table_entry, node);
node              551 drivers/infiniband/core/uverbs_cmd.c 		rb_erase(&entry->node, &dev->xrcd_tree);
node              169 drivers/infiniband/hw/hfi1/affinity.c 	int node;
node              200 drivers/infiniband/hw/hfi1/affinity.c 			node = pcibus_to_node(dev->bus);
node              201 drivers/infiniband/hw/hfi1/affinity.c 			if (node < 0)
node              204 drivers/infiniband/hw/hfi1/affinity.c 			hfi1_per_node_cntr[node]++;
node              218 drivers/infiniband/hw/hfi1/affinity.c 	for (node = 0; node < node_affinity.num_possible_nodes; node++)
node              219 drivers/infiniband/hw/hfi1/affinity.c 		hfi1_per_node_cntr[node] = 1;
node              246 drivers/infiniband/hw/hfi1/affinity.c static struct hfi1_affinity_node *node_affinity_allocate(int node)
node              253 drivers/infiniband/hw/hfi1/affinity.c 	entry->node = node;
node              270 drivers/infiniband/hw/hfi1/affinity.c static struct hfi1_affinity_node *node_affinity_lookup(int node)
node              277 drivers/infiniband/hw/hfi1/affinity.c 		if (entry->node == node)
node              500 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node              557 drivers/infiniband/hw/hfi1/affinity.c 				       hfi1_per_node_cntr[dd->node];
node              566 drivers/infiniband/hw/hfi1/affinity.c 		    hfi1_per_node_cntr[dd->node] != 0)
node              634 drivers/infiniband/hw/hfi1/affinity.c 	int node = pcibus_to_node(dd->pcidev->bus);
node              644 drivers/infiniband/hw/hfi1/affinity.c 	if (node < 0) {
node              646 drivers/infiniband/hw/hfi1/affinity.c 		node = 0;
node              648 drivers/infiniband/hw/hfi1/affinity.c 	dd->node = node;
node              650 drivers/infiniband/hw/hfi1/affinity.c 	local_mask = cpumask_of_node(dd->node);
node              655 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node              662 drivers/infiniband/hw/hfi1/affinity.c 		entry = node_affinity_allocate(node);
node              704 drivers/infiniband/hw/hfi1/affinity.c 				  hfi1_per_node_cntr[dd->node];
node              768 drivers/infiniband/hw/hfi1/affinity.c 	if (dd->node < 0)
node              772 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node              783 drivers/infiniband/hw/hfi1/affinity.c 	dd->node = NUMA_NO_NODE;
node              803 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node              899 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node              974 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(dd->node);
node             1037 drivers/infiniband/hw/hfi1/affinity.c int hfi1_get_proc_affinity(int node)
node             1114 drivers/infiniband/hw/hfi1/affinity.c 	entry = node_affinity_lookup(node);
node             1153 drivers/infiniband/hw/hfi1/affinity.c 	node_mask = cpumask_of_node(node);
node             1154 drivers/infiniband/hw/hfi1/affinity.c 	hfi1_cdbg(PROC, "Device on NUMA %u, CPUs %*pbl", node,
node             1160 drivers/infiniband/hw/hfi1/affinity.c 	hfi1_cdbg(PROC, "Available CPUs on NUMA %u: %*pbl", node,
node               95 drivers/infiniband/hw/hfi1/affinity.h int hfi1_get_proc_affinity(int node);
node              100 drivers/infiniband/hw/hfi1/affinity.h 	int node;
node              968 drivers/infiniband/hw/hfi1/file_ops.c 	fd->rec_cpu_num = hfi1_get_proc_affinity(dd->node);
node              985 drivers/infiniband/hw/hfi1/file_ops.c 	uctxt->sc = sc_alloc(dd, SC_USER, uctxt->rcvhdrqentsize, dd->node);
node             1213 drivers/infiniband/hw/hfi1/hfi.h 	int node; /* home node of this chip */
node              138 drivers/infiniband/hw/hfi1/init.c 	ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd);
node              159 drivers/infiniband/hw/hfi1/init.c 	rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node);
node              178 drivers/infiniband/hw/hfi1/init.c 			       GFP_KERNEL, dd->node);
node             1293 drivers/infiniband/hw/hfi1/init.c 	dd->node = NUMA_NO_NODE;
node              127 drivers/infiniband/hw/hfi1/mad.c 	struct trap_node *node, *q;
node              142 drivers/infiniband/hw/hfi1/mad.c 		list_for_each_entry_safe(node, q, &trap_list, list) {
node              143 drivers/infiniband/hw/hfi1/mad.c 			list_del(&node->list);
node              144 drivers/infiniband/hw/hfi1/mad.c 			if (node != trap)
node              145 drivers/infiniband/hw/hfi1/mad.c 				kfree(node);
node              159 drivers/infiniband/hw/hfi1/mad.c 	struct trap_node *node;
node              183 drivers/infiniband/hw/hfi1/mad.c 	list_for_each_entry(node, &trap_list->list, list) {
node              184 drivers/infiniband/hw/hfi1/mad.c 		if (node == trap) {
node              185 drivers/infiniband/hw/hfi1/mad.c 			node->retry++;
node              207 drivers/infiniband/hw/hfi1/mad.c 	node = NULL;
node              219 drivers/infiniband/hw/hfi1/mad.c 		node = list_first_entry(&trap_list->list, struct trap_node,
node              221 drivers/infiniband/hw/hfi1/mad.c 		node->in_use = 1;
node              225 drivers/infiniband/hw/hfi1/mad.c 	return node;
node               82 drivers/infiniband/hw/hfi1/mmu_rb.c INTERVAL_TREE_DEFINE(struct mmu_rb_node, node, unsigned long, __last,
node               85 drivers/infiniband/hw/hfi1/mmu_rb.c static unsigned long mmu_node_start(struct mmu_rb_node *node)
node               87 drivers/infiniband/hw/hfi1/mmu_rb.c 	return node->addr & PAGE_MASK;
node               90 drivers/infiniband/hw/hfi1/mmu_rb.c static unsigned long mmu_node_last(struct mmu_rb_node *node)
node               92 drivers/infiniband/hw/hfi1/mmu_rb.c 	return PAGE_ALIGN(node->addr + node->len) - 1;
node              132 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct rb_node *node;
node              148 drivers/infiniband/hw/hfi1/mmu_rb.c 	while ((node = rb_first_cached(&handler->root))) {
node              149 drivers/infiniband/hw/hfi1/mmu_rb.c 		rbnode = rb_entry(node, struct mmu_rb_node, node);
node              150 drivers/infiniband/hw/hfi1/mmu_rb.c 		rb_erase_cached(node, &handler->root);
node              164 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct mmu_rb_node *node;
node              170 drivers/infiniband/hw/hfi1/mmu_rb.c 	node = __mmu_rb_search(handler, mnode->addr, mnode->len);
node              171 drivers/infiniband/hw/hfi1/mmu_rb.c 	if (node) {
node              193 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct mmu_rb_node *node = NULL;
node              197 drivers/infiniband/hw/hfi1/mmu_rb.c 		node = __mmu_int_rb_iter_first(&handler->root, addr,
node              200 drivers/infiniband/hw/hfi1/mmu_rb.c 		for (node = __mmu_int_rb_iter_first(&handler->root, addr,
node              202 drivers/infiniband/hw/hfi1/mmu_rb.c 		     node;
node              203 drivers/infiniband/hw/hfi1/mmu_rb.c 		     node = __mmu_int_rb_iter_next(node, addr,
node              205 drivers/infiniband/hw/hfi1/mmu_rb.c 			if (handler->ops->filter(node, addr, len))
node              206 drivers/infiniband/hw/hfi1/mmu_rb.c 				return node;
node              209 drivers/infiniband/hw/hfi1/mmu_rb.c 	return node;
node              216 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct mmu_rb_node *node;
node              221 drivers/infiniband/hw/hfi1/mmu_rb.c 	node = __mmu_rb_search(handler, addr, len);
node              222 drivers/infiniband/hw/hfi1/mmu_rb.c 	if (node) {
node              223 drivers/infiniband/hw/hfi1/mmu_rb.c 		if (node->addr == addr && node->len == len)
node              225 drivers/infiniband/hw/hfi1/mmu_rb.c 		__mmu_int_rb_remove(node, &handler->root);
node              226 drivers/infiniband/hw/hfi1/mmu_rb.c 		list_del(&node->list); /* remove from LRU list */
node              231 drivers/infiniband/hw/hfi1/mmu_rb.c 	*rb_node = node;
node              271 drivers/infiniband/hw/hfi1/mmu_rb.c 			struct mmu_rb_node *node)
node              276 drivers/infiniband/hw/hfi1/mmu_rb.c 	trace_hfi1_mmu_rb_remove(node->addr, node->len);
node              278 drivers/infiniband/hw/hfi1/mmu_rb.c 	__mmu_int_rb_remove(node, &handler->root);
node              279 drivers/infiniband/hw/hfi1/mmu_rb.c 	list_del(&node->list); /* remove from LRU list */
node              282 drivers/infiniband/hw/hfi1/mmu_rb.c 	handler->ops->remove(handler->ops_arg, node);
node              291 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct mmu_rb_node *node, *ptr = NULL;
node              296 drivers/infiniband/hw/hfi1/mmu_rb.c 	for (node = __mmu_int_rb_iter_first(root, range->start, range->end-1);
node              297 drivers/infiniband/hw/hfi1/mmu_rb.c 	     node; node = ptr) {
node              299 drivers/infiniband/hw/hfi1/mmu_rb.c 		ptr = __mmu_int_rb_iter_next(node, range->start,
node              301 drivers/infiniband/hw/hfi1/mmu_rb.c 		trace_hfi1_mmu_mem_invalidate(node->addr, node->len);
node              302 drivers/infiniband/hw/hfi1/mmu_rb.c 		if (handler->ops->invalidate(handler->ops_arg, node)) {
node              303 drivers/infiniband/hw/hfi1/mmu_rb.c 			__mmu_int_rb_remove(node, root);
node              305 drivers/infiniband/hw/hfi1/mmu_rb.c 			list_move(&node->list, &handler->del_list);
node              325 drivers/infiniband/hw/hfi1/mmu_rb.c 	struct mmu_rb_node *node;
node              328 drivers/infiniband/hw/hfi1/mmu_rb.c 		node = list_first_entry(del_list, struct mmu_rb_node, list);
node              329 drivers/infiniband/hw/hfi1/mmu_rb.c 		list_del(&node->list);
node              330 drivers/infiniband/hw/hfi1/mmu_rb.c 		handler->ops->remove(handler->ops_arg, node);
node               56 drivers/infiniband/hw/hfi1/mmu_rb.h 	struct rb_node node;
node               65 drivers/infiniband/hw/hfi1/mmu_rb.h 	bool (*filter)(struct mmu_rb_node *node, unsigned long addr,
node               69 drivers/infiniband/hw/hfi1/mmu_rb.h 	int (*invalidate)(void *ops_arg, struct mmu_rb_node *node);
node              566 drivers/infiniband/hw/hfi1/pio.c 	sc->hw_free = &sc->dd->cr_base[sc->node].va[gc].cr[index];
node              568 drivers/infiniband/hw/hfi1/pio.c 	       &((struct credit_return *)sc->dd->cr_base[sc->node].dma)[gc];
node              740 drivers/infiniband/hw/hfi1/pio.c 	sc->node = numa;
node             2023 drivers/infiniband/hw/hfi1/pio.c 				  dd->rcd[0]->rcvhdrqentsize, dd->node);
node             2032 drivers/infiniband/hw/hfi1/pio.c 					       GFP_KERNEL, dd->node);
node             2047 drivers/infiniband/hw/hfi1/pio.c 					 dd->rcd[0]->rcvhdrqentsize, dd->node);
node             2057 drivers/infiniband/hw/hfi1/pio.c 		sc_alloc(dd, SC_KERNEL, dd->rcd[0]->rcvhdrqentsize, dd->node);
node             2127 drivers/infiniband/hw/hfi1/pio.c 			set_dev_node(&dd->pcidev->dev, dd->node);
node             2135 drivers/infiniband/hw/hfi1/pio.c 	set_dev_node(&dd->pcidev->dev, dd->node);
node              106 drivers/infiniband/hw/hfi1/pio.h 	int node;			/* context home node */
node              389 drivers/infiniband/hw/hfi1/qp.c 			       cpumask_first(cpumask_of_node(dd->node)));
node              724 drivers/infiniband/hw/hfi1/qp.c 	priv = kzalloc_node(sizeof(*priv), GFP_KERNEL, rdi->dparms.node);
node              731 drivers/infiniband/hw/hfi1/qp.c 				   rdi->dparms.node);
node              589 drivers/infiniband/hw/hfi1/ruc.c 			cpumask_first(cpumask_of_node(ps.ppd->dd->node));
node              842 drivers/infiniband/hw/hfi1/sdma.c 	struct rhash_head node;
node              849 drivers/infiniband/hw/hfi1/sdma.c 	.head_offset = offsetof(struct sdma_rht_node, node),
node              998 drivers/infiniband/hw/hfi1/sdma.c 						     &rht_node->node,
node             1065 drivers/infiniband/hw/hfi1/sdma.c 							     &rht_node->node,
node             1400 drivers/infiniband/hw/hfi1/sdma.c 				    GFP_KERNEL, dd->node);
node             1478 drivers/infiniband/hw/hfi1/sdma.c 				      GFP_KERNEL, dd->node);
node              372 drivers/infiniband/hw/hfi1/tid_rdma.c 					    GFP_KERNEL, dd->node);
node              380 drivers/infiniband/hw/hfi1/tid_rdma.c 					    dd->node);
node              392 drivers/infiniband/hw/hfi1/tid_rdma.c 					    dd->node);
node              624 drivers/infiniband/hw/hfi1/tid_rdma.c 			     cpumask_first(cpumask_of_node(dd->node)),
node             1201 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct kern_tid_node *node = &flow->tnode[flow->tnode_cnt++];
node             1210 drivers/infiniband/hw/hfi1/tid_rdma.c 	node->grp = grp;
node             1211 drivers/infiniband/hw/hfi1/tid_rdma.c 	node->map = grp->map;
node             1212 drivers/infiniband/hw/hfi1/tid_rdma.c 	node->cnt = cnt;
node             1298 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct kern_tid_node *node = &flow->tnode[grp_num];
node             1299 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_group *grp = node->grp;
node             1308 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (node->map & BIT(i) || cnt >= node->cnt) {
node             1330 drivers/infiniband/hw/hfi1/tid_rdma.c 		pair = !(i & 0x1) && !((node->map >> i) & 0x3) &&
node             1331 drivers/infiniband/hw/hfi1/tid_rdma.c 			node->cnt >= cnt + 2;
node             1365 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct kern_tid_node *node = &flow->tnode[grp_num];
node             1366 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_group *grp = node->grp;
node             1373 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (node->map & BIT(i) || cnt >= node->cnt) {
node             5363 drivers/infiniband/hw/hfi1/tid_rdma.c 		cpumask_first(cpumask_of_node(ps.ppd->dd->node));
node             5414 drivers/infiniband/hw/hfi1/tid_rdma.c 				   cpumask_first(cpumask_of_node(dd->node)));
node               62 drivers/infiniband/hw/hfi1/user_exp_rcv.c static int tid_rb_insert(void *arg, struct mmu_rb_node *node);
node               65 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void tid_rb_remove(void *arg, struct mmu_rb_node *node);
node               74 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void clear_tid_node(struct hfi1_filedata *fd, struct tid_rb_node *node);
node              193 drivers/infiniband/hw/hfi1/user_exp_rcv.c 			    struct tid_rb_node *node,
node              202 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		pci_unmap_single(dd->pcidev, node->dma_addr,
node              203 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				 node->mmu.len, PCI_DMA_FROMDEVICE);
node              204 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		pages = &node->pages[idx];
node              755 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	struct tid_rb_node *node;
node              764 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node = kzalloc(sizeof(*node) + (sizeof(struct page *) * npages),
node              766 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	if (!node)
node              775 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		kfree(node);
node              779 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->mmu.addr = tbuf->vaddr + (pageidx * PAGE_SIZE);
node              780 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->mmu.len = npages * PAGE_SIZE;
node              781 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->phys = page_to_phys(pages[0]);
node              782 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->npages = npages;
node              783 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->rcventry = rcventry;
node              784 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->dma_addr = phys;
node              785 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->grp = grp;
node              786 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->freed = false;
node              787 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	memcpy(node->pages, pages, sizeof(struct page *) * npages);
node              790 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		ret = tid_rb_insert(fd, &node->mmu);
node              792 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		ret = hfi1_mmu_rb_insert(fd->handler, &node->mmu);
node              796 drivers/infiniband/hw/hfi1/user_exp_rcv.c 			  node->rcventry, node->mmu.addr, node->phys, ret);
node              799 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		kfree(node);
node              804 drivers/infiniband/hw/hfi1/user_exp_rcv.c 			       node->mmu.addr, node->phys, phys);
node              813 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	struct tid_rb_node *node;
node              828 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node = fd->entry_to_rb[rcventry];
node              829 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	if (!node || node->rcventry != (uctxt->expected_base + rcventry))
node              833 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		*grp = node->grp;
node              836 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		cacheless_tid_rb_remove(fd, node);
node              838 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		hfi1_mmu_rb_remove(fd->handler, &node->mmu);
node              843 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void clear_tid_node(struct hfi1_filedata *fd, struct tid_rb_node *node)
node              848 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	trace_hfi1_exp_tid_unreg(uctxt->ctxt, fd->subctxt, node->rcventry,
node              849 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				 node->npages, node->mmu.addr, node->phys,
node              850 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				 node->dma_addr);
node              856 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	hfi1_put_tid(dd, node->rcventry, PT_INVALID_FLUSH, 0, 0);
node              858 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	unpin_rcv_pages(fd, NULL, node, 0, node->npages, true);
node              860 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->grp->used--;
node              861 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->grp->map &= ~(1 << (node->rcventry - node->grp->base));
node              863 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	if (node->grp->used == node->grp->size - 1)
node              864 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		tid_group_move(node->grp, &uctxt->tid_full_list,
node              866 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	else if (!node->grp->used)
node              867 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		tid_group_move(node->grp, &uctxt->tid_used_list,
node              869 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	kfree(node);
node              889 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				struct tid_rb_node *node;
node              891 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				node = fd->entry_to_rb[rcventry -
node              893 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				if (!node || node->rcventry != rcventry)
node              896 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				cacheless_tid_rb_remove(fd, node);
node              913 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	struct tid_rb_node *node =
node              916 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	if (node->freed)
node              919 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	trace_hfi1_exp_tid_inval(uctxt->ctxt, fdata->subctxt, node->mmu.addr,
node              920 drivers/infiniband/hw/hfi1/user_exp_rcv.c 				 node->rcventry, node->npages, node->dma_addr);
node              921 drivers/infiniband/hw/hfi1/user_exp_rcv.c 	node->freed = true;
node              926 drivers/infiniband/hw/hfi1/user_exp_rcv.c 			rcventry2tidinfo(node->rcventry - uctxt->expected_base);
node              928 drivers/infiniband/hw/hfi1/user_exp_rcv.c 			EXP_TID_SET(LEN, node->npages);
node              951 drivers/infiniband/hw/hfi1/user_exp_rcv.c static int tid_rb_insert(void *arg, struct mmu_rb_node *node)
node              955 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		container_of(node, struct tid_rb_node, mmu);
node              971 drivers/infiniband/hw/hfi1/user_exp_rcv.c static void tid_rb_remove(void *arg, struct mmu_rb_node *node)
node              975 drivers/infiniband/hw/hfi1/user_exp_rcv.c 		container_of(node, struct tid_rb_node, mmu);
node              108 drivers/infiniband/hw/hfi1/user_sdma.c static bool sdma_rb_filter(struct mmu_rb_node *node, unsigned long addr,
node              969 drivers/infiniband/hw/hfi1/user_sdma.c 			  struct sdma_mmu_node *node,
node              979 drivers/infiniband/hw/hfi1/user_sdma.c 	memcpy(pages, node->pages, node->npages * sizeof(*pages));
node              981 drivers/infiniband/hw/hfi1/user_sdma.c 	npages -= node->npages;
node              991 drivers/infiniband/hw/hfi1/user_sdma.c 					 (node->npages * PAGE_SIZE)), npages, 0,
node              992 drivers/infiniband/hw/hfi1/user_sdma.c 					 pages + node->npages);
node              998 drivers/infiniband/hw/hfi1/user_sdma.c 		unpin_vector_pages(pq->mm, pages, node->npages, pinned);
node             1001 drivers/infiniband/hw/hfi1/user_sdma.c 	kfree(node->pages);
node             1002 drivers/infiniband/hw/hfi1/user_sdma.c 	node->rb.len = iovec->iov.iov_len;
node             1003 drivers/infiniband/hw/hfi1/user_sdma.c 	node->pages = pages;
node             1008 drivers/infiniband/hw/hfi1/user_sdma.c static void unpin_sdma_pages(struct sdma_mmu_node *node)
node             1010 drivers/infiniband/hw/hfi1/user_sdma.c 	if (node->npages) {
node             1011 drivers/infiniband/hw/hfi1/user_sdma.c 		unpin_vector_pages(node->pq->mm, node->pages, 0, node->npages);
node             1012 drivers/infiniband/hw/hfi1/user_sdma.c 		atomic_sub(node->npages, &node->pq->n_locked);
node             1021 drivers/infiniband/hw/hfi1/user_sdma.c 	struct sdma_mmu_node *node = NULL;
node             1032 drivers/infiniband/hw/hfi1/user_sdma.c 		node = container_of(rb_node, struct sdma_mmu_node, rb);
node             1034 drivers/infiniband/hw/hfi1/user_sdma.c 			atomic_inc(&node->refcount);
node             1035 drivers/infiniband/hw/hfi1/user_sdma.c 			iovec->pages = node->pages;
node             1036 drivers/infiniband/hw/hfi1/user_sdma.c 			iovec->npages = node->npages;
node             1037 drivers/infiniband/hw/hfi1/user_sdma.c 			iovec->node = node;
node             1042 drivers/infiniband/hw/hfi1/user_sdma.c 	if (!node) {
node             1043 drivers/infiniband/hw/hfi1/user_sdma.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node             1044 drivers/infiniband/hw/hfi1/user_sdma.c 		if (!node)
node             1047 drivers/infiniband/hw/hfi1/user_sdma.c 		node->rb.addr = (unsigned long)iovec->iov.iov_base;
node             1048 drivers/infiniband/hw/hfi1/user_sdma.c 		node->pq = pq;
node             1049 drivers/infiniband/hw/hfi1/user_sdma.c 		atomic_set(&node->refcount, 0);
node             1054 drivers/infiniband/hw/hfi1/user_sdma.c 	if (node->npages < npages) {
node             1055 drivers/infiniband/hw/hfi1/user_sdma.c 		pinned = pin_sdma_pages(req, iovec, node, npages);
node             1060 drivers/infiniband/hw/hfi1/user_sdma.c 		node->npages += pinned;
node             1061 drivers/infiniband/hw/hfi1/user_sdma.c 		npages = node->npages;
node             1063 drivers/infiniband/hw/hfi1/user_sdma.c 	iovec->pages = node->pages;
node             1065 drivers/infiniband/hw/hfi1/user_sdma.c 	iovec->node = node;
node             1067 drivers/infiniband/hw/hfi1/user_sdma.c 	ret = hfi1_mmu_rb_insert(req->pq->handler, &node->rb);
node             1069 drivers/infiniband/hw/hfi1/user_sdma.c 		iovec->node = NULL;
node             1074 drivers/infiniband/hw/hfi1/user_sdma.c 	unpin_sdma_pages(node);
node             1075 drivers/infiniband/hw/hfi1/user_sdma.c 	kfree(node);
node             1455 drivers/infiniband/hw/hfi1/user_sdma.c 		struct sdma_mmu_node *node = req->iovs[i].node;
node             1457 drivers/infiniband/hw/hfi1/user_sdma.c 		if (!node)
node             1460 drivers/infiniband/hw/hfi1/user_sdma.c 		req->iovs[i].node = NULL;
node             1464 drivers/infiniband/hw/hfi1/user_sdma.c 					   &node->rb);
node             1466 drivers/infiniband/hw/hfi1/user_sdma.c 			atomic_dec(&node->refcount);
node             1486 drivers/infiniband/hw/hfi1/user_sdma.c static bool sdma_rb_filter(struct mmu_rb_node *node, unsigned long addr,
node             1489 drivers/infiniband/hw/hfi1/user_sdma.c 	return (bool)(node->addr == addr);
node             1494 drivers/infiniband/hw/hfi1/user_sdma.c 	struct sdma_mmu_node *node =
node             1497 drivers/infiniband/hw/hfi1/user_sdma.c 	atomic_inc(&node->refcount);
node             1509 drivers/infiniband/hw/hfi1/user_sdma.c 	struct sdma_mmu_node *node =
node             1514 drivers/infiniband/hw/hfi1/user_sdma.c 	if (atomic_read(&node->refcount))
node             1518 drivers/infiniband/hw/hfi1/user_sdma.c 	evict_data->cleared += node->npages;
node             1529 drivers/infiniband/hw/hfi1/user_sdma.c 	struct sdma_mmu_node *node =
node             1532 drivers/infiniband/hw/hfi1/user_sdma.c 	unpin_sdma_pages(node);
node             1533 drivers/infiniband/hw/hfi1/user_sdma.c 	kfree(node);
node             1538 drivers/infiniband/hw/hfi1/user_sdma.c 	struct sdma_mmu_node *node =
node             1541 drivers/infiniband/hw/hfi1/user_sdma.c 	if (!atomic_read(&node->refcount))
node              164 drivers/infiniband/hw/hfi1/user_sdma.h 	struct sdma_mmu_node *node;
node             1902 drivers/infiniband/hw/hfi1/verbs.c 	dd->verbs_dev.rdi.dparms.node = dd->node;
node              110 drivers/infiniband/hw/hfi1/vnic_main.c 	ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt);
node               45 drivers/infiniband/hw/mlx4/cm.c 	struct rb_node node;
node              147 drivers/infiniband/hw/mlx4/cm.c 	struct rb_node *node = sl_id_map->rb_node;
node              149 drivers/infiniband/hw/mlx4/cm.c 	while (node) {
node              151 drivers/infiniband/hw/mlx4/cm.c 			rb_entry(node, struct id_map_entry, node);
node              154 drivers/infiniband/hw/mlx4/cm.c 			node = node->rb_left;
node              156 drivers/infiniband/hw/mlx4/cm.c 			node = node->rb_right;
node              158 drivers/infiniband/hw/mlx4/cm.c 			node = node->rb_left;
node              160 drivers/infiniband/hw/mlx4/cm.c 			node = node->rb_right;
node              181 drivers/infiniband/hw/mlx4/cm.c 		rb_erase(&found_ent->node, sl_id_map);
node              202 drivers/infiniband/hw/mlx4/cm.c 		rb_replace_node(&ent->node, &new->node, sl_id_map);
node              209 drivers/infiniband/hw/mlx4/cm.c 		ent = rb_entry(parent, struct id_map_entry, node);
node              217 drivers/infiniband/hw/mlx4/cm.c 	rb_link_node(&new->node, parent, link);
node              218 drivers/infiniband/hw/mlx4/cm.c 	rb_insert_color(&new->node, sl_id_map);
node              411 drivers/infiniband/hw/mlx4/cm.c 					 struct id_map_entry, node);
node              413 drivers/infiniband/hw/mlx4/cm.c 			rb_erase(&ent->node, sl_id_map);
node              422 drivers/infiniband/hw/mlx4/cm.c 				rb_entry(nd, struct id_map_entry, node);
node              429 drivers/infiniband/hw/mlx4/cm.c 			rb_erase(&map->node, sl_id_map);
node              103 drivers/infiniband/hw/mlx4/mcg.c 	struct rb_node		node;
node              165 drivers/infiniband/hw/mlx4/mcg.c 	struct rb_node *node = ctx->mcg_table.rb_node;
node              169 drivers/infiniband/hw/mlx4/mcg.c 	while (node) {
node              170 drivers/infiniband/hw/mlx4/mcg.c 		group = rb_entry(node, struct mcast_group, node);
node              176 drivers/infiniband/hw/mlx4/mcg.c 			node = node->rb_left;
node              178 drivers/infiniband/hw/mlx4/mcg.c 			node = node->rb_right;
node              193 drivers/infiniband/hw/mlx4/mcg.c 		cur_group = rb_entry(parent, struct mcast_group, node);
node              204 drivers/infiniband/hw/mlx4/mcg.c 	rb_link_node(&group->node, parent, link);
node              205 drivers/infiniband/hw/mlx4/mcg.c 	rb_insert_color(&group->node, &ctx->mcg_table);
node              461 drivers/infiniband/hw/mlx4/mcg.c 			rb_erase(&group->node, &ctx->mcg_table);
node             1069 drivers/infiniband/hw/mlx4/mcg.c 	rb_erase(&group->node, &group->demux->mcg_table);
node             1103 drivers/infiniband/hw/mlx4/mcg.c 		group = rb_entry(p, struct mcast_group, node);
node             1232 drivers/infiniband/hw/mlx4/mcg.c 		group = rb_entry(p, struct mcast_group, node);
node              257 drivers/infiniband/hw/mlx5/cong.c 	enum mlx5_ib_cong_node_type node;
node              271 drivers/infiniband/hw/mlx5/cong.c 	node = mlx5_ib_param_to_node(offset);
node              273 drivers/infiniband/hw/mlx5/cong.c 	err = mlx5_cmd_query_cong_params(mdev, node, out, outlen);
node              293 drivers/infiniband/hw/mlx5/cong.c 	enum mlx5_ib_cong_node_type node;
node              312 drivers/infiniband/hw/mlx5/cong.c 	node = mlx5_ib_param_to_node(offset);
node              313 drivers/infiniband/hw/mlx5/cong.c 	MLX5_SET(modify_cong_params_in, in, cong_protocol, node);
node              578 drivers/infiniband/hw/mlx5/odp.c 	struct rb_node *node;
node              581 drivers/infiniband/hw/mlx5/odp.c 	for (node = rb_first_cached(&per_mm->umem_tree); node;
node              582 drivers/infiniband/hw/mlx5/odp.c 	     node = rb_next(node)) {
node              584 drivers/infiniband/hw/mlx5/odp.c 			rb_entry(node, struct ib_umem_odp, interval_tree.rb);
node               61 drivers/infiniband/hw/qib/qib_user_sdma.c 	struct rb_node node;
node              144 drivers/infiniband/hw/qib/qib_user_sdma.c 	struct rb_node *node = root->rb_node;
node              146 drivers/infiniband/hw/qib/qib_user_sdma.c 	while (node) {
node              147 drivers/infiniband/hw/qib/qib_user_sdma.c 		sdma_rb_node = rb_entry(node, struct qib_user_sdma_rb_node,
node              148 drivers/infiniband/hw/qib/qib_user_sdma.c 					node);
node              150 drivers/infiniband/hw/qib/qib_user_sdma.c 			node = node->rb_left;
node              152 drivers/infiniband/hw/qib/qib_user_sdma.c 			node = node->rb_right;
node              162 drivers/infiniband/hw/qib/qib_user_sdma.c 	struct rb_node **node = &(root->rb_node);
node              166 drivers/infiniband/hw/qib/qib_user_sdma.c 	while (*node) {
node              167 drivers/infiniband/hw/qib/qib_user_sdma.c 		got = rb_entry(*node, struct qib_user_sdma_rb_node, node);
node              168 drivers/infiniband/hw/qib/qib_user_sdma.c 		parent = *node;
node              170 drivers/infiniband/hw/qib/qib_user_sdma.c 			node = &((*node)->rb_left);
node              172 drivers/infiniband/hw/qib/qib_user_sdma.c 			node = &((*node)->rb_right);
node              177 drivers/infiniband/hw/qib/qib_user_sdma.c 	rb_link_node(&new->node, parent, node);
node              178 drivers/infiniband/hw/qib/qib_user_sdma.c 	rb_insert_color(&new->node, root);
node             1103 drivers/infiniband/hw/qib/qib_user_sdma.c 		rb_erase(&pq->sdma_rb_node->node, &qib_user_sdma_rb_root);
node             1603 drivers/infiniband/hw/qib/qib_verbs.c 	dd->verbs_dev.rdi.dparms.node = dd->assigned_node_id;
node               42 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define START(node) ((node)->start)
node               43 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define LAST(node) ((node)->last)
node               45 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MAKE_NODE(node, start, end, ref_cnt, flags, err, err_out)	\
node               47 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 			node = usnic_uiom_interval_node_alloc(start,	\
node               49 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 				if (!node) {				\
node               55 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MARK_FOR_ADD(node, list) (list_add_tail(&node->link, list))
node               57 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c #define MAKE_NODE_AND_APPEND(node, start, end, ref_cnt, flags, err,	\
node               60 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 					MAKE_NODE(node, start, end,	\
node               63 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 					MARK_FOR_ADD(node, list);	\
node              107 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 	struct usnic_uiom_interval_node *node;
node              111 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 	for (node = usnic_uiom_interval_tree_iter_first(root, start, last);
node              112 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 		node;
node              113 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 		node = usnic_uiom_interval_tree_iter_next(node, start, last))
node              114 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c 		list_add_tail(&node->link, list);
node               50 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_insert(struct usnic_uiom_interval_node *node,
node               53 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_remove(struct usnic_uiom_interval_node *node,
node               60 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.h usnic_uiom_interval_tree_iter_next(struct usnic_uiom_interval_node *node,
node              240 drivers/infiniband/sw/rdmavt/cq.c 		k_wc = vzalloc_node(sz, rdi->dparms.node);
node              289 drivers/infiniband/sw/rdmavt/cq.c 			cpumask_first(cpumask_of_node(rdi->dparms.node));
node              406 drivers/infiniband/sw/rdmavt/cq.c 		k_wc = vzalloc_node(sz, rdi->dparms.node);
node              167 drivers/infiniband/sw/rdmavt/mmap.c 	ip = kmalloc_node(sizeof(*ip), GFP_KERNEL, rdi->dparms.node);
node               91 drivers/infiniband/sw/rdmavt/mr.c 			       vmalloc_node(lk_tab_size, rdi->dparms.node);
node              146 drivers/infiniband/sw/rdmavt/mr.c 					  dev->dparms.node);
node              174 drivers/infiniband/sw/rdmavt/qp.c 	int node = rdi->dparms.node;
node              181 drivers/infiniband/sw/rdmavt/qp.c 	rdi->wss = kzalloc_node(sizeof(*rdi->wss), GFP_KERNEL, node);
node              219 drivers/infiniband/sw/rdmavt/qp.c 				    GFP_KERNEL, node);
node              419 drivers/infiniband/sw/rdmavt/qp.c 				   rdi->dparms.node);
node              429 drivers/infiniband/sw/rdmavt/qp.c 			     GFP_KERNEL, rdi->dparms.node);
node              822 drivers/infiniband/sw/rdmavt/qp.c int rvt_alloc_rq(struct rvt_rq *rq, u32 size, int node,
node              830 drivers/infiniband/sw/rdmavt/qp.c 		rq->kwq = kzalloc_node(sizeof(*rq->kwq), GFP_KERNEL, node);
node              837 drivers/infiniband/sw/rdmavt/qp.c 			vzalloc_node(sizeof(struct rvt_krwq) + size, node);
node             1032 drivers/infiniband/sw/rdmavt/qp.c static int alloc_ud_wq_attr(struct rvt_qp *qp, int node)
node             1040 drivers/infiniband/sw/rdmavt/qp.c 					       GFP_KERNEL, node);
node             1113 drivers/infiniband/sw/rdmavt/qp.c 		swq = vzalloc_node(array_size(sz, sqsize), rdi->dparms.node);
node             1129 drivers/infiniband/sw/rdmavt/qp.c 				  rdi->dparms.node);
node             1140 drivers/infiniband/sw/rdmavt/qp.c 					     rdi->dparms.node);
node             1171 drivers/infiniband/sw/rdmavt/qp.c 					   rdi->dparms.node, udata);
node             1196 drivers/infiniband/sw/rdmavt/qp.c 		err = alloc_ud_wq_attr(qp, rdi->dparms.node);
node               71 drivers/infiniband/sw/rdmavt/qp.h int rvt_alloc_rq(struct rvt_rq *rq, u32 size, int node,
node              101 drivers/infiniband/sw/rdmavt/srq.c 			 dev->dparms.node, udata)) {
node              189 drivers/infiniband/sw/rdmavt/srq.c 		if (rvt_alloc_rq(&tmp_rq, size * sz, dev->dparms.node,
node              291 drivers/infiniband/sw/rxe/rxe_pool.c 		elem = rb_entry(parent, struct rxe_pool_entry, node);
node              304 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_link_node(&new->node, parent, link);
node              305 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_insert_color(&new->node, &pool->tree);
node              319 drivers/infiniband/sw/rxe/rxe_pool.c 		elem = rb_entry(parent, struct rxe_pool_entry, node);
node              335 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_link_node(&new->node, parent, link);
node              336 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_insert_color(&new->node, &pool->tree);
node              360 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_erase(&elem->node, &pool->tree);
node              384 drivers/infiniband/sw/rxe/rxe_pool.c 	rb_erase(&elem->node, &pool->tree);
node              479 drivers/infiniband/sw/rxe/rxe_pool.c 	struct rb_node *node = NULL;
node              488 drivers/infiniband/sw/rxe/rxe_pool.c 	node = pool->tree.rb_node;
node              490 drivers/infiniband/sw/rxe/rxe_pool.c 	while (node) {
node              491 drivers/infiniband/sw/rxe/rxe_pool.c 		elem = rb_entry(node, struct rxe_pool_entry, node);
node              494 drivers/infiniband/sw/rxe/rxe_pool.c 			node = node->rb_left;
node              496 drivers/infiniband/sw/rxe/rxe_pool.c 			node = node->rb_right;
node              505 drivers/infiniband/sw/rxe/rxe_pool.c 	return node ? elem : NULL;
node              510 drivers/infiniband/sw/rxe/rxe_pool.c 	struct rb_node *node = NULL;
node              520 drivers/infiniband/sw/rxe/rxe_pool.c 	node = pool->tree.rb_node;
node              522 drivers/infiniband/sw/rxe/rxe_pool.c 	while (node) {
node              523 drivers/infiniband/sw/rxe/rxe_pool.c 		elem = rb_entry(node, struct rxe_pool_entry, node);
node              529 drivers/infiniband/sw/rxe/rxe_pool.c 			node = node->rb_left;
node              531 drivers/infiniband/sw/rxe/rxe_pool.c 			node = node->rb_right;
node              536 drivers/infiniband/sw/rxe/rxe_pool.c 	if (node)
node              541 drivers/infiniband/sw/rxe/rxe_pool.c 	return node ? elem : NULL;
node               88 drivers/infiniband/sw/rxe/rxe_pool.h 	struct rb_node		node;
node              188 drivers/infiniband/sw/siw/siw_main.c 	int i, num_cpus, cpu, min_use, node = sdev->numa_node, tx_cpu = -1;
node              190 drivers/infiniband/sw/siw/siw_main.c 	if (node < 0)
node              193 drivers/infiniband/sw/siw/siw_main.c 		tx_cpumask = siw_cpu_info.tx_valid_cpus[node];
node              221 drivers/infiniband/sw/siw/siw_main.c 		"tx cpu %d, node %d, %d qp's\n", tx_cpu, node, min_use);
node              393 drivers/infiniband/ulp/isert/ib_isert.c 	INIT_LIST_HEAD(&isert_conn->node);
node              545 drivers/infiniband/ulp/isert/ib_isert.c 	list_add_tail(&isert_conn->node, &isert_np->accepted);
node              606 drivers/infiniband/ulp/isert/ib_isert.c 	list_move_tail(&isert_conn->node, &isert_np->pending);
node              637 drivers/infiniband/ulp/isert/ib_isert.c 	if (!list_empty(&isert_conn->node)) {
node              642 drivers/infiniband/ulp/isert/ib_isert.c 		list_del_init(&isert_conn->node);
node              739 drivers/infiniband/ulp/isert/ib_isert.c 	list_del_init(&isert_conn->node);
node             2489 drivers/infiniband/ulp/isert/ib_isert.c 			struct isert_conn, node);
node             2490 drivers/infiniband/ulp/isert/ib_isert.c 	list_del_init(&isert_conn->node);
node             2525 drivers/infiniband/ulp/isert/ib_isert.c 					 node) {
node             2536 drivers/infiniband/ulp/isert/ib_isert.c 					 node) {
node              152 drivers/infiniband/ulp/isert/ib_isert.h 	struct list_head	node;
node              105 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	struct opa_vnic_mac_tbl_node *node;
node              112 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) {
node              113 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		hash_del(&node->hlist);
node              114 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		kfree(node);
node              155 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	struct opa_vnic_mac_tbl_node *node;
node              168 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each(mactbl, bkt, node, hlist) {
node              169 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct __opa_vnic_mactable_entry *nentry = &node->entry;
node              172 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if ((node->index < loffset) ||
node              173 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		    (node->index >= (loffset + lnum_entries)))
node              177 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		entry = &tbl->tbl_entries[node->index - loffset];
node              207 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	struct opa_vnic_mac_tbl_node *node, *new_node;
node              241 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node              242 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if (!node) {
node              247 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		node->index = loffset + i;
node              248 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		nentry = &node->entry;
node              254 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		key = node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX];
node              255 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		vnic_hash_add(new_mactbl, &node->hlist, key);
node              263 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each(old_mactbl, bkt, node, hlist) {
node              264 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if ((node->index >= loffset) &&
node              265 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		    (node->index < (loffset + lnum_entries)))
node              274 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		new_node->index = node->index;
node              275 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		memcpy(&new_node->entry, &node->entry, sizeof(node->entry));
node              301 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	struct opa_vnic_mac_tbl_node *node;
node              312 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 	vnic_hash_for_each_possible(mactbl, node, hlist, key) {
node              313 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		struct __opa_vnic_mactable_entry *entry = &node->entry;
node              319 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 		if (!memcmp(node->entry.mac_addr, mac_hdr->h_dest,
node              320 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 			    ARRAY_SIZE(node->entry.mac_addr))) {
node              322 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c 			dlid = OPA_VNIC_DLID_SD_GET_DLID(node->entry.dlid_sd);
node              276 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_add(hashtable, node, key)                                   \
node              277 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h 	hlist_add_head(node,                                                  \
node             3778 drivers/infiniband/ulp/srp/ib_srp.c 	int ret, node_idx, node, cpu, i;
node             3912 drivers/infiniband/ulp/srp/ib_srp.c 	for_each_online_node(node) {
node             3924 drivers/infiniband/ulp/srp/ib_srp.c 			if (cpu_to_node(cpu) != node)
node               48 drivers/input/evdev.c 	struct list_head node;
node              305 drivers/input/evdev.c 		list_for_each_entry_rcu(client, &evdev->client_list, node)
node              376 drivers/input/evdev.c 	list_add_tail_rcu(&client->node, &evdev->client_list);
node              384 drivers/input/evdev.c 	list_del_rcu(&client->node);
node              428 drivers/input/evdev.c 	list_for_each_entry(client, &evdev->client_list, node)
node              263 drivers/input/gameport/gameport.c 	struct list_head node;
node              278 drivers/input/gameport/gameport.c 					 struct gameport_event, node);
node              279 drivers/input/gameport/gameport.c 		list_del_init(&event->node);
node              299 drivers/input/gameport/gameport.c 	list_for_each_entry_safe(e, next, &gameport_event_list, node) {
node              309 drivers/input/gameport/gameport.c 			list_del_init(&e->node);
node              368 drivers/input/gameport/gameport.c 	list_for_each_entry_reverse(event, &gameport_event_list, node) {
node              395 drivers/input/gameport/gameport.c 	list_add_tail(&event->node, &gameport_event_list);
node              414 drivers/input/gameport/gameport.c 	list_for_each_entry_safe(event, next, &gameport_event_list, node) {
node              416 drivers/input/gameport/gameport.c 			list_del_init(&event->node);
node              440 drivers/input/gameport/gameport.c 	list_for_each_entry(event, &gameport_event_list, node) {
node              539 drivers/input/gameport/gameport.c 	INIT_LIST_HEAD(&gameport->node);
node              559 drivers/input/gameport/gameport.c 	list_add_tail(&gameport->node, &gameport_list);
node              597 drivers/input/gameport/gameport.c 	list_del_init(&gameport->node);
node              765 drivers/input/gameport/gameport.c 	list_for_each_entry(gameport, &gameport_list, node) {
node               37 drivers/input/gameport/ns558.c 	struct list_head node;
node              146 drivers/input/gameport/ns558.c 	list_add(&ns558->node, &ns558_list);
node              219 drivers/input/gameport/ns558.c 	list_add_tail(&ns558->node, &ns558_list);
node              260 drivers/input/gameport/ns558.c 	list_for_each_entry_safe(ns558, safe, &ns558_list, node) {
node             1165 drivers/input/input.c 	struct input_dev *dev = container_of(v, struct input_dev, node);
node             1259 drivers/input/input.c 	struct input_handler *handler = container_of(v, struct input_handler, node);
node             1826 drivers/input/input.c 		INIT_LIST_HEAD(&dev->node);
node             2097 drivers/input/input.c 	list_del_init(&dev->node);
node             2228 drivers/input/input.c 	list_add_tail(&dev->node, &input_dev_list);
node             2230 drivers/input/input.c 	list_for_each_entry(handler, &input_handler_list, node)
node             2300 drivers/input/input.c 	list_add_tail(&handler->node, &input_handler_list);
node             2302 drivers/input/input.c 	list_for_each_entry(dev, &input_dev_list, node)
node             2329 drivers/input/input.c 	list_del_init(&handler->node);
node               66 drivers/input/joydev.c 	struct list_head node;
node              147 drivers/input/joydev.c 	list_for_each_entry_rcu(client, &joydev->client_list, node)
node              173 drivers/input/joydev.c 	list_add_tail_rcu(&client->node, &joydev->client_list);
node              181 drivers/input/joydev.c 	list_del_rcu(&client->node);
node              238 drivers/input/joydev.c 	list_for_each_entry(client, &joydev->client_list, node)
node              251 drivers/input/keyboard/cap11xx.c 	struct device_node *node = dev->of_node, *child;
node              253 drivers/input/keyboard/cap11xx.c 	int cnt = of_get_child_count(node);
node              280 drivers/input/keyboard/cap11xx.c 	for_each_child_of_node(node, child) {
node              326 drivers/input/keyboard/cap11xx.c 	struct device_node *node;
node              378 drivers/input/keyboard/cap11xx.c 	node = dev->of_node;
node              380 drivers/input/keyboard/cap11xx.c 	if (!of_property_read_u32(node, "microchip,sensor-gain", &gain32)) {
node              387 drivers/input/keyboard/cap11xx.c 	if (of_property_read_bool(node, "microchip,irq-active-high")) {
node              398 drivers/input/keyboard/cap11xx.c 	of_property_read_u32_array(node, "linux,keycodes",
node              420 drivers/input/keyboard/cap11xx.c 	if (of_property_read_bool(node, "autorepeat"))
node              455 drivers/input/keyboard/cap11xx.c 	irq = irq_of_parse_and_map(node, 0);
node              243 drivers/input/keyboard/mtk-pmic-keys.c 	struct device_node *node = pdev->dev.of_node, *child;
node              270 drivers/input/keyboard/mtk-pmic-keys.c 	keycount = of_get_available_child_count(node);
node              276 drivers/input/keyboard/mtk-pmic-keys.c 	for_each_child_of_node(node, child) {
node              169 drivers/input/misc/atmel_captouch.c 	struct device_node *node;
node              206 drivers/input/misc/atmel_captouch.c 	node = dev->of_node;
node              207 drivers/input/misc/atmel_captouch.c 	if (!node) {
node              212 drivers/input/misc/atmel_captouch.c 	if (of_property_read_bool(node, "autorepeat"))
node              215 drivers/input/misc/atmel_captouch.c 	capdev->num_btn = of_property_count_u32_elems(node, "linux,keymap");
node              219 drivers/input/misc/atmel_captouch.c 	err = of_property_read_u32_array(node, "linux,keycodes",
node              119 drivers/input/misc/regulator-haptic.c 	struct device_node *node;
node              122 drivers/input/misc/regulator-haptic.c 	node = dev->of_node;
node              123 drivers/input/misc/regulator-haptic.c 	if(!node) {
node              128 drivers/input/misc/regulator-haptic.c 	error = of_property_read_u32(node, "max-microvolt", &haptic->max_volt);
node              134 drivers/input/misc/regulator-haptic.c 	error = of_property_read_u32(node, "min-microvolt", &haptic->min_volt);
node              169 drivers/input/misc/twl4030-vibra.c 	struct device_node *node;
node              174 drivers/input/misc/twl4030-vibra.c 	node = of_get_child_by_name(parent, "codec");
node              175 drivers/input/misc/twl4030-vibra.c 	if (node) {
node              176 drivers/input/misc/twl4030-vibra.c 		of_node_put(node);
node               21 drivers/input/mouse/psmouse-smbus.c 	struct list_head node;
node               38 drivers/input/mouse/psmouse-smbus.c 	list_for_each_entry(smbdev, &psmouse_smbus_list, node) {
node               66 drivers/input/mouse/psmouse-smbus.c 	list_for_each_entry_safe(smbdev, tmp, &psmouse_smbus_list, node) {
node               80 drivers/input/mouse/psmouse-smbus.c 			list_del(&smbdev->node);
node              173 drivers/input/mouse/psmouse-smbus.c 		list_del(&smbdev->node);
node              216 drivers/input/mouse/psmouse-smbus.c 	list_for_each_entry_safe(smbdev, tmp, &psmouse_smbus_list, node) {
node              218 drivers/input/mouse/psmouse-smbus.c 			list_del(&smbdev->node);
node              263 drivers/input/mouse/psmouse-smbus.c 	list_add_tail(&smbdev->node, &psmouse_smbus_list);
node              283 drivers/input/mouse/psmouse-smbus.c 		list_del(&smbdev->node);
node               97 drivers/input/mousedev.c 	struct list_head node;
node              270 drivers/input/mousedev.c 	list_for_each_entry_rcu(client, &mousedev->client_list, node) {
node              507 drivers/input/mousedev.c 	list_add_tail_rcu(&client->node, &mousedev->client_list);
node              515 drivers/input/mousedev.c 	list_del_rcu(&client->node);
node              805 drivers/input/mousedev.c 	list_for_each_entry(client, &mousedev->client_list, node)
node              158 drivers/input/rmi4/rmi_bus.c 	struct device_node *node = fn->rmi_dev->xport->dev->of_node;
node              162 drivers/input/rmi4/rmi_bus.c 	fn->dev.of_node = of_get_child_by_name(node, of_name);
node               40 drivers/input/rmi4/rmi_bus.h 	struct list_head node;
node               44 drivers/input/rmi4/rmi_driver.c 					 &data->function_list, node) {
node               45 drivers/input/rmi4/rmi_driver.c 		list_del(&fn->node);
node              104 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
node              119 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
node              248 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
node              281 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
node              315 drivers/input/rmi4/rmi_driver.c 	list_for_each_entry(entry, &data->function_list, node) {
node              850 drivers/input/rmi4/rmi_driver.c 	INIT_LIST_HEAD(&fn->node);
node              871 drivers/input/rmi4/rmi_driver.c 	list_add_tail(&fn->node, &data->function_list);
node               90 drivers/input/serio/gscps2.c 	struct list_head node;
node              231 drivers/input/serio/gscps2.c 	list_for_each_entry(ps2port, &ps2port_list, node) {
node              249 drivers/input/serio/gscps2.c 	list_for_each_entry(ps2port, &ps2port_list, node) {
node              392 drivers/input/serio/gscps2.c 	list_add_tail(&ps2port->node, &ps2port_list);
node              421 drivers/input/serio/gscps2.c 	list_del(&ps2port->node);
node              583 drivers/input/serio/hil_mlc.c static inline void hilse_setup_input(hil_mlc *mlc, const struct hilse_node *node)
node              586 drivers/input/serio/hil_mlc.c 	switch (node->act) {
node              588 drivers/input/serio/hil_mlc.c 		mlc->imatch = node->object.packet;
node              592 drivers/input/serio/hil_mlc.c 		mlc->imatch = node->object.packet;
node              596 drivers/input/serio/hil_mlc.c 		mlc->imatch = node->object.packet;
node              605 drivers/input/serio/hil_mlc.c 	mlc->intimeout = usecs_to_jiffies(node->arg);
node              619 drivers/input/serio/hil_mlc.c 	const struct hilse_node *node;
node              633 drivers/input/serio/hil_mlc.c 	node = hil_mlc_se + mlc->seidx;
node              635 drivers/input/serio/hil_mlc.c 	switch (node->act) {
node              640 drivers/input/serio/hil_mlc.c 		BUG_ON(node->object.func == NULL);
node              641 drivers/input/serio/hil_mlc.c 		rc = node->object.func(mlc, node->arg);
node              642 drivers/input/serio/hil_mlc.c 		nextidx = (rc > 0) ? node->ugly :
node              643 drivers/input/serio/hil_mlc.c 			((rc < 0) ? node->bad : node->good);
node              654 drivers/input/serio/hil_mlc.c 		rc = mlc->in(mlc, node->arg);
node              662 drivers/input/serio/hil_mlc.c 			nextidx = node->ugly;
node              664 drivers/input/serio/hil_mlc.c 			nextidx = node->good;
node              666 drivers/input/serio/hil_mlc.c 			nextidx = node->bad;
node              673 drivers/input/serio/hil_mlc.c 		pack = node->object.packet;
node              679 drivers/input/serio/hil_mlc.c 		pack = node->object.packet;
node              685 drivers/input/serio/hil_mlc.c 		pack = node->object.packet;
node              689 drivers/input/serio/hil_mlc.c 			if ((node + 1)->act & HILSE_IN)
node              690 drivers/input/serio/hil_mlc.c 				hilse_setup_input(mlc, node + 1);
node              718 drivers/input/serio/hil_mlc.c 		nextidx = mlc->cts(mlc) ? node->bad : node->good;
node              144 drivers/input/serio/serio.c 	struct list_head node;
node              159 drivers/input/serio/serio.c 					 struct serio_event, node);
node              160 drivers/input/serio/serio.c 		list_del_init(&event->node);
node              181 drivers/input/serio/serio.c 	list_for_each_entry_safe(e, next, &serio_event_list, node) {
node              191 drivers/input/serio/serio.c 			list_del_init(&e->node);
node              256 drivers/input/serio/serio.c 	list_for_each_entry_reverse(event, &serio_event_list, node) {
node              283 drivers/input/serio/serio.c 	list_add_tail(&event->node, &serio_event_list);
node              302 drivers/input/serio/serio.c 	list_for_each_entry_safe(event, next, &serio_event_list, node) {
node              304 drivers/input/serio/serio.c 			list_del_init(&event->node);
node              326 drivers/input/serio/serio.c 	list_for_each_entry(event, &serio_event_list, node) {
node              503 drivers/input/serio/serio.c 	INIT_LIST_HEAD(&serio->node);
node              537 drivers/input/serio/serio.c 	list_add_tail(&serio->node, &serio_list);
node              575 drivers/input/serio/serio.c 	list_del_init(&serio->node);
node              863 drivers/input/serio/serio.c 	list_for_each_entry(serio, &serio_list, node) {
node               38 drivers/input/serio/serio_raw.c 	struct list_head node;
node               45 drivers/input/serio/serio_raw.c 	struct list_head node;
node               66 drivers/input/serio/serio_raw.c 	list_for_each_entry(serio_raw, &serio_raw_list, node) {
node              107 drivers/input/serio/serio_raw.c 	list_add_tail(&client->node, &serio_raw->client_list);
node              130 drivers/input/serio/serio_raw.c 	list_del(&client->node);
node              282 drivers/input/serio/serio_raw.c 		list_for_each_entry(client, &serio_raw->client_list, node)
node              321 drivers/input/serio/serio_raw.c 	list_add_tail(&serio_raw->node, &serio_raw_list);
node              347 drivers/input/serio/serio_raw.c 	list_del_init(&serio_raw->node);
node              383 drivers/input/serio/serio_raw.c 	list_for_each_entry(client, &serio_raw->client_list, node)
node              399 drivers/input/serio/serio_raw.c 	list_del_init(&serio_raw->node);
node             1185 drivers/input/touchscreen/ads7846.c 	struct device_node *node = dev->of_node;
node             1189 drivers/input/touchscreen/ads7846.c 	if (!node) {
node             1206 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,vref-delay-usecs",
node             1208 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,vref-mv", &pdata->vref_mv);
node             1209 drivers/input/touchscreen/ads7846.c 	pdata->keep_vref_on = of_property_read_bool(node, "ti,keep-vref-on");
node             1211 drivers/input/touchscreen/ads7846.c 	pdata->swap_xy = of_property_read_bool(node, "ti,swap-xy");
node             1213 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,settle-delay-usec",
node             1215 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,penirq-recheck-delay-usecs",
node             1218 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,x-plate-ohms", &pdata->x_plate_ohms);
node             1219 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,y-plate-ohms", &pdata->y_plate_ohms);
node             1221 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,x-min", &pdata->x_min);
node             1222 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,y-min", &pdata->y_min);
node             1223 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,x-max", &pdata->x_max);
node             1224 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,y-max", &pdata->y_max);
node             1230 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,pressure-min", &pdata->pressure_min);
node             1231 drivers/input/touchscreen/ads7846.c 	if (!of_property_read_u32(node, "touchscreen-min-pressure", &value))
node             1233 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,pressure-max", &pdata->pressure_max);
node             1235 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,debounce-max", &pdata->debounce_max);
node             1236 drivers/input/touchscreen/ads7846.c 	if (!of_property_read_u32(node, "touchscreen-average-samples", &value))
node             1238 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,debounce-tol", &pdata->debounce_tol);
node             1239 drivers/input/touchscreen/ads7846.c 	of_property_read_u16(node, "ti,debounce-rep", &pdata->debounce_rep);
node             1241 drivers/input/touchscreen/ads7846.c 	of_property_read_u32(node, "ti,pendown-gpio-debounce",
node             1244 drivers/input/touchscreen/ads7846.c 	pdata->wakeup = of_property_read_bool(node, "wakeup-source") ||
node             1245 drivers/input/touchscreen/ads7846.c 			of_property_read_bool(node, "linux,wakeup");
node              606 drivers/input/touchscreen/mxs-lradc-ts.c 	struct device_node *node = dev->parent->of_node;
node              626 drivers/input/touchscreen/mxs-lradc-ts.c 	ret = of_property_read_u32(node, "fsl,lradc-touchscreen-wires",
node              631 drivers/input/touchscreen/mxs-lradc-ts.c 	if (of_property_read_u32(node, "fsl,ave-ctrl", &adapt)) {
node              643 drivers/input/touchscreen/mxs-lradc-ts.c 	if (of_property_read_u32(node, "fsl,ave-delay", &adapt)) {
node              655 drivers/input/touchscreen/mxs-lradc-ts.c 	if (of_property_read_u32(node, "fsl,settling", &adapt)) {
node              678 drivers/input/touchscreen/mxs-lradc-ts.c 		virq = irq_of_parse_and_map(node, irq);
node              885 drivers/input/touchscreen/sur40.c 	struct sur40_buffer *buf, *node;
node              888 drivers/input/touchscreen/sur40.c 	list_for_each_entry_safe(buf, node, &sur40->buf_list, list) {
node              344 drivers/input/touchscreen/ti_am335x_tsc.c 	struct device_node *node = pdev->dev.of_node;
node              347 drivers/input/touchscreen/ti_am335x_tsc.c 	if (!node)
node              350 drivers/input/touchscreen/ti_am335x_tsc.c 	err = of_property_read_u32(node, "ti,wires", &ts_dev->wires);
node              362 drivers/input/touchscreen/ti_am335x_tsc.c 	err = of_property_read_u32(node, "ti,x-plate-resistance",
node              371 drivers/input/touchscreen/ti_am335x_tsc.c 	err = of_property_read_u32(node, "ti,coordinate-readouts",
node              375 drivers/input/touchscreen/ti_am335x_tsc.c 		err = of_property_read_u32(node, "ti,coordiante-readouts",
node              388 drivers/input/touchscreen/ti_am335x_tsc.c 	err = of_property_read_u32(node, "ti,charge-delay",
node              399 drivers/input/touchscreen/ti_am335x_tsc.c 	return of_property_read_u32_array(node, "ti,wire-config",
node               38 drivers/interconnect/core.c 	struct icc_node *node;
node              105 drivers/interconnect/core.c 	struct icc_node *node = dst;
node              116 drivers/interconnect/core.c 		node->provider->users++;
node              117 drivers/interconnect/core.c 		hlist_add_head(&path->reqs[i].req_node, &node->req_list);
node              118 drivers/interconnect/core.c 		path->reqs[i].node = node;
node              121 drivers/interconnect/core.c 		node = node->reverse;
node              131 drivers/interconnect/core.c 	struct icc_node *n, *node = NULL;
node              146 drivers/interconnect/core.c 		list_for_each_entry_safe(node, n, &traverse_list, search_list) {
node              147 drivers/interconnect/core.c 			if (node == dst) {
node              153 drivers/interconnect/core.c 			for (i = 0; i < node->num_links; i++) {
node              154 drivers/interconnect/core.c 				struct icc_node *tmp = node->links[i];
node              165 drivers/interconnect/core.c 				tmp->reverse = node;
node              200 drivers/interconnect/core.c static int aggregate_requests(struct icc_node *node)
node              202 drivers/interconnect/core.c 	struct icc_provider *p = node->provider;
node              205 drivers/interconnect/core.c 	node->avg_bw = 0;
node              206 drivers/interconnect/core.c 	node->peak_bw = 0;
node              209 drivers/interconnect/core.c 		p->pre_aggregate(node);
node              211 drivers/interconnect/core.c 	hlist_for_each_entry(r, &node->req_list, req_node)
node              212 drivers/interconnect/core.c 		p->aggregate(node, r->tag, r->avg_bw, r->peak_bw,
node              213 drivers/interconnect/core.c 			     &node->avg_bw, &node->peak_bw);
node              225 drivers/interconnect/core.c 		next = path->reqs[i].node;
node              284 drivers/interconnect/core.c 	struct icc_node *node = ERR_PTR(-EPROBE_DEFER);
node              293 drivers/interconnect/core.c 			node = provider->xlate(spec, provider->data);
node              294 drivers/interconnect/core.c 		if (!IS_ERR(node))
node              299 drivers/interconnect/core.c 	return node;
node              434 drivers/interconnect/core.c 	struct icc_node *node;
node              448 drivers/interconnect/core.c 		node = path->reqs[i].node;
node              455 drivers/interconnect/core.c 		aggregate_requests(node);
node              464 drivers/interconnect/core.c 			node = path->reqs[i].node;
node              467 drivers/interconnect/core.c 			aggregate_requests(node);
node              528 drivers/interconnect/core.c 	struct icc_node *node;
node              541 drivers/interconnect/core.c 		node = path->reqs[i].node;
node              543 drivers/interconnect/core.c 		if (!WARN_ON(!node->provider->users))
node              544 drivers/interconnect/core.c 			node->provider->users--;
node              554 drivers/interconnect/core.c 	struct icc_node *node;
node              557 drivers/interconnect/core.c 	node = node_find(id);
node              558 drivers/interconnect/core.c 	if (node)
node              559 drivers/interconnect/core.c 		return node;
node              561 drivers/interconnect/core.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              562 drivers/interconnect/core.c 	if (!node)
node              565 drivers/interconnect/core.c 	id = idr_alloc(&icc_idr, node, id, id + 1, GFP_KERNEL);
node              568 drivers/interconnect/core.c 		kfree(node);
node              572 drivers/interconnect/core.c 	node->id = id;
node              574 drivers/interconnect/core.c 	return node;
node              585 drivers/interconnect/core.c 	struct icc_node *node;
node              589 drivers/interconnect/core.c 	node = icc_node_create_nolock(id);
node              593 drivers/interconnect/core.c 	return node;
node              603 drivers/interconnect/core.c 	struct icc_node *node;
node              607 drivers/interconnect/core.c 	node = node_find(id);
node              608 drivers/interconnect/core.c 	if (node) {
node              609 drivers/interconnect/core.c 		idr_remove(&icc_idr, node->id);
node              610 drivers/interconnect/core.c 		WARN_ON(!hlist_empty(&node->req_list));
node              615 drivers/interconnect/core.c 	kfree(node);
node              632 drivers/interconnect/core.c int icc_link_create(struct icc_node *node, const int dst_id)
node              638 drivers/interconnect/core.c 	if (!node->provider)
node              653 drivers/interconnect/core.c 	new = krealloc(node->links,
node              654 drivers/interconnect/core.c 		       (node->num_links + 1) * sizeof(*node->links),
node              661 drivers/interconnect/core.c 	node->links = new;
node              662 drivers/interconnect/core.c 	node->links[node->num_links++] = dst;
node              720 drivers/interconnect/core.c void icc_node_add(struct icc_node *node, struct icc_provider *provider)
node              724 drivers/interconnect/core.c 	node->provider = provider;
node              725 drivers/interconnect/core.c 	list_add_tail(&node->node_list, &provider->nodes);
node              735 drivers/interconnect/core.c void icc_node_del(struct icc_node *node)
node              739 drivers/interconnect/core.c 	list_del(&node->node_list);
node              330 drivers/interconnect/qcom/qcs404.c static int qcom_icc_aggregate(struct icc_node *node, u32 tag, u32 avg_bw,
node              417 drivers/interconnect/qcom/qcs404.c 	struct icc_node *node, *tmp;
node              473 drivers/interconnect/qcom/qcs404.c 		node = icc_node_create(qnodes[i]->id);
node              474 drivers/interconnect/qcom/qcs404.c 		if (IS_ERR(node)) {
node              475 drivers/interconnect/qcom/qcs404.c 			ret = PTR_ERR(node);
node              479 drivers/interconnect/qcom/qcs404.c 		node->name = qnodes[i]->name;
node              480 drivers/interconnect/qcom/qcs404.c 		node->data = qnodes[i];
node              481 drivers/interconnect/qcom/qcs404.c 		icc_node_add(node, provider);
node              483 drivers/interconnect/qcom/qcs404.c 		dev_dbg(dev, "registered node %s\n", node->name);
node              487 drivers/interconnect/qcom/qcs404.c 			icc_link_create(node, qnodes[i]->links[j]);
node              489 drivers/interconnect/qcom/qcs404.c 		data->nodes[i] = node;
node              497 drivers/interconnect/qcom/qcs404.c 	list_for_each_entry_safe(node, tmp, &provider->nodes, node_list) {
node              498 drivers/interconnect/qcom/qcs404.c 		icc_node_del(node);
node              499 drivers/interconnect/qcom/qcs404.c 		icc_node_destroy(node->id);
node              633 drivers/interconnect/qcom/sdm845.c static void qcom_icc_pre_aggregate(struct icc_node *node)
node              638 drivers/interconnect/qcom/sdm845.c 	qn = node->data;
node              646 drivers/interconnect/qcom/sdm845.c static int qcom_icc_aggregate(struct icc_node *node, u32 tag, u32 avg_bw,
node              652 drivers/interconnect/qcom/sdm845.c 	qn = node->data;
node              676 drivers/interconnect/qcom/sdm845.c 	struct icc_node *node;
node              683 drivers/interconnect/qcom/sdm845.c 		node = dst;
node              685 drivers/interconnect/qcom/sdm845.c 		node = src;
node              687 drivers/interconnect/qcom/sdm845.c 	qp = to_qcom_provider(node->provider);
node              778 drivers/interconnect/qcom/sdm845.c 	struct icc_node *node;
node              820 drivers/interconnect/qcom/sdm845.c 		node = icc_node_create(qnodes[i]->id);
node              821 drivers/interconnect/qcom/sdm845.c 		if (IS_ERR(node)) {
node              822 drivers/interconnect/qcom/sdm845.c 			ret = PTR_ERR(node);
node              826 drivers/interconnect/qcom/sdm845.c 		node->name = qnodes[i]->name;
node              827 drivers/interconnect/qcom/sdm845.c 		node->data = qnodes[i];
node              828 drivers/interconnect/qcom/sdm845.c 		icc_node_add(node, provider);
node              830 drivers/interconnect/qcom/sdm845.c 		dev_dbg(&pdev->dev, "registered node %p %s %d\n", node,
node              831 drivers/interconnect/qcom/sdm845.c 			qnodes[i]->name, node->id);
node              835 drivers/interconnect/qcom/sdm845.c 			icc_link_create(node, qnodes[i]->links[j]);
node              837 drivers/interconnect/qcom/sdm845.c 		data->nodes[i] = node;
node              858 drivers/interconnect/qcom/sdm845.c 	list_for_each_entry(node, &provider->nodes, node_list) {
node              859 drivers/interconnect/qcom/sdm845.c 		icc_node_del(node);
node              860 drivers/interconnect/qcom/sdm845.c 		icc_node_destroy(node->id);
node              215 drivers/iommu/amd_iommu.c 	struct llist_node *node;
node              220 drivers/iommu/amd_iommu.c 	node = dev_data_list.first;
node              221 drivers/iommu/amd_iommu.c 	llist_for_each_entry(dev_data, node, dev_data_list) {
node             3523 drivers/iommu/arm-smmu-v3.c 	struct acpi_iort_node *node;
node             3525 drivers/iommu/arm-smmu-v3.c 	node = *(struct acpi_iort_node **)dev_get_platdata(dev);
node             3528 drivers/iommu/arm-smmu-v3.c 	iort_smmu = (struct acpi_iort_smmu_v3 *)node->node_data;
node              155 drivers/iommu/arm-smmu.c 	struct device_node *np = it->node;
node              160 drivers/iommu/arm-smmu.c 		if (it->node == np) {
node              164 drivers/iommu/arm-smmu.c 	it->node = np;
node              188 drivers/iommu/arm-smmu.c 	it.node = np;
node             1920 drivers/iommu/arm-smmu.c 	struct acpi_iort_node *node =
node             1926 drivers/iommu/arm-smmu.c 	iort_smmu = (struct acpi_iort_smmu *)node->node_data;
node              227 drivers/iommu/dma-iommu.c 		if (window->node.next == &bridge->dma_ranges &&
node              965 drivers/iommu/dma-iommu.c 	int node = dev_to_node(dev);
node              971 drivers/iommu/dma-iommu.c 		page = alloc_pages_node(node, gfp, get_order(alloc_size));
node              476 drivers/iommu/dmar.c 			int node = acpi_map_pxm_to_node(rhsa->proximity_domain);
node              478 drivers/iommu/dmar.c 			if (!node_online(node))
node              479 drivers/iommu/dmar.c 				node = NUMA_NO_NODE;
node              480 drivers/iommu/dmar.c 			drhd->iommu->node = node;
node             1064 drivers/iommu/dmar.c 	iommu->node = NUMA_NO_NODE;
node             1480 drivers/iommu/dmar.c 	desc_page = alloc_pages_node(iommu->node, GFP_ATOMIC | __GFP_ZERO,
node             1796 drivers/iommu/dmar.c 	irq = dmar_alloc_hwirq(iommu->seq_id, iommu->node, iommu);
node              512 drivers/iommu/fsl_pamu.c 	struct device_node *node;
node              519 drivers/iommu/fsl_pamu.c 		node = of_find_matching_node(NULL, l3_device_ids);
node              520 drivers/iommu/fsl_pamu.c 		if (node) {
node              521 drivers/iommu/fsl_pamu.c 			prop = of_get_property(node, "cache-stash-id", NULL);
node              524 drivers/iommu/fsl_pamu.c 					 node);
node              525 drivers/iommu/fsl_pamu.c 				of_node_put(node);
node              528 drivers/iommu/fsl_pamu.c 			of_node_put(node);
node              534 drivers/iommu/fsl_pamu.c 	for_each_of_cpu_node(node) {
node              535 drivers/iommu/fsl_pamu.c 		prop = of_get_property(node, "reg", &len);
node              548 drivers/iommu/fsl_pamu.c 			prop = of_get_property(node, "cache-stash-id", NULL);
node              551 drivers/iommu/fsl_pamu.c 					 node);
node              552 drivers/iommu/fsl_pamu.c 				of_node_put(node);
node              555 drivers/iommu/fsl_pamu.c 			of_node_put(node);
node              559 drivers/iommu/fsl_pamu.c 		prop = of_get_property(node, "next-level-cache", NULL);
node              561 drivers/iommu/fsl_pamu.c 			pr_debug("can't find next-level-cache at %pOF\n", node);
node              562 drivers/iommu/fsl_pamu.c 			of_node_put(node);
node              565 drivers/iommu/fsl_pamu.c 		of_node_put(node);
node              568 drivers/iommu/fsl_pamu.c 		node = of_find_node_by_phandle(*prop);
node              569 drivers/iommu/fsl_pamu.c 		if (!node) {
node              715 drivers/iommu/fsl_pamu.c 	struct device_node *node = NULL;
node              718 drivers/iommu/fsl_pamu.c 	for_each_node_with_property(node, "fsl,liodn") {
node              719 drivers/iommu/fsl_pamu.c 		prop = of_get_property(node, "fsl,liodn", &len);
node              738 drivers/iommu/fsl_pamu.c 			if (of_device_is_compatible(node, "fsl,qman-portal"))
node              740 drivers/iommu/fsl_pamu.c 			if (of_device_is_compatible(node, "fsl,qman"))
node              742 drivers/iommu/fsl_pamu.c 			if (of_device_is_compatible(node, "fsl,bman"))
node              514 drivers/iommu/intel-iommu.c void *alloc_pgtable_page(int node)
node              519 drivers/iommu/intel-iommu.c 	page = alloc_pages_node(node, GFP_ATOMIC | __GFP_ZERO, 0);
node              718 drivers/iommu/intel-iommu.c 		context = alloc_pgtable_page(iommu->node);
node             1195 drivers/iommu/intel-iommu.c 	root = (struct root_entry *)alloc_pgtable_page(iommu->node);
node             1756 drivers/iommu/intel-iommu.c 		domain->nid			 = iommu->node;
node             1894 drivers/iommu/intel-iommu.c 	domain->nid = iommu->node;
node             3042 drivers/iommu/intel-iommu.c 			new_ce = alloc_pgtable_page(iommu->node);
node             4891 drivers/iommu/intel-iommu.c 					    &adev->physical_node_list, node) {
node              155 drivers/iommu/intel-pasid.c 	pages = alloc_pages_node(info->iommu->node,
node              245 drivers/iommu/intel-pasid.c 		entries = alloc_pgtable_page(info->iommu->node);
node               54 drivers/iommu/intel-svm.c 	irq = dmar_alloc_hwirq(DMAR_UNITS_SUPPORTED + iommu->seq_id, iommu->node, iommu);
node              543 drivers/iommu/intel_irq_remapping.c 	pages = alloc_pages_node(iommu->node, GFP_KERNEL | __GFP_ZERO,
node               42 drivers/iommu/iova.c 	iovad->cached_node = &iovad->anchor.node;
node               43 drivers/iommu/iova.c 	iovad->cached32_node = &iovad->anchor.node;
node               51 drivers/iommu/iova.c 	rb_link_node(&iovad->anchor.node, NULL, &iovad->rbroot.rb_node);
node               52 drivers/iommu/iova.c 	rb_insert_color(&iovad->anchor.node, &iovad->rbroot);
node              129 drivers/iommu/iova.c 		iovad->cached32_node = &new->node;
node              131 drivers/iommu/iova.c 		iovad->cached_node = &new->node;
node              139 drivers/iommu/iova.c 	cached_iova = rb_entry(iovad->cached32_node, struct iova, node);
node              143 drivers/iommu/iova.c 		iovad->cached32_node = rb_next(&free->node);
node              147 drivers/iommu/iova.c 	cached_iova = rb_entry(iovad->cached_node, struct iova, node);
node              149 drivers/iommu/iova.c 		iovad->cached_node = rb_next(&free->node);
node              162 drivers/iommu/iova.c 		struct iova *this = rb_entry(*new, struct iova, node);
node              176 drivers/iommu/iova.c 	rb_link_node(&iova->node, parent, new);
node              177 drivers/iommu/iova.c 	rb_insert_color(&iova->node, root);
node              200 drivers/iommu/iova.c 	curr_iova = rb_entry(curr, struct iova, node);
node              206 drivers/iommu/iova.c 		curr_iova = rb_entry(curr, struct iova, node);
node              320 drivers/iommu/iova.c 	struct rb_node *node = iovad->rbroot.rb_node;
node              324 drivers/iommu/iova.c 	while (node) {
node              325 drivers/iommu/iova.c 		struct iova *iova = rb_entry(node, struct iova, node);
node              328 drivers/iommu/iova.c 			node = node->rb_left;
node              330 drivers/iommu/iova.c 			node = node->rb_right;
node              342 drivers/iommu/iova.c 	rb_erase(&iova->node, &iovad->rbroot);
node              599 drivers/iommu/iova.c 	rbtree_postorder_for_each_entry_safe(iova, tmp, &iovad->rbroot, node)
node              605 drivers/iommu/iova.c __is_range_overlap(struct rb_node *node,
node              608 drivers/iommu/iova.c 	struct iova *iova = rb_entry(node, struct iova, node);
node              664 drivers/iommu/iova.c 	struct rb_node *node;
node              674 drivers/iommu/iova.c 	for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) {
node              675 drivers/iommu/iova.c 		if (__is_range_overlap(node, pfn_lo, pfn_hi)) {
node              676 drivers/iommu/iova.c 			iova = rb_entry(node, struct iova, node);
node              709 drivers/iommu/iova.c 	struct rb_node *node;
node              712 drivers/iommu/iova.c 	for (node = rb_first(&from->rbroot); node; node = rb_next(node)) {
node              713 drivers/iommu/iova.c 		struct iova *iova = rb_entry(node, struct iova, node);
node              748 drivers/iommu/iova.c 	rb_erase(&iova->node, &iovad->rbroot);
node              433 drivers/iommu/mtk_iommu_v1.c 		iommu_spec.np = of_node_get(it.node);
node              599 drivers/iommu/mtk_iommu_v1.c 		larb_spec.np = of_node_get(it.node);
node               40 drivers/iommu/omap-iommu.c 	struct list_head node;
node             1263 drivers/iommu/omap-iommu.c 	list_for_each_entry_safe(orphan_dev, tmp, &orphan_dev_list, node) {
node             1266 drivers/iommu/omap-iommu.c 			list_del(&orphan_dev->node);
node             1707 drivers/iommu/omap-iommu.c 					    node) {
node             1719 drivers/iommu/omap-iommu.c 			list_add(&orphan_dev->node, &orphan_dev_list);
node              108 drivers/iommu/rockchip-iommu.c 	struct list_head node; /* entry in rk_iommu_domain.iommus */
node              629 drivers/iommu/rockchip-iommu.c 		iommu = list_entry(pos, struct rk_iommu, node);
node              918 drivers/iommu/rockchip-iommu.c 	list_del_init(&iommu->node);
node              957 drivers/iommu/rockchip-iommu.c 	list_add_tail(&iommu->node, &rk_domain->iommus);
node              353 drivers/iommu/virtio-iommu.c 	struct interval_tree_node *node, *next;
node              358 drivers/iommu/virtio-iommu.c 		node = next;
node              359 drivers/iommu/virtio-iommu.c 		mapping = container_of(node, struct viommu_mapping, iova);
node              360 drivers/iommu/virtio-iommu.c 		next = interval_tree_iter_next(node, iova, last);
node              372 drivers/iommu/virtio-iommu.c 		interval_tree_remove(node, &vdomain->mappings);
node              392 drivers/iommu/virtio-iommu.c 	struct interval_tree_node *node;
node              396 drivers/iommu/virtio-iommu.c 	node = interval_tree_iter_first(&vdomain->mappings, 0, -1UL);
node              397 drivers/iommu/virtio-iommu.c 	while (node) {
node              398 drivers/iommu/virtio-iommu.c 		mapping = container_of(node, struct viommu_mapping, iova);
node              412 drivers/iommu/virtio-iommu.c 		node = interval_tree_iter_next(node, 0, -1UL);
node              788 drivers/iommu/virtio-iommu.c 	struct interval_tree_node *node;
node              792 drivers/iommu/virtio-iommu.c 	node = interval_tree_iter_first(&vdomain->mappings, iova, iova);
node              793 drivers/iommu/virtio-iommu.c 	if (node) {
node              794 drivers/iommu/virtio-iommu.c 		mapping = container_of(node, struct viommu_mapping, iova);
node              140 drivers/irqchip/irq-al-fic.c static int al_fic_register(struct device_node *node,
node              146 drivers/irqchip/irq-al-fic.c 	fic->domain = irq_domain_add_linear(node,
node              200 drivers/irqchip/irq-al-fic.c static struct al_fic *al_fic_wire_init(struct device_node *node,
node              225 drivers/irqchip/irq-al-fic.c 	ret = al_fic_register(node, fic);
node              241 drivers/irqchip/irq-al-fic.c static int __init al_fic_init_dt(struct device_node *node,
node              251 drivers/irqchip/irq-al-fic.c 		       node->name);
node              255 drivers/irqchip/irq-al-fic.c 	base = of_iomap(node, 0);
node              257 drivers/irqchip/irq-al-fic.c 		pr_err("%s: fail to map memory\n", node->name);
node              261 drivers/irqchip/irq-al-fic.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              263 drivers/irqchip/irq-al-fic.c 		pr_err("%s: fail to map irq\n", node->name);
node              268 drivers/irqchip/irq-al-fic.c 	fic = al_fic_wire_init(node,
node              270 drivers/irqchip/irq-al-fic.c 			       node->name,
node              274 drivers/irqchip/irq-al-fic.c 		       node->name,
node              191 drivers/irqchip/irq-alpine-msi.c 				    struct device_node *node)
node              196 drivers/irqchip/irq-alpine-msi.c 	gic_node = of_irq_find_parent(node);
node              218 drivers/irqchip/irq-alpine-msi.c 	msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(node),
node              230 drivers/irqchip/irq-alpine-msi.c static int alpine_msix_init(struct device_node *node,
node              243 drivers/irqchip/irq-alpine-msi.c 	ret = of_address_to_resource(node, 0, &res);
node              259 drivers/irqchip/irq-alpine-msi.c 	if (of_property_read_u32(node, "al,msi-base-spi", &priv->spi_first)) {
node              265 drivers/irqchip/irq-alpine-msi.c 	if (of_property_read_u32(node, "al,msi-num-spis", &priv->num_spis)) {
node              282 drivers/irqchip/irq-alpine-msi.c 	ret = alpine_msix_init_domains(priv, node);
node              271 drivers/irqchip/irq-armada-370-xp.c static int armada_370_xp_msi_init(struct device_node *node,
node              286 drivers/irqchip/irq-armada-370-xp.c 		pci_msi_create_irq_domain(of_node_to_fwnode(node),
node              306 drivers/irqchip/irq-armada-370-xp.c static inline int armada_370_xp_msi_init(struct device_node *node,
node              646 drivers/irqchip/irq-armada-370-xp.c static int __init armada_370_xp_mpic_of_init(struct device_node *node,
node              653 drivers/irqchip/irq-armada-370-xp.c 	BUG_ON(of_address_to_resource(node, 0, &main_int_res));
node              654 drivers/irqchip/irq-armada-370-xp.c 	BUG_ON(of_address_to_resource(node, 1, &per_cpu_int_res));
node              658 drivers/irqchip/irq-armada-370-xp.c 				   node->full_name));
node              661 drivers/irqchip/irq-armada-370-xp.c 				   node->full_name));
node              678 drivers/irqchip/irq-armada-370-xp.c 		irq_domain_add_linear(node, nr_irqs,
node              687 drivers/irqchip/irq-armada-370-xp.c 	armada_370_xp_msi_init(node, main_int_res.start);
node              689 drivers/irqchip/irq-armada-370-xp.c 	parent_irq = irq_of_parse_and_map(node, 0);
node               65 drivers/irqchip/irq-aspeed-i2c-ic.c static int __init aspeed_i2c_ic_of_init(struct device_node *node,
node               75 drivers/irqchip/irq-aspeed-i2c-ic.c 	i2c_ic->base = of_iomap(node, 0);
node               81 drivers/irqchip/irq-aspeed-i2c-ic.c 	i2c_ic->parent_irq = irq_of_parse_and_map(node, 0);
node               87 drivers/irqchip/irq-aspeed-i2c-ic.c 	i2c_ic->irq_domain = irq_domain_add_linear(node, ASPEED_I2C_IC_NUM_BUS,
node              184 drivers/irqchip/irq-aspeed-vic.c static int __init avic_of_init(struct device_node *node,
node              195 drivers/irqchip/irq-aspeed-vic.c 	regs = of_iomap(node, 0);
node              214 drivers/irqchip/irq-aspeed-vic.c 	vic->dom = irq_domain_add_simple(node, NUM_IRQS, 0,
node               57 drivers/irqchip/irq-ath79-cpu.c 	struct device_node *node, struct device_node *parent)
node               63 drivers/irqchip/irq-ath79-cpu.c 		node, "qca,ddr-wb-channels", "#qca,ddr-wb-channel-cells");
node               70 drivers/irqchip/irq-ath79-cpu.c 			node, "qca,ddr-wb-channel-interrupts", i, &irq);
node               75 drivers/irqchip/irq-ath79-cpu.c 			node, "qca,ddr-wb-channels",
node               84 drivers/irqchip/irq-ath79-cpu.c 	return mips_cpu_irq_of_init(node, parent);
node              132 drivers/irqchip/irq-ath79-misc.c 	struct device_node *node, struct device_node *parent)
node              138 drivers/irqchip/irq-ath79-misc.c 	irq = irq_of_parse_and_map(node, 0);
node              144 drivers/irqchip/irq-ath79-misc.c 	base = of_iomap(node, 0);
node              150 drivers/irqchip/irq-ath79-misc.c 	domain = irq_domain_add_linear(node, ATH79_MISC_IRQ_COUNT,
node              162 drivers/irqchip/irq-ath79-misc.c 	struct device_node *node, struct device_node *parent)
node              165 drivers/irqchip/irq-ath79-misc.c 	return ath79_misc_intc_of_init(node, parent);
node              172 drivers/irqchip/irq-ath79-misc.c 	struct device_node *node, struct device_node *parent)
node              175 drivers/irqchip/irq-ath79-misc.c 	return ath79_misc_intc_of_init(node, parent);
node              112 drivers/irqchip/irq-ativic32.c int __init ativic32_init_irq(struct device_node *node, struct device_node *parent)
node              130 drivers/irqchip/irq-ativic32.c 	root_domain = irq_domain_add_linear(node, nr_ints,
node              134 drivers/irqchip/irq-ativic32.c 		panic("%s: unable to create IRQ domain\n", node->full_name);
node              111 drivers/irqchip/irq-atmel-aic-common.c 	struct device_node *node = irq_domain_get_of_node(domain);
node              123 drivers/irqchip/irq-atmel-aic-common.c 	of_property_for_each_u32(node, "atmel,external-irqs", prop, p, hwirq) {
node              208 drivers/irqchip/irq-atmel-aic-common.c struct irq_domain *__init aic_common_of_init(struct device_node *node,
node              223 drivers/irqchip/irq-atmel-aic-common.c 	reg_base = of_iomap(node, 0);
node              233 drivers/irqchip/irq-atmel-aic-common.c 	domain = irq_domain_add_linear(node, nchips * 32, ops, aic);
node               31 drivers/irqchip/irq-atmel-aic-common.h struct irq_domain *__init aic_common_of_init(struct device_node *node,
node              241 drivers/irqchip/irq-atmel-aic.c static int __init aic_of_init(struct device_node *node,
node              250 drivers/irqchip/irq-atmel-aic.c 	domain = aic_common_of_init(node, &aic_irq_ops, "atmel-aic",
node              320 drivers/irqchip/irq-atmel-aic5.c static int __init aic5_of_init(struct device_node *node,
node              335 drivers/irqchip/irq-atmel-aic5.c 	domain = aic_common_of_init(node, &aic5_irq_ops, "atmel-aic5",
node              363 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d2_aic5_of_init(struct device_node *node,
node              373 drivers/irqchip/irq-atmel-aic5.c 	return aic5_of_init(node, parent, NR_SAMA5D2_IRQS);
node              379 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d3_aic5_of_init(struct device_node *node,
node              382 drivers/irqchip/irq-atmel-aic5.c 	return aic5_of_init(node, parent, NR_SAMA5D3_IRQS);
node              388 drivers/irqchip/irq-atmel-aic5.c static int __init sama5d4_aic5_of_init(struct device_node *node,
node              391 drivers/irqchip/irq-atmel-aic5.c 	return aic5_of_init(node, parent, NR_SAMA5D4_IRQS);
node              397 drivers/irqchip/irq-atmel-aic5.c static int __init sam9x60_aic5_of_init(struct device_node *node,
node              400 drivers/irqchip/irq-atmel-aic5.c 	return aic5_of_init(node, parent, NR_SAM9X60_IRQS);
node              132 drivers/irqchip/irq-bcm2835.c static int __init armctrl_of_init(struct device_node *node,
node              139 drivers/irqchip/irq-bcm2835.c 	base = of_iomap(node, 0);
node              141 drivers/irqchip/irq-bcm2835.c 		panic("%pOF: unable to map IC registers\n", node);
node              143 drivers/irqchip/irq-bcm2835.c 	intc.domain = irq_domain_add_linear(node, MAKE_HWIRQ(NR_BANKS, 0),
node              146 drivers/irqchip/irq-bcm2835.c 		panic("%pOF: unable to create IRQ domain\n", node);
node              163 drivers/irqchip/irq-bcm2835.c 		int parent_irq = irq_of_parse_and_map(node, 0);
node              167 drivers/irqchip/irq-bcm2835.c 			      node);
node              177 drivers/irqchip/irq-bcm2835.c static int __init bcm2835_armctrl_of_init(struct device_node *node,
node              180 drivers/irqchip/irq-bcm2835.c 	return armctrl_of_init(node, parent, false);
node              183 drivers/irqchip/irq-bcm2835.c static int __init bcm2836_armctrl_of_init(struct device_node *node,
node              186 drivers/irqchip/irq-bcm2835.c 	return armctrl_of_init(node, parent, true);
node              219 drivers/irqchip/irq-bcm2836.c static int __init bcm2836_arm_irqchip_l1_intc_of_init(struct device_node *node,
node              222 drivers/irqchip/irq-bcm2836.c 	intc.base = of_iomap(node, 0);
node              224 drivers/irqchip/irq-bcm2836.c 		panic("%pOF: unable to map local interrupt registers\n", node);
node              229 drivers/irqchip/irq-bcm2836.c 	intc.domain = irq_domain_add_linear(node, LAST_IRQ + 1,
node              233 drivers/irqchip/irq-bcm2836.c 		panic("%pOF: unable to create IRQ domain\n", node);
node              195 drivers/irqchip/irq-crossbar.c static int __init crossbar_of_init(struct device_node *node)
node              207 drivers/irqchip/irq-crossbar.c 	cb->crossbar_base = of_iomap(node, 0);
node              211 drivers/irqchip/irq-crossbar.c 	of_property_read_u32(node, "ti,max-crossbar-sources",
node              219 drivers/irqchip/irq-crossbar.c 	of_property_read_u32(node, "ti,max-irqs", &max);
node              235 drivers/irqchip/irq-crossbar.c 	irqsr = of_get_property(node, "ti,irqs-reserved", &size);
node              240 drivers/irqchip/irq-crossbar.c 			of_property_read_u32_index(node,
node              253 drivers/irqchip/irq-crossbar.c 	irqsr = of_get_property(node, "ti,irqs-skip", &size);
node              258 drivers/irqchip/irq-crossbar.c 			of_property_read_u32_index(node,
node              275 drivers/irqchip/irq-crossbar.c 	of_property_read_u32(node, "ti,reg-size", &reg_size);
node              306 drivers/irqchip/irq-crossbar.c 	of_property_read_u32(node, "ti,irqs-safe-map", &cb->safe_map);
node              333 drivers/irqchip/irq-crossbar.c static int __init irqcrossbar_init(struct device_node *node,
node              340 drivers/irqchip/irq-crossbar.c 		pr_err("%pOF: no parent, giving up\n", node);
node              346 drivers/irqchip/irq-crossbar.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              350 drivers/irqchip/irq-crossbar.c 	err = crossbar_of_init(node);
node              356 drivers/irqchip/irq-crossbar.c 					  node, &crossbar_domain_ops,
node              359 drivers/irqchip/irq-crossbar.c 		pr_err("%pOF: failed to allocated domain\n", node);
node               60 drivers/irqchip/irq-csky-apb-intc.c static void __init ck_set_gc(struct device_node *node, void __iomem *reg_base,
node               71 drivers/irqchip/irq-csky-apb-intc.c 	if (of_find_property(node, "csky,support-pulse-signal", NULL))
node              102 drivers/irqchip/irq-csky-apb-intc.c ck_intc_init_comm(struct device_node *node, struct device_node *parent)
node              111 drivers/irqchip/irq-csky-apb-intc.c 	reg_base = of_iomap(node, 0);
node              113 drivers/irqchip/irq-csky-apb-intc.c 		pr_err("C-SKY Intc unable to map: %p.\n", node);
node              117 drivers/irqchip/irq-csky-apb-intc.c 	root_domain = irq_domain_add_linear(node, nr_irq,
node              164 drivers/irqchip/irq-csky-apb-intc.c gx_intc_init(struct device_node *node, struct device_node *parent)
node              168 drivers/irqchip/irq-csky-apb-intc.c 	ret = ck_intc_init_comm(node, parent);
node              186 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base, GX_INTC_NEN31_00, 0);
node              187 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base, GX_INTC_NEN63_32, 32);
node              231 drivers/irqchip/irq-csky-apb-intc.c ck_intc_init(struct device_node *node, struct device_node *parent)
node              235 drivers/irqchip/irq-csky-apb-intc.c 	ret = ck_intc_init_comm(node, parent);
node              246 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base, CK_INTC_NEN31_00, 0);
node              247 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base, CK_INTC_NEN63_32, 32);
node              258 drivers/irqchip/irq-csky-apb-intc.c ck_dual_intc_init(struct device_node *node, struct device_node *parent)
node              265 drivers/irqchip/irq-csky-apb-intc.c 	ret = ck_intc_init(node, parent);
node              273 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base + CK_INTC_DUAL_BASE, CK_INTC_NEN31_00, 64);
node              274 drivers/irqchip/irq-csky-apb-intc.c 	ck_set_gc(node, reg_base + CK_INTC_DUAL_BASE, CK_INTC_NEN63_32, 96);
node              228 drivers/irqchip/irq-csky-mpintc.c csky_mpintc_init(struct device_node *node, struct device_node *parent)
node              239 drivers/irqchip/irq-csky-mpintc.c 	ret = of_property_read_u32(node, "csky,num-irqs", &nr_irq);
node              258 drivers/irqchip/irq-csky-mpintc.c 	root_domain = irq_domain_add_linear(node, nr_irq, &csky_irqdomain_ops,
node              160 drivers/irqchip/irq-davinci-cp-intc.c 			struct device_node *node)
node              217 drivers/irqchip/irq-davinci-cp-intc.c 					node, config->num_irqs, irq_base, 0,
node              238 drivers/irqchip/irq-davinci-cp-intc.c static int __init davinci_cp_intc_of_init(struct device_node *node,
node              244 drivers/irqchip/irq-davinci-cp-intc.c 	ret = of_address_to_resource(node, 0, &config.reg);
node              251 drivers/irqchip/irq-davinci-cp-intc.c 	ret = of_property_read_u32(node, "ti,intc-size", &config.num_irqs);
node              258 drivers/irqchip/irq-davinci-cp-intc.c 	return davinci_cp_intc_do_init(&config, node);
node               71 drivers/irqchip/irq-digicolor.c static int __init digicolor_of_init(struct device_node *node,
node               79 drivers/irqchip/irq-digicolor.c 	reg_base = of_iomap(node, 0);
node               81 drivers/irqchip/irq-digicolor.c 		pr_err("%pOF: unable to map IC registers\n", node);
node               89 drivers/irqchip/irq-digicolor.c 	ucregs = syscon_regmap_lookup_by_phandle(node, "syscon");
node               91 drivers/irqchip/irq-digicolor.c 		pr_err("%pOF: unable to map UC registers\n", node);
node               98 drivers/irqchip/irq-digicolor.c 		irq_domain_add_linear(node, 64, &irq_generic_chip_ops, NULL);
node              100 drivers/irqchip/irq-digicolor.c 		pr_err("%pOF: unable to create IRQ domain\n", node);
node              108 drivers/irqchip/irq-digicolor.c 		pr_err("%pOF: unable to allocate IRQ gc\n", node);
node              135 drivers/irqchip/irq-eznps.c static int __init nps400_of_init(struct device_node *node,
node              145 drivers/irqchip/irq-eznps.c 	nps400_root_domain = irq_domain_add_linear(node, NPS_NR_CPU_IRQS,
node              166 drivers/irqchip/irq-ftintc010.c int __init ft010_of_init_irq(struct device_node *node,
node              177 drivers/irqchip/irq-ftintc010.c 	f->base = of_iomap(node, 0);
node              184 drivers/irqchip/irq-ftintc010.c 	f->domain = irq_domain_add_simple(node, FT010_NUM_IRQS, 0,
node               45 drivers/irqchip/irq-gic-realview.c realview_gic_of_init(struct device_node *node, struct device_node *parent)
node               73 drivers/irqchip/irq-gic-realview.c 	return gic_of_init(node, parent);
node              419 drivers/irqchip/irq-gic-v2m.c 	struct device_node *node = to_of_node(parent_handle);
node              422 drivers/irqchip/irq-gic-v2m.c 	for (child = of_find_matching_node(node, gicv2m_device_id); child;
node             3806 drivers/irqchip/irq-gic-v3-its.c static int __init its_of_probe(struct device_node *node)
node             3811 drivers/irqchip/irq-gic-v3-its.c 	for (np = of_find_matching_node(node, its_device_id); np;
node             3866 drivers/irqchip/irq-gic-v3-its.c 	int node;
node             3879 drivers/irqchip/irq-gic-v3-its.c 	node = acpi_map_pxm_to_node(its_affinity->proximity_domain);
node             3881 drivers/irqchip/irq-gic-v3-its.c 	if (node == NUMA_NO_NODE || node >= MAX_NUMNODES) {
node             3882 drivers/irqchip/irq-gic-v3-its.c 		pr_err("SRAT: Invalid NUMA node %d in ITS affinity\n", node);
node             3886 drivers/irqchip/irq-gic-v3-its.c 	its_srat_maps[its_in_srat].numa_node = node;
node             3890 drivers/irqchip/irq-gic-v3-its.c 		its_affinity->proximity_domain, its_affinity->its_id, node);
node             1700 drivers/irqchip/irq-gic-v3.c static void __init gic_of_setup_kvm_info(struct device_node *node)
node             1708 drivers/irqchip/irq-gic-v3.c 	gic_v3_kvm_info.maint_irq = irq_of_parse_and_map(node, 0);
node             1712 drivers/irqchip/irq-gic-v3.c 	if (of_property_read_u32(node, "#redistributor-regions",
node             1717 drivers/irqchip/irq-gic-v3.c 	ret = of_address_to_resource(node, gicv_idx, &r);
node             1725 drivers/irqchip/irq-gic-v3.c static int __init gic_of_init(struct device_node *node, struct device_node *parent)
node             1733 drivers/irqchip/irq-gic-v3.c 	dist_base = of_iomap(node, 0);
node             1735 drivers/irqchip/irq-gic-v3.c 		pr_err("%pOF: unable to map gic dist registers\n", node);
node             1741 drivers/irqchip/irq-gic-v3.c 		pr_err("%pOF: no distributor detected, giving up\n", node);
node             1745 drivers/irqchip/irq-gic-v3.c 	if (of_property_read_u32(node, "#redistributor-regions", &nr_redist_regions))
node             1759 drivers/irqchip/irq-gic-v3.c 		ret = of_address_to_resource(node, 1 + i, &res);
node             1760 drivers/irqchip/irq-gic-v3.c 		rdist_regs[i].redist_base = of_iomap(node, 1 + i);
node             1762 drivers/irqchip/irq-gic-v3.c 			pr_err("%pOF: couldn't map region %d\n", node, i);
node             1769 drivers/irqchip/irq-gic-v3.c 	if (of_property_read_u64(node, "redistributor-stride", &redist_stride))
node             1772 drivers/irqchip/irq-gic-v3.c 	gic_enable_of_quirks(node, gic_quirks, &gic_data);
node             1775 drivers/irqchip/irq-gic-v3.c 			     redist_stride, &node->fwnode);
node             1779 drivers/irqchip/irq-gic-v3.c 	gic_populate_ppi_partitions(node);
node             1782 drivers/irqchip/irq-gic-v3.c 		gic_of_setup_kvm_info(node);
node              963 drivers/irqchip/irq-gic.c static void __init gic_init_physaddr(struct device_node *node)
node              966 drivers/irqchip/irq-gic.c 	if (of_address_to_resource(node, 0, &res) == 0) {
node              973 drivers/irqchip/irq-gic.c #define gic_init_physaddr(node)  do { } while (0)
node             1275 drivers/irqchip/irq-gic.c static bool gic_check_eoimode(struct device_node *node, void __iomem **base)
node             1279 drivers/irqchip/irq-gic.c 	of_address_to_resource(node, 1, &cpuif_res);
node             1353 drivers/irqchip/irq-gic.c static int gic_of_setup(struct gic_chip_data *gic, struct device_node *node)
node             1355 drivers/irqchip/irq-gic.c 	if (!gic || !node)
node             1358 drivers/irqchip/irq-gic.c 	gic->raw_dist_base = of_iomap(node, 0);
node             1362 drivers/irqchip/irq-gic.c 	gic->raw_cpu_base = of_iomap(node, 1);
node             1366 drivers/irqchip/irq-gic.c 	if (of_property_read_u32(node, "cpu-offset", &gic->percpu_offset))
node             1405 drivers/irqchip/irq-gic.c static void __init gic_of_setup_kvm_info(struct device_node *node)
node             1413 drivers/irqchip/irq-gic.c 	gic_v2_kvm_info.maint_irq = irq_of_parse_and_map(node, 0);
node             1417 drivers/irqchip/irq-gic.c 	ret = of_address_to_resource(node, 2, vctrl_res);
node             1421 drivers/irqchip/irq-gic.c 	ret = of_address_to_resource(node, 3, vcpu_res);
node             1430 drivers/irqchip/irq-gic.c gic_of_init(struct device_node *node, struct device_node *parent)
node             1435 drivers/irqchip/irq-gic.c 	if (WARN_ON(!node))
node             1443 drivers/irqchip/irq-gic.c 	ret = gic_of_setup(gic, node);
node             1451 drivers/irqchip/irq-gic.c 	if (gic_cnt == 0 && !gic_check_eoimode(node, &gic->raw_cpu_base))
node             1454 drivers/irqchip/irq-gic.c 	ret = __gic_init_bases(gic, &node->fwnode);
node             1461 drivers/irqchip/irq-gic.c 		gic_init_physaddr(node);
node             1462 drivers/irqchip/irq-gic.c 		gic_of_setup_kvm_info(node);
node             1466 drivers/irqchip/irq-gic.c 		irq = irq_of_parse_and_map(node, 0);
node             1471 drivers/irqchip/irq-gic.c 		gicv2m_init(&node->fwnode, gic_data[gic_cnt].domain);
node              362 drivers/irqchip/irq-hip04.c hip04_of_init(struct device_node *node, struct device_node *parent)
node              367 drivers/irqchip/irq-hip04.c 	if (WARN_ON(!node))
node              370 drivers/irqchip/irq-hip04.c 	hip04_data.dist_base = of_iomap(node, 0);
node              373 drivers/irqchip/irq-hip04.c 	hip04_data.cpu_base = of_iomap(node, 1);
node              401 drivers/irqchip/irq-hip04.c 	hip04_data.domain = irq_domain_add_legacy(node, nr_irqs, irq_base,
node              312 drivers/irqchip/irq-i8259.c struct irq_domain * __init __init_i8259_irqs(struct device_node *node)
node              321 drivers/irqchip/irq-i8259.c 	domain = irq_domain_add_legacy(node, 16, I8259A_IRQ_BASE, 0,
node              349 drivers/irqchip/irq-i8259.c int __init i8259_of_init(struct device_node *node, struct device_node *parent)
node              354 drivers/irqchip/irq-i8259.c 	domain = __init_i8259_irqs(node);
node              356 drivers/irqchip/irq-i8259.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              300 drivers/irqchip/irq-imgpdc.c 	struct device_node *node = pdev->dev.of_node;
node              307 drivers/irqchip/irq-imgpdc.c 	if (!node)
node              333 drivers/irqchip/irq-imgpdc.c 	ret = of_property_read_u32(node, "num-perips", &val);
node              345 drivers/irqchip/irq-imgpdc.c 	ret = of_property_read_u32(node, "num-syswakes", &val);
node              382 drivers/irqchip/irq-imgpdc.c 	priv->domain = irq_domain_add_linear(node, 16, &irq_generic_chip_ops,
node              202 drivers/irqchip/irq-imx-gpcv2.c static int __init imx_gpcv2_irqchip_init(struct device_node *node,
node              212 drivers/irqchip/irq-imx-gpcv2.c 		pr_err("%pOF: no parent, giving up\n", node);
node              216 drivers/irqchip/irq-imx-gpcv2.c 	id = of_match_node(gpcv2_of_match, node);
node              218 drivers/irqchip/irq-imx-gpcv2.c 		pr_err("%pOF: unknown compatibility string\n", node);
node              226 drivers/irqchip/irq-imx-gpcv2.c 		pr_err("%pOF: unable to get parent domain\n", node);
node              232 drivers/irqchip/irq-imx-gpcv2.c 		pr_err("%pOF: kzalloc failed!\n", node);
node              238 drivers/irqchip/irq-imx-gpcv2.c 	cd->gpc_base = of_iomap(node, 0);
node              240 drivers/irqchip/irq-imx-gpcv2.c 		pr_err("%pOF: unable to map gpc registers\n", node);
node              246 drivers/irqchip/irq-imx-gpcv2.c 				node, &gpcv2_irqchip_data_domain_ops, cd);
node              287 drivers/irqchip/irq-imx-gpcv2.c 	of_node_clear_flag(node, OF_POPULATED);
node               78 drivers/irqchip/irq-ingenic.c static int __init ingenic_intc_of_init(struct device_node *node,
node               94 drivers/irqchip/irq-ingenic.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              105 drivers/irqchip/irq-ingenic.c 	intc->base = of_iomap(node, 0);
node              111 drivers/irqchip/irq-ingenic.c 	domain = irq_domain_add_legacy(node, num_chips * 32,
node              158 drivers/irqchip/irq-ingenic.c static int __init intc_1chip_of_init(struct device_node *node,
node              161 drivers/irqchip/irq-ingenic.c 	return ingenic_intc_of_init(node, 1);
node              166 drivers/irqchip/irq-ingenic.c static int __init intc_2chip_of_init(struct device_node *node,
node              169 drivers/irqchip/irq-ingenic.c 	return ingenic_intc_of_init(node, 2);
node               65 drivers/irqchip/irq-jcore-aic.c static int __init aic_irq_of_init(struct device_node *node,
node               75 drivers/irqchip/irq-jcore-aic.c 	if (of_device_is_compatible(node, "jcore,aic1")) {
node               79 drivers/irqchip/irq-jcore-aic.c 			void __iomem *base = of_iomap(node, cpu);
node              103 drivers/irqchip/irq-jcore-aic.c 	domain = irq_domain_add_linear(node, dom_sz, &jcore_aic_irqdomain_ops,
node              174 drivers/irqchip/irq-lpc32xx.c static int __init lpc32xx_of_ic_init(struct device_node *node,
node              178 drivers/irqchip/irq-lpc32xx.c 	bool is_mic = of_device_is_compatible(node, "nxp,lpc3220-mic");
node              179 drivers/irqchip/irq-lpc32xx.c 	const __be32 *reg = of_get_property(node, "reg", NULL);
node              186 drivers/irqchip/irq-lpc32xx.c 	irqc->base = of_iomap(node, 0);
node              188 drivers/irqchip/irq-lpc32xx.c 		pr_err("%pOF: unable to map registers\n", node);
node              202 drivers/irqchip/irq-lpc32xx.c 	irqc->domain = irq_domain_add_linear(node, NR_LPC32XX_IC_IRQS,
node              216 drivers/irqchip/irq-lpc32xx.c 		for (i = 0; i < of_irq_count(node); i++) {
node              217 drivers/irqchip/irq-lpc32xx.c 			parent_irq = irq_of_parse_and_map(node, i);
node              103 drivers/irqchip/irq-ls1x.c static int __init ls1x_intc_of_init(struct device_node *node,
node              115 drivers/irqchip/irq-ls1x.c 	priv->intc_base = of_iomap(node, 0);
node              121 drivers/irqchip/irq-ls1x.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              129 drivers/irqchip/irq-ls1x.c 	priv->domain = irq_domain_add_linear(node, 32, &irq_generic_chip_ops,
node              138 drivers/irqchip/irq-ls1x.c 		node->full_name, handle_level_irq,
node              352 drivers/irqchip/irq-meson-gpio.c static int __init meson_gpio_irq_parse_dt(struct device_node *node,
node              358 drivers/irqchip/irq-meson-gpio.c 	match = of_match_node(meson_irq_gpio_matches, node);
node              364 drivers/irqchip/irq-meson-gpio.c 	ret = of_property_read_variable_u32_array(node,
node              377 drivers/irqchip/irq-meson-gpio.c static int __init meson_gpio_irq_of_init(struct device_node *node,
node              401 drivers/irqchip/irq-meson-gpio.c 	ctl->base = of_iomap(node, 0);
node              407 drivers/irqchip/irq-meson-gpio.c 	ret = meson_gpio_irq_parse_dt(node, ctl);
node              413 drivers/irqchip/irq-meson-gpio.c 					     of_node_to_fwnode(node),
node              208 drivers/irqchip/irq-mips-cpu.c static int mips_cpu_ipi_match(struct irq_domain *d, struct device_node *node,
node              216 drivers/irqchip/irq-mips-cpu.c 		return (!node || (to_of_node(d->fwnode) == node)) && is_ipi;
node              634 drivers/irqchip/irq-mips-gic.c int gic_ipi_domain_match(struct irq_domain *d, struct device_node *node,
node              642 drivers/irqchip/irq-mips-gic.c 		return (!node || to_of_node(d->fwnode) == node) && is_ipi;
node              671 drivers/irqchip/irq-mips-gic.c static int __init gic_of_init(struct device_node *node,
node              683 drivers/irqchip/irq-mips-gic.c 	while (!of_property_read_u32_index(node, "mti,reserved-cpu-vectors",
node              693 drivers/irqchip/irq-mips-gic.c 	if (of_address_to_resource(node, 0, &res)) {
node              759 drivers/irqchip/irq-mips-gic.c 	gic_irq_domain = irq_domain_add_simple(node, GIC_NUM_LOCAL_INTRS +
node              770 drivers/irqchip/irq-mips-gic.c 						  node, &gic_ipi_domain_ops, NULL);
node              778 drivers/irqchip/irq-mips-gic.c 	if (node &&
node              779 drivers/irqchip/irq-mips-gic.c 	    !of_property_read_u32_array(node, "mti,reserved-ipi-vectors", v, 2)) {
node              190 drivers/irqchip/irq-mmp.c static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
node              373 drivers/irqchip/irq-mmp.c static int __init mmp_init_bases(struct device_node *node)
node              377 drivers/irqchip/irq-mmp.c 	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
node              383 drivers/irqchip/irq-mmp.c 	mmp_icu_base = of_iomap(node, 0);
node              390 drivers/irqchip/irq-mmp.c 	icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
node              414 drivers/irqchip/irq-mmp.c static int __init mmp_of_init(struct device_node *node,
node              419 drivers/irqchip/irq-mmp.c 	ret = mmp_init_bases(node);
node              432 drivers/irqchip/irq-mmp.c static int __init mmp2_of_init(struct device_node *node,
node              437 drivers/irqchip/irq-mmp.c 	ret = mmp_init_bases(node);
node              450 drivers/irqchip/irq-mmp.c static int __init mmp3_of_init(struct device_node *node,
node              455 drivers/irqchip/irq-mmp.c 	mmp_icu2_base = of_iomap(node, 1);
node              461 drivers/irqchip/irq-mmp.c 	ret = mmp_init_bases(node);
node              482 drivers/irqchip/irq-mmp.c static int __init mmp2_mux_of_init(struct device_node *node,
node              493 drivers/irqchip/irq-mmp.c 	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
node              506 drivers/irqchip/irq-mmp.c 	ret = of_property_read_variable_u32_array(node, "reg", reg,
node              515 drivers/irqchip/irq-mmp.c 	icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
node              520 drivers/irqchip/irq-mmp.c 	icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
node              533 drivers/irqchip/irq-mmp.c 	if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
node               61 drivers/irqchip/irq-mscc-ocelot.c static int __init ocelot_irq_init(struct device_node *node,
node               68 drivers/irqchip/irq-mscc-ocelot.c 	parent_irq = irq_of_parse_and_map(node, 0);
node               72 drivers/irqchip/irq-mscc-ocelot.c 	domain = irq_domain_add_linear(node, OCELOT_NR_IRQ,
node               75 drivers/irqchip/irq-mscc-ocelot.c 		pr_err("%pOFn: unable to add irq domain\n", node);
node               83 drivers/irqchip/irq-mscc-ocelot.c 		pr_err("%pOFn: unable to alloc irq domain gc\n", node);
node               88 drivers/irqchip/irq-mscc-ocelot.c 	gc->reg_base = of_iomap(node, 0);
node               90 drivers/irqchip/irq-mscc-ocelot.c 		pr_err("%pOFn: unable to map resource\n", node);
node              243 drivers/irqchip/irq-mtk-cirq.c static int __init mtk_cirq_of_init(struct device_node *node,
node              260 drivers/irqchip/irq-mtk-cirq.c 	cirq_data->base = of_iomap(node, 0);
node              267 drivers/irqchip/irq-mtk-cirq.c 	ret = of_property_read_u32_index(node, "mediatek,ext-irq-range", 0,
node              272 drivers/irqchip/irq-mtk-cirq.c 	ret = of_property_read_u32_index(node, "mediatek,ext-irq-range", 1,
node              279 drivers/irqchip/irq-mtk-cirq.c 					  irq_num, node,
node              122 drivers/irqchip/irq-mtk-sysirq.c static int __init mtk_sysirq_of_init(struct device_node *node,
node              139 drivers/irqchip/irq-mtk-sysirq.c 	while (of_get_address(node, i++, NULL, NULL))
node              167 drivers/irqchip/irq-mtk-sysirq.c 		ret = of_address_to_resource(node, i, &res);
node              171 drivers/irqchip/irq-mtk-sysirq.c 		chip_data->intpol_bases[i] = of_iomap(node, i);
node              173 drivers/irqchip/irq-mtk-sysirq.c 			pr_err("%pOF: couldn't map region %d\n", node, i);
node              209 drivers/irqchip/irq-mtk-sysirq.c 	domain = irq_domain_add_hierarchy(domain_parent, 0, intpol_num, node,
node              172 drivers/irqchip/irq-mvebu-gicp.c 	struct device_node *node = pdev->dev.of_node;
node              187 drivers/irqchip/irq-mvebu-gicp.c 	ret = of_property_count_u32_elems(node, "marvell,spi-ranges");
node              202 drivers/irqchip/irq-mvebu-gicp.c 		of_property_read_u32_index(node, "marvell,spi-ranges",
node              206 drivers/irqchip/irq-mvebu-gicp.c 		of_property_read_u32_index(node, "marvell,spi-ranges",
node              219 drivers/irqchip/irq-mvebu-gicp.c 	irq_parent_dn = of_irq_find_parent(node);
node              233 drivers/irqchip/irq-mvebu-gicp.c 						   of_node_to_fwnode(node),
node              239 drivers/irqchip/irq-mvebu-gicp.c 	plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node),
node              161 drivers/irqchip/irq-mvebu-odmi.c static int __init mvebu_odmi_init(struct device_node *node,
node              167 drivers/irqchip/irq-mvebu-odmi.c 	if (of_property_read_u32(node, "marvell,odmi-frames", &odmis_count))
node              184 drivers/irqchip/irq-mvebu-odmi.c 		ret = of_address_to_resource(node, i, &odmi->res);
node              188 drivers/irqchip/irq-mvebu-odmi.c 		odmi->base = of_io_request_and_map(node, i, "odmi");
node              194 drivers/irqchip/irq-mvebu-odmi.c 		if (of_property_read_u32_index(node, "marvell,spi-base",
node              201 drivers/irqchip/irq-mvebu-odmi.c 	inner_domain = irq_domain_create_linear(of_node_to_fwnode(node),
node              211 drivers/irqchip/irq-mvebu-odmi.c 	plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node),
node              124 drivers/irqchip/irq-mvebu-pic.c 	struct device_node *node = pdev->dev.of_node;
node              144 drivers/irqchip/irq-mvebu-pic.c 	pic->parent_irq = irq_of_parse_and_map(node, 0);
node              150 drivers/irqchip/irq-mvebu-pic.c 	pic->domain = irq_domain_add_linear(node, PIC_MAX_IRQS,
node              370 drivers/irqchip/irq-mvebu-sei.c 	struct device_node *node = pdev->dev.of_node;
node              404 drivers/irqchip/irq-mvebu-sei.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              411 drivers/irqchip/irq-mvebu-sei.c 	sei->sei_domain = irq_domain_create_linear(of_node_to_fwnode(node),
node              427 drivers/irqchip/irq-mvebu-sei.c 						     of_node_to_fwnode(node),
node              441 drivers/irqchip/irq-mvebu-sei.c 						     of_node_to_fwnode(node),
node              452 drivers/irqchip/irq-mvebu-sei.c 	plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node),
node               83 drivers/irqchip/irq-nvic.c static int __init nvic_of_init(struct device_node *node,
node               93 drivers/irqchip/irq-nvic.c 	nvic_base = of_iomap(node, 0);
node              104 drivers/irqchip/irq-nvic.c 		irq_domain_add_linear(node, irqs, &nvic_irq_domain_ops, NULL);
node              243 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq_of(struct device_node *node)
node              247 drivers/irqchip/irq-omap-intc.c 	omap_irq_base = of_iomap(node, 0);
node              251 drivers/irqchip/irq-omap-intc.c 	domain = irq_domain_add_linear(node, omap_nr_irqs,
node              263 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq_legacy(u32 base, struct device_node *node)
node              277 drivers/irqchip/irq-omap-intc.c 	domain = irq_domain_add_legacy(node, omap_nr_irqs, irq_base, 0,
node              297 drivers/irqchip/irq-omap-intc.c static int __init omap_init_irq(u32 base, struct device_node *node)
node              307 drivers/irqchip/irq-omap-intc.c 	if (of_device_is_compatible(node, "ti,omap2-intc") ||
node              308 drivers/irqchip/irq-omap-intc.c 			of_device_is_compatible(node, "ti,omap3-intc")) {
node              311 drivers/irqchip/irq-omap-intc.c 		if (of_address_to_resource(node, 0, &res))
node              315 drivers/irqchip/irq-omap-intc.c 		ret = omap_init_irq_legacy(base, node);
node              316 drivers/irqchip/irq-omap-intc.c 	} else if (node) {
node              317 drivers/irqchip/irq-omap-intc.c 		ret = omap_init_irq_of(node);
node              363 drivers/irqchip/irq-omap-intc.c static int __init intc_of_init(struct device_node *node,
node              371 drivers/irqchip/irq-omap-intc.c 	if (WARN_ON(!node))
node              374 drivers/irqchip/irq-omap-intc.c 	if (of_device_is_compatible(node, "ti,dm814-intc") ||
node              375 drivers/irqchip/irq-omap-intc.c 	    of_device_is_compatible(node, "ti,dm816-intc") ||
node              376 drivers/irqchip/irq-omap-intc.c 	    of_device_is_compatible(node, "ti,am33xx-intc")) {
node              381 drivers/irqchip/irq-omap-intc.c 	ret = omap_init_irq(-1, of_node_get(node));
node              147 drivers/irqchip/irq-ompic.c static int __init ompic_of_init(struct device_node *node,
node              160 drivers/irqchip/irq-ompic.c 	if (of_address_to_resource(node, 0, &res)) {
node              179 drivers/irqchip/irq-ompic.c 	irq = irq_of_parse_and_map(node, 0);
node              142 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_init(struct device_node *node,
node              148 drivers/irqchip/irq-or1k-pic.c 	root_domain = irq_domain_add_linear(node, 32, &or1k_irq_domain_ops,
node              156 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_or1200_init(struct device_node *node,
node              159 drivers/irqchip/irq-or1k-pic.c 	return or1k_pic_init(node, &or1k_pic_or1200);
node              164 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_level_init(struct device_node *node,
node              167 drivers/irqchip/irq-or1k-pic.c 	return or1k_pic_init(node, &or1k_pic_level);
node              172 drivers/irqchip/irq-or1k-pic.c static int __init or1k_pic_edge_init(struct device_node *node,
node              175 drivers/irqchip/irq-or1k-pic.c 	return or1k_pic_init(node, &or1k_pic_edge);
node              192 drivers/irqchip/irq-pic32-evic.c 	struct device_node *node = irq_domain_get_of_node(domain);
node              200 drivers/irqchip/irq-pic32-evic.c 	of_property_for_each_u32(node, pname, prop, p, hwirq) {
node              212 drivers/irqchip/irq-pic32-evic.c static int __init pic32_of_init(struct device_node *node,
node              223 drivers/irqchip/irq-pic32-evic.c 	evic_base = of_iomap(node, 0);
node              233 drivers/irqchip/irq-pic32-evic.c 	evic_irq_domain = irq_domain_add_linear(node, nchips * 32,
node               84 drivers/irqchip/irq-rda-intc.c static int __init rda8810_intc_init(struct device_node *node,
node               87 drivers/irqchip/irq-rda-intc.c 	rda_intc_base = of_io_request_and_map(node, 0, "rda-intc");
node               94 drivers/irqchip/irq-rda-intc.c 	rda_irq_domain = irq_domain_create_linear(&node->fwnode, RDA_NR_IRQS,
node              201 drivers/irqchip/irq-sifive-plic.c static int plic_find_hart_id(struct device_node *node)
node              203 drivers/irqchip/irq-sifive-plic.c 	for (; node; node = node->parent) {
node              204 drivers/irqchip/irq-sifive-plic.c 		if (of_device_is_compatible(node, "riscv"))
node              205 drivers/irqchip/irq-sifive-plic.c 			return riscv_of_processor_hartid(node);
node              211 drivers/irqchip/irq-sifive-plic.c static int __init plic_init(struct device_node *node,
node              222 drivers/irqchip/irq-sifive-plic.c 	plic_regs = of_iomap(node, 0);
node              227 drivers/irqchip/irq-sifive-plic.c 	of_property_read_u32(node, "riscv,ndev", &nr_irqs);
node              231 drivers/irqchip/irq-sifive-plic.c 	nr_contexts = of_irq_count(node);
node              238 drivers/irqchip/irq-sifive-plic.c 	plic_irqdomain = irq_domain_add_linear(node, nr_irqs + 1,
node              250 drivers/irqchip/irq-sifive-plic.c 		if (of_irq_parse_one(node, i, &parent)) {
node              210 drivers/irqchip/irq-sni-exiu.c static int __init exiu_dt_init(struct device_node *node,
node              218 drivers/irqchip/irq-sni-exiu.c 		pr_err("%pOF: no parent, giving up\n", node);
node              224 drivers/irqchip/irq-sni-exiu.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              228 drivers/irqchip/irq-sni-exiu.c 	if (of_address_to_resource(node, 0, &res)) {
node              229 drivers/irqchip/irq-sni-exiu.c 		pr_err("%pOF: failed to parse memory resource\n", node);
node              233 drivers/irqchip/irq-sni-exiu.c 	data = exiu_init(of_node_to_fwnode(node), &res);
node              237 drivers/irqchip/irq-sni-exiu.c 	domain = irq_domain_add_hierarchy(parent_domain, 0, NUM_IRQS, node,
node              240 drivers/irqchip/irq-sni-exiu.c 		pr_err("%pOF: failed to allocate domain\n", node);
node              244 drivers/irqchip/irq-sni-exiu.c 	pr_info("%pOF: %d interrupts forwarded to %pOF\n", node, NUM_IRQS,
node              653 drivers/irqchip/irq-stm32-exti.c 					   struct device_node *node)
node              668 drivers/irqchip/irq-stm32-exti.c 	host_data->base = of_iomap(node, 0);
node              670 drivers/irqchip/irq-stm32-exti.c 		pr_err("%pOF: Unable to map registers\n", node);
node              689 drivers/irqchip/irq-stm32-exti.c 					   struct device_node *node)
node              709 drivers/irqchip/irq-stm32-exti.c 	pr_info("%pOF: bank%d\n", node, bank_idx);
node              715 drivers/irqchip/irq-stm32-exti.c 				  struct device_node *node)
node              723 drivers/irqchip/irq-stm32-exti.c 	host_data = stm32_exti_host_init(drv_data, node);
node              727 drivers/irqchip/irq-stm32-exti.c 	domain = irq_domain_add_linear(node, drv_data->bank_nr * IRQS_PER_BANK,
node              731 drivers/irqchip/irq-stm32-exti.c 		       node);
node              740 drivers/irqchip/irq-stm32-exti.c 		       node);
node              749 drivers/irqchip/irq-stm32-exti.c 		chip_data = stm32_exti_chip_init(host_data, i, node);
node              768 drivers/irqchip/irq-stm32-exti.c 	nr_irqs = of_irq_count(node);
node              770 drivers/irqchip/irq-stm32-exti.c 		unsigned int irq = irq_of_parse_and_map(node, i);
node              107 drivers/irqchip/irq-sun4i.c static int __init sun4i_of_init(struct device_node *node,
node              110 drivers/irqchip/irq-sun4i.c 	irq_ic_data->irq_base = of_iomap(node, 0);
node              113 drivers/irqchip/irq-sun4i.c 			node);
node              136 drivers/irqchip/irq-sun4i.c 	irq_ic_data->irq_domain = irq_domain_add_linear(node, 3 * 32,
node              139 drivers/irqchip/irq-sun4i.c 		panic("%pOF: unable to create IRQ domain\n", node);
node              146 drivers/irqchip/irq-sun4i.c static int __init sun4i_ic_of_init(struct device_node *node,
node              158 drivers/irqchip/irq-sun4i.c 	return sun4i_of_init(node, parent);
node              163 drivers/irqchip/irq-sun4i.c static int __init suniv_ic_of_init(struct device_node *node,
node              175 drivers/irqchip/irq-sun4i.c 	return sun4i_of_init(node, parent);
node              159 drivers/irqchip/irq-sunxi-nmi.c static int __init sunxi_sc_nmi_irq_init(struct device_node *node,
node              169 drivers/irqchip/irq-sunxi-nmi.c 	domain = irq_domain_add_linear(node, 1, &irq_generic_chip_ops, NULL);
node              183 drivers/irqchip/irq-sunxi-nmi.c 	irq = irq_of_parse_and_map(node, 0);
node              191 drivers/irqchip/irq-sunxi-nmi.c 	gc->reg_base = of_io_request_and_map(node, 0, of_node_full_name(node));
node              235 drivers/irqchip/irq-sunxi-nmi.c static int __init sun6i_r_intc_irq_init(struct device_node *node,
node              238 drivers/irqchip/irq-sunxi-nmi.c 	return sunxi_sc_nmi_irq_init(node, &sun6i_r_intc_reg_offs);
node              243 drivers/irqchip/irq-sunxi-nmi.c static int __init sun6i_sc_nmi_irq_init(struct device_node *node,
node              246 drivers/irqchip/irq-sunxi-nmi.c 	return sunxi_sc_nmi_irq_init(node, &sun6i_reg_offs);
node              250 drivers/irqchip/irq-sunxi-nmi.c static int __init sun7i_sc_nmi_irq_init(struct device_node *node,
node              253 drivers/irqchip/irq-sunxi-nmi.c 	return sunxi_sc_nmi_irq_init(node, &sun7i_reg_offs);
node              257 drivers/irqchip/irq-sunxi-nmi.c static int __init sun9i_nmi_irq_init(struct device_node *node,
node              260 drivers/irqchip/irq-sunxi-nmi.c 	return sunxi_sc_nmi_irq_init(node, &sun9i_reg_offs);
node              173 drivers/irqchip/irq-tango.c 				  struct device_node *node)
node              181 drivers/irqchip/irq-tango.c 	irq = irq_of_parse_and_map(node, 0);
node              183 drivers/irqchip/irq-tango.c 		panic("%pOFn: failed to get IRQ", node);
node              185 drivers/irqchip/irq-tango.c 	err = of_address_to_resource(node, 0, &res);
node              187 drivers/irqchip/irq-tango.c 		panic("%pOFn: failed to get address", node);
node              193 drivers/irqchip/irq-tango.c 	dom = irq_domain_add_linear(node, 64, &irq_generic_chip_ops, chip);
node              195 drivers/irqchip/irq-tango.c 		panic("%pOFn: failed to create irqdomain", node);
node              197 drivers/irqchip/irq-tango.c 	err = irq_alloc_domain_generic_chips(dom, 32, 2, node->name,
node              200 drivers/irqchip/irq-tango.c 		panic("%pOFn: failed to allocate irqchip", node);
node              209 drivers/irqchip/irq-tango.c static int __init tangox_of_irq_init(struct device_node *node,
node              216 drivers/irqchip/irq-tango.c 	base = of_iomap(node, 0);
node              218 drivers/irqchip/irq-tango.c 		panic("%pOFn: of_iomap failed", node);
node              220 drivers/irqchip/irq-tango.c 	of_address_to_resource(node, 0, &res);
node              222 drivers/irqchip/irq-tango.c 	for_each_child_of_node(node, c)
node              275 drivers/irqchip/irq-tegra.c static int __init tegra_ictlr_init(struct device_node *node,
node              285 drivers/irqchip/irq-tegra.c 		pr_err("%pOF: no parent, giving up\n", node);
node              291 drivers/irqchip/irq-tegra.c 		pr_err("%pOF: unable to obtain parent domain\n", node);
node              295 drivers/irqchip/irq-tegra.c 	match = of_match_node(ictlr_matches, node);
node              308 drivers/irqchip/irq-tegra.c 		base = of_iomap(node, i);
node              323 drivers/irqchip/irq-tegra.c 		pr_err("%pOF: no valid regions, giving up\n", node);
node              330 drivers/irqchip/irq-tegra.c 	     node, num_ictlrs, soc->num_ictlrs);
node              334 drivers/irqchip/irq-tegra.c 					  node, &tegra_ictlr_domain_ops,
node              337 drivers/irqchip/irq-tegra.c 		pr_err("%pOF: failed to allocated domain\n", node);
node              345 drivers/irqchip/irq-tegra.c 		node, num_ictlrs * 32, parent);
node              520 drivers/irqchip/irq-ti-sci-inta.c 	struct device_node *parent_node, *node;
node              526 drivers/irqchip/irq-ti-sci-inta.c 	node = dev_of_node(dev);
node              527 drivers/irqchip/irq-ti-sci-inta.c 	parent_node = of_irq_find_parent(node);
node              584 drivers/irqchip/irq-ti-sci-inta.c 	msi_domain = ti_sci_inta_msi_create_irq_domain(of_node_to_fwnode(node),
node               94 drivers/irqchip/irq-ts4800.c 	struct device_node *node = pdev->dev.of_node;
node              111 drivers/irqchip/irq-ts4800.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              122 drivers/irqchip/irq-ts4800.c 	data->domain = irq_domain_add_linear(node, 8, &ts4800_ic_ops, data);
node              150 drivers/irqchip/irq-versatile-fpga.c 			  int parent_irq, u32 valid, struct device_node *node)
node              173 drivers/irqchip/irq-versatile-fpga.c 	f->domain = irq_domain_add_simple(node, fls(valid), irq_start,
node              195 drivers/irqchip/irq-versatile-fpga.c int __init fpga_irq_of_init(struct device_node *node,
node              203 drivers/irqchip/irq-versatile-fpga.c 	if (WARN_ON(!node))
node              206 drivers/irqchip/irq-versatile-fpga.c 	base = of_iomap(node, 0);
node              209 drivers/irqchip/irq-versatile-fpga.c 	if (of_property_read_u32(node, "clear-mask", &clear_mask))
node              212 drivers/irqchip/irq-versatile-fpga.c 	if (of_property_read_u32(node, "valid-mask", &valid_mask))
node              219 drivers/irqchip/irq-versatile-fpga.c 	parent_irq = irq_of_parse_and_map(node, 0);
node              225 drivers/irqchip/irq-versatile-fpga.c 	fpga_irq_init(base, node->name, 0, parent_irq, valid_mask, node);
node              232 drivers/irqchip/irq-versatile-fpga.c 	if (of_device_is_compatible(node, "arm,versatile-sic"))
node              178 drivers/irqchip/irq-vf610-mscm-ir.c static int __init vf610_mscm_ir_of_init(struct device_node *node,
node              195 drivers/irqchip/irq-vf610-mscm-ir.c 	mscm_ir_data->mscm_ir_base = of_io_request_and_map(node, 0, "mscm-ir");
node              202 drivers/irqchip/irq-vf610-mscm-ir.c 	mscm_cp_regmap = syscon_regmap_lookup_by_phandle(node, "fsl,cpucfg");
node              213 drivers/irqchip/irq-vf610-mscm-ir.c 					  MSCM_IRSPRC_NUM, node,
node              268 drivers/irqchip/irq-vic.c 				struct device_node *node)
node              290 drivers/irqchip/irq-vic.c 	v->domain = irq_domain_add_simple(node, fls(valid_sources), irq,
node              402 drivers/irqchip/irq-vic.c 			       u32 vic_sources, struct device_node *node)
node              428 drivers/irqchip/irq-vic.c 	vic_register(base, 0, irq_start, vic_sources, 0, node);
node              433 drivers/irqchip/irq-vic.c 			      struct device_node *node)
node              451 drivers/irqchip/irq-vic.c 		vic_init_st(base, irq_start, vic_sources, node);
node              468 drivers/irqchip/irq-vic.c 	vic_register(base, parent_irq, irq_start, vic_sources, resume_sources, node);
node              506 drivers/irqchip/irq-vic.c static int __init vic_of_init(struct device_node *node,
node              516 drivers/irqchip/irq-vic.c 	regs = of_iomap(node, 0);
node              520 drivers/irqchip/irq-vic.c 	of_property_read_u32(node, "valid-mask", &interrupt_mask);
node              521 drivers/irqchip/irq-vic.c 	of_property_read_u32(node, "valid-wakeup-mask", &wakeup_mask);
node              526 drivers/irqchip/irq-vic.c 	__vic_init(regs, 0, 0, interrupt_mask, wakeup_mask, node);
node              190 drivers/irqchip/irq-vt8500.c static int __init vt8500_irq_init(struct device_node *node,
node              194 drivers/irqchip/irq-vt8500.c 	struct device_node *np = node;
node              203 drivers/irqchip/irq-vt8500.c 	intc[active_cnt].domain = irq_domain_add_linear(node, 64,
node               69 drivers/irqchip/irq-zevio.c static int __init zevio_of_init(struct device_node *node,
node               79 drivers/irqchip/irq-zevio.c 	zevio_irq_io = of_iomap(node, 0);
node               95 drivers/irqchip/irq-zevio.c 	zevio_irq_domain = irq_domain_add_linear(node, MAX_INTRS,
node              271 drivers/irqchip/qcom-pdc.c static int qcom_pdc_init(struct device_node *node, struct device_node *parent)
node              276 drivers/irqchip/qcom-pdc.c 	pdc_base = of_iomap(node, 0);
node              278 drivers/irqchip/qcom-pdc.c 		pr_err("%pOF: unable to map PDC registers\n", node);
node              284 drivers/irqchip/qcom-pdc.c 		pr_err("%pOF: unable to find PDC's parent domain\n", node);
node              289 drivers/irqchip/qcom-pdc.c 	ret = pdc_setup_pin_mapping(node);
node              291 drivers/irqchip/qcom-pdc.c 		pr_err("%pOF: failed to init PDC pin-hwirq mapping\n", node);
node              296 drivers/irqchip/qcom-pdc.c 						 of_fwnode_handle(node),
node              299 drivers/irqchip/qcom-pdc.c 		pr_err("%pOF: GIC domain add failed\n", node);
node              305 drivers/leds/led-class.c 	list_add_tail(&led_cdev->node, &leds_list);
node              361 drivers/leds/led-class.c 	list_del(&led_cdev->node);
node              248 drivers/leds/led-triggers.c 	list_for_each_entry(led_cdev, &leds_list, node) {
node              277 drivers/leds/led-triggers.c 	list_for_each_entry(led_cdev, &leds_list, node) {
node              602 drivers/leds/leds-max77693.c 	struct device_node *node = dev->of_node, *child_node;
node              607 drivers/leds/leds-max77693.c 	of_property_read_u32(node, "maxim,boost-mode", &cfg->boost_mode);
node              608 drivers/leds/leds-max77693.c 	of_property_read_u32(node, "maxim,boost-mvout", &cfg->boost_vout);
node              609 drivers/leds/leds-max77693.c 	of_property_read_u32(node, "maxim,mvsys-min", &cfg->low_vsys);
node              611 drivers/leds/leds-max77693.c 	for_each_available_child_of_node(node, child_node) {
node               48 drivers/leds/trigger/ledtrig-panic.c 	list_for_each_entry(led_cdev, &leds_list, node)
node              383 drivers/lightnvm/core.c 	tqueue = blk_alloc_queue_node(GFP_KERNEL, dev->q->node);
node             1167 drivers/lightnvm/core.c struct nvm_dev *nvm_alloc_dev(int node)
node             1171 drivers/lightnvm/core.c 	dev = kzalloc_node(sizeof(struct nvm_dev), GFP_KERNEL, node);
node              160 drivers/macintosh/ans-lcd.c 	struct device_node* node;
node              162 drivers/macintosh/ans-lcd.c 	node = of_find_node_by_name(NULL, "lcd");
node              163 drivers/macintosh/ans-lcd.c 	if (!node || !of_node_name_eq(node->parent, "gc")) {
node              164 drivers/macintosh/ans-lcd.c 		of_node_put(node);
node              167 drivers/macintosh/ans-lcd.c 	of_node_put(node);
node              156 drivers/macintosh/windfarm_smu_controls.c static struct smu_fan_control *smu_fan_create(struct device_node *node,
node              168 drivers/macintosh/windfarm_smu_controls.c 	l = of_get_property(node, "location", NULL);
node              231 drivers/macintosh/windfarm_smu_controls.c 	v = of_get_property(node, "min-value", NULL);
node              235 drivers/macintosh/windfarm_smu_controls.c 	v = of_get_property(node, "max-value", NULL);
node              241 drivers/macintosh/windfarm_smu_controls.c 	reg = of_get_property(node, "reg", NULL);
node               35 drivers/macintosh/windfarm_smu_sat.c 	struct device_node	*node;
node              210 drivers/macintosh/windfarm_smu_sat.c 	sat->node = of_node_get(dev);
node              196 drivers/macintosh/windfarm_smu_sensors.c static struct smu_ad_sensor *smu_ads_create(struct device_node *node)
node              205 drivers/macintosh/windfarm_smu_sensors.c 	l = of_get_property(node, "location", NULL);
node              216 drivers/macintosh/windfarm_smu_sensors.c 	if (of_node_is_type(node, "temp-sensor") &&
node              225 drivers/macintosh/windfarm_smu_sensors.c 	} else if (of_node_is_type(node, "current-sensor") &&
node              234 drivers/macintosh/windfarm_smu_sensors.c 	} else if (of_node_is_type(node, "voltage-sensor") &&
node              243 drivers/macintosh/windfarm_smu_sensors.c 	} else if (of_node_is_type(node, "power-sensor") &&
node              255 drivers/macintosh/windfarm_smu_sensors.c 	v = of_get_property(node, "reg", NULL);
node              264 drivers/mailbox/hi6220-mailbox.c 	struct device_node *node = pdev->dev.of_node;
node              331 drivers/mailbox/hi6220-mailbox.c 	if (of_find_property(node, "hi6220,mbox-tx-noirq", NULL))
node              353 drivers/mailbox/mailbox.c 	list_for_each_entry(mbox, &mbox_cons, node)
node              516 drivers/mailbox/mailbox.c 	list_add_tail(&mbox->node, &mbox_cons);
node              536 drivers/mailbox/mailbox.c 	list_del(&mbox->node);
node              676 drivers/mailbox/omap-mailbox.c 	struct device_node *node;
node              684 drivers/mailbox/omap-mailbox.c 	node = of_find_node_by_phandle(phandle);
node              685 drivers/mailbox/omap-mailbox.c 	if (!node) {
node              691 drivers/mailbox/omap-mailbox.c 	mbox = omap_mbox_device_find(mdev, node->name);
node              692 drivers/mailbox/omap-mailbox.c 	of_node_put(node);
node              705 drivers/mailbox/omap-mailbox.c 	struct device_node *node = pdev->dev.of_node;
node              714 drivers/mailbox/omap-mailbox.c 	if (!node) {
node              724 drivers/mailbox/omap-mailbox.c 	if (of_property_read_u32(node, "ti,mbox-num-users", &num_users))
node              727 drivers/mailbox/omap-mailbox.c 	if (of_property_read_u32(node, "ti,mbox-num-fifos", &num_fifos))
node              730 drivers/mailbox/omap-mailbox.c 	info_count = of_get_available_child_count(node);
node              744 drivers/mailbox/omap-mailbox.c 		child = of_get_next_available_child(node, child);
node              436 drivers/mailbox/zynqmp-ipi-mailbox.c static int zynqmp_ipi_mbox_get_buf_res(struct device_node *node,
node              442 drivers/mailbox/zynqmp-ipi-mailbox.c 	index = of_property_match_string(node, "reg-names", name);
node              444 drivers/mailbox/zynqmp-ipi-mailbox.c 		ret = of_address_to_resource(node, index, res);
node              474 drivers/mailbox/zynqmp-ipi-mailbox.c 				 struct device_node *node)
node              488 drivers/mailbox/zynqmp-ipi-mailbox.c 	ipi_mbox->dev.of_node = node;
node              489 drivers/mailbox/zynqmp-ipi-mailbox.c 	dev_set_name(&ipi_mbox->dev, "%s", of_node_full_name(node));
node              502 drivers/mailbox/zynqmp-ipi-mailbox.c 	ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res);
node              518 drivers/mailbox/zynqmp-ipi-mailbox.c 	ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res);
node              541 drivers/mailbox/zynqmp-ipi-mailbox.c 	ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res);
node              557 drivers/mailbox/zynqmp-ipi-mailbox.c 	ret = zynqmp_ipi_mbox_get_buf_res(node, name, &res);
node              579 drivers/mailbox/zynqmp-ipi-mailbox.c 	ret = of_property_read_u32(node, "xlnx,ipi-id", &ipi_mbox->remote_id);
node              226 drivers/md/bcache/bcache.h 	struct rb_node		node;
node             1791 drivers/md/bcache/btree.c 					&dc->writeback_keys.keys, node)
node             2508 drivers/md/bcache/btree.c 		if (RB_INSERT(&buf->keys, w, node, keybuf_cmp))
node             2550 drivers/md/bcache/btree.c 		w = RB_FIRST(&buf->keys, struct keybuf_key, node);
node             2553 drivers/md/bcache/btree.c 		w = RB_LAST(&buf->keys, struct keybuf_key, node);
node             2565 drivers/md/bcache/btree.c 	rb_erase(&w->node, &buf->keys);
node             2589 drivers/md/bcache/btree.c 	w = RB_GREATER(&buf->keys, s, node, keybuf_nonoverlapping_cmp);
node             2593 drivers/md/bcache/btree.c 		w = RB_NEXT(w, node);
node             2611 drivers/md/bcache/btree.c 	w = RB_FIRST(&buf->keys, struct keybuf_key, node);
node             2614 drivers/md/bcache/btree.c 		w = RB_NEXT(w, node);
node              120 drivers/md/dm-bio-prison-v1.c 			rb_entry(*new, struct dm_bio_prison_cell, node);
node              140 drivers/md/dm-bio-prison-v1.c 	rb_link_node(&cell_prealloc->node, parent, new);
node              141 drivers/md/dm-bio-prison-v1.c 	rb_insert_color(&cell_prealloc->node, &prison->cells);
node              188 drivers/md/dm-bio-prison-v1.c 	rb_erase(&cell->node, &prison->cells);
node              216 drivers/md/dm-bio-prison-v1.c 	rb_erase(&cell->node, &prison->cells);
node              257 drivers/md/dm-bio-prison-v1.c 	rb_erase(&cell->node, &prison->cells);
node              266 drivers/md/dm-bio-prison-v1.c 		rb_erase(&cell->node, &prison->cells);
node               42 drivers/md/dm-bio-prison-v1.h 	struct rb_node node;
node              124 drivers/md/dm-bio-prison-v2.c 			rb_entry(*new, struct dm_bio_prison_cell_v2, node);
node              143 drivers/md/dm-bio-prison-v2.c 	rb_link_node(&cell_prealloc->node, parent, new);
node              144 drivers/md/dm-bio-prison-v2.c 	rb_insert_color(&cell_prealloc->node, &prison->cells);
node              204 drivers/md/dm-bio-prison-v2.c 			rb_erase(&cell->node, &prison->cells);
node              336 drivers/md/dm-bio-prison-v2.c 	rb_erase(&cell->node, &prison->cells);
node               51 drivers/md/dm-bio-prison-v2.h 	struct rb_node node;
node              134 drivers/md/dm-bufio.c 	struct rb_node node;
node              254 drivers/md/dm-bufio.c 		b = container_of(n, struct dm_buffer, node);
node              271 drivers/md/dm-bufio.c 		found = container_of(*new, struct dm_buffer, node);
node              283 drivers/md/dm-bufio.c 	rb_link_node(&b->node, parent, new);
node              284 drivers/md/dm-bufio.c 	rb_insert_color(&b->node, &c->buffer_tree);
node              289 drivers/md/dm-bufio.c 	rb_erase(&b->node, &c->buffer_tree);
node               15 drivers/md/dm-cache-background-tracker.c 	struct rb_node node;
node               87 drivers/md/dm-cache-background-tracker.c 		w = container_of(*new, struct bt_work, node);
node              102 drivers/md/dm-cache-background-tracker.c 	rb_link_node(&nw->node, parent, new);
node              103 drivers/md/dm-cache-background-tracker.c 	rb_insert_color(&nw->node, &b->pending);
node              116 drivers/md/dm-cache-background-tracker.c 		w = container_of(*new, struct bt_work, node);
node              236 drivers/md/dm-cache-background-tracker.c 	rb_erase(&w->node, &b->pending);
node             1509 drivers/md/dm-crypt.c #define crypt_io_from_node(node) rb_entry((node), struct dm_crypt_io, rb_node)
node               18 drivers/md/dm-dust.c 	struct rb_node node;
node               37 drivers/md/dm-dust.c 	struct rb_node *node = root->rb_node;
node               39 drivers/md/dm-dust.c 	while (node) {
node               40 drivers/md/dm-dust.c 		struct badblock *bblk = rb_entry(node, struct badblock, node);
node               43 drivers/md/dm-dust.c 			node = node->rb_left;
node               45 drivers/md/dm-dust.c 			node = node->rb_right;
node               61 drivers/md/dm-dust.c 		bblk = rb_entry(parent, struct badblock, node);
node               71 drivers/md/dm-dust.c 	rb_link_node(&new->node, parent, link);
node               72 drivers/md/dm-dust.c 	rb_insert_color(&new->node, root);
node               94 drivers/md/dm-dust.c 	rb_erase(&bblock->node, &dd->badblocklist);
node              183 drivers/md/dm-dust.c 		rb_erase(&bblk->node, &dd->badblocklist);
node              228 drivers/md/dm-dust.c 	struct rb_node *node = NULL, *nnode = NULL;
node              237 drivers/md/dm-dust.c 		node = nnode;
node              238 drivers/md/dm-dust.c 		nnode = rb_next(node);
node              239 drivers/md/dm-dust.c 		rb_erase(node, tree);
node              241 drivers/md/dm-dust.c 		kfree(node);
node              140 drivers/md/dm-integrity.c 	struct rb_node node;
node              272 drivers/md/dm-integrity.c 		struct rb_node node;
node             1116 drivers/md/dm-integrity.c 		struct dm_integrity_range *range = container_of(*n, struct dm_integrity_range, node);
node             1120 drivers/md/dm-integrity.c 			n = &range->node.rb_left;
node             1122 drivers/md/dm-integrity.c 			n = &range->node.rb_right;
node             1128 drivers/md/dm-integrity.c 	rb_link_node(&new_range->node, parent, n);
node             1129 drivers/md/dm-integrity.c 	rb_insert_color(&new_range->node, &ic->in_progress);
node             1136 drivers/md/dm-integrity.c 	rb_erase(&range->node, &ic->in_progress);
node             1181 drivers/md/dm-integrity.c static void init_journal_node(struct journal_node *node)
node             1183 drivers/md/dm-integrity.c 	RB_CLEAR_NODE(&node->node);
node             1184 drivers/md/dm-integrity.c 	node->sector = (sector_t)-1;
node             1187 drivers/md/dm-integrity.c static void add_journal_node(struct dm_integrity_c *ic, struct journal_node *node, sector_t sector)
node             1192 drivers/md/dm-integrity.c 	node->sector = sector;
node             1193 drivers/md/dm-integrity.c 	BUG_ON(!RB_EMPTY_NODE(&node->node));
node             1201 drivers/md/dm-integrity.c 		j = container_of(parent, struct journal_node, node);
node             1203 drivers/md/dm-integrity.c 			link = &j->node.rb_left;
node             1205 drivers/md/dm-integrity.c 			link = &j->node.rb_right;
node             1208 drivers/md/dm-integrity.c 	rb_link_node(&node->node, parent, link);
node             1209 drivers/md/dm-integrity.c 	rb_insert_color(&node->node, &ic->journal_tree_root);
node             1212 drivers/md/dm-integrity.c static void remove_journal_node(struct dm_integrity_c *ic, struct journal_node *node)
node             1214 drivers/md/dm-integrity.c 	BUG_ON(RB_EMPTY_NODE(&node->node));
node             1215 drivers/md/dm-integrity.c 	rb_erase(&node->node, &ic->journal_tree_root);
node             1216 drivers/md/dm-integrity.c 	init_journal_node(node);
node             1227 drivers/md/dm-integrity.c 		struct journal_node *j = container_of(n, struct journal_node, node);
node             1233 drivers/md/dm-integrity.c 			n = j->node.rb_left;
node             1235 drivers/md/dm-integrity.c 			n = j->node.rb_right;
node             1244 drivers/md/dm-integrity.c 	struct journal_node *node, *next_node;
node             1249 drivers/md/dm-integrity.c 	node = &ic->journal_tree[pos];
node             1250 drivers/md/dm-integrity.c 	if (unlikely(RB_EMPTY_NODE(&node->node)))
node             1252 drivers/md/dm-integrity.c 	if (unlikely(node->sector != sector))
node             1255 drivers/md/dm-integrity.c 	next = rb_next(&node->node);
node             1259 drivers/md/dm-integrity.c 	next_node = container_of(next, struct journal_node, node);
node             1263 drivers/md/dm-integrity.c static bool find_newer_committed_node(struct dm_integrity_c *ic, struct journal_node *node)
node             1269 drivers/md/dm-integrity.c 	BUG_ON(RB_EMPTY_NODE(&node->node));
node             1271 drivers/md/dm-integrity.c 	next = rb_next(&node->node);
node             1275 drivers/md/dm-integrity.c 	next_node = container_of(next, struct journal_node, node);
node             1277 drivers/md/dm-integrity.c 	if (next_node->sector != node->sector)
node              240 drivers/md/dm-snap.c 	struct hlist_node node;
node              247 drivers/md/dm-snap.c 	INIT_HLIST_NODE(&c->node);
node              253 drivers/md/dm-snap.c 	return !hlist_unhashed(&c->node);
node              263 drivers/md/dm-snap.c 	hlist_add_head(&c->node,
node              274 drivers/md/dm-snap.c 	hlist_del(&c->node);
node              286 drivers/md/dm-snap.c 	    &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) {
node              143 drivers/md/dm-stats.c static void *dm_kvzalloc(size_t alloc_size, int node)
node              150 drivers/md/dm-stats.c 	p = kvzalloc_node(alloc_size, GFP_KERNEL | __GFP_NOMEMALLOC, node);
node              125 drivers/md/dm-table.c 	sector_t *node;
node              128 drivers/md/dm-table.c 		node = get_node(t, l, n);
node              131 drivers/md/dm-table.c 			node[k] = high(t, l + 1, get_child(n, k));
node             1366 drivers/md/dm-table.c 	sector_t *node;
node             1373 drivers/md/dm-table.c 		node = get_node(t, l, n);
node             1376 drivers/md/dm-table.c 			if (node[k] >= sector)
node             2108 drivers/md/dm-thin.c #define thin_pbd(node) rb_entry((node), struct dm_thin_endio_hook, rb_node)
node             2136 drivers/md/dm-thin.c 	struct rb_node *node;
node             2140 drivers/md/dm-thin.c 	for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) {
node             2141 drivers/md/dm-thin.c 		pbd = thin_pbd(node);
node              541 drivers/md/dm-writecache.c 	struct rb_node *node = wc->tree.rb_node;
node              543 drivers/md/dm-writecache.c 	if (unlikely(!node))
node              547 drivers/md/dm-writecache.c 		e = container_of(node, struct wc_entry, rb_node);
node              551 drivers/md/dm-writecache.c 		node = (read_original_sector(wc, e) >= block ?
node              553 drivers/md/dm-writecache.c 		if (unlikely(!node)) {
node              559 drivers/md/dm-writecache.c 				node = rb_next(&e->rb_node);
node              560 drivers/md/dm-writecache.c 				if (unlikely(!node))
node              562 drivers/md/dm-writecache.c 				e = container_of(node, struct wc_entry, rb_node);
node              571 drivers/md/dm-writecache.c 			node = rb_prev(&e->rb_node);
node              573 drivers/md/dm-writecache.c 			node = rb_next(&e->rb_node);
node              574 drivers/md/dm-writecache.c 		if (unlikely(!node))
node              576 drivers/md/dm-writecache.c 		e2 = container_of(node, struct wc_entry, rb_node);
node              586 drivers/md/dm-writecache.c 	struct rb_node **node = &wc->tree.rb_node, *parent = NULL;
node              588 drivers/md/dm-writecache.c 	while (*node) {
node              589 drivers/md/dm-writecache.c 		e = container_of(*node, struct wc_entry, rb_node);
node              592 drivers/md/dm-writecache.c 			node = &parent->rb_left;
node              594 drivers/md/dm-writecache.c 			node = &parent->rb_right;
node              596 drivers/md/dm-writecache.c 	rb_link_node(&ins->rb_node, parent, node);
node              610 drivers/md/dm-writecache.c 		struct rb_node **node = &wc->freetree.rb_node, *parent = NULL;
node              611 drivers/md/dm-writecache.c 		if (unlikely(!*node))
node              613 drivers/md/dm-writecache.c 		while (*node) {
node              614 drivers/md/dm-writecache.c 			parent = *node;
node              615 drivers/md/dm-writecache.c 			if (&e->rb_node < *node)
node              616 drivers/md/dm-writecache.c 				node = &parent->rb_left;
node              618 drivers/md/dm-writecache.c 				node = &parent->rb_right;
node              620 drivers/md/dm-writecache.c 		rb_link_node(&e->rb_node, parent, node);
node              800 drivers/md/dm-writecache.c 		struct rb_node *node = rb_next(&e->rb_node);
node              811 drivers/md/dm-writecache.c 		if (unlikely(!node))
node              814 drivers/md/dm-writecache.c 		e = container_of(node, struct wc_entry, rb_node);
node             1604 drivers/md/dm-writecache.c 	struct rb_node *node, *next_node;
node             1652 drivers/md/dm-writecache.c 		node = rb_prev(&e->rb_node);
node             1653 drivers/md/dm-writecache.c 		if (node) {
node             1654 drivers/md/dm-writecache.c 			f = container_of(node, struct wc_entry, rb_node);
node              101 drivers/md/dm-zoned-metadata.c 	struct rb_node		node;
node              281 drivers/md/dm-zoned-metadata.c 			rb_erase(&mblk->node, &zmd->mblk_rbtree);
node              300 drivers/md/dm-zoned-metadata.c 	RB_CLEAR_NODE(&mblk->node);
node              334 drivers/md/dm-zoned-metadata.c 		b = container_of(*new, struct dmz_mblock, node);
node              340 drivers/md/dm-zoned-metadata.c 	rb_link_node(&mblk->node, parent, new);
node              341 drivers/md/dm-zoned-metadata.c 	rb_insert_color(&mblk->node, root);
node              352 drivers/md/dm-zoned-metadata.c 	struct rb_node *node = root->rb_node;
node              355 drivers/md/dm-zoned-metadata.c 	while (node) {
node              356 drivers/md/dm-zoned-metadata.c 		mblk = container_of(node, struct dmz_mblock, node);
node              368 drivers/md/dm-zoned-metadata.c 		node = (mblk->no < mblk_no) ? node->rb_left : node->rb_right;
node              471 drivers/md/dm-zoned-metadata.c 		rb_erase(&mblk->node, &zmd->mblk_rbtree);
node              521 drivers/md/dm-zoned-metadata.c 			rb_erase(&mblk->node, &zmd->mblk_rbtree);
node             2367 drivers/md/dm-zoned-metadata.c 		rb_erase(&mblk->node, &zmd->mblk_rbtree);
node             2375 drivers/md/dm-zoned-metadata.c 		rb_erase(&mblk->node, &zmd->mblk_rbtree);
node             2381 drivers/md/dm-zoned-metadata.c 	rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
node               85 drivers/md/persistent-data/dm-btree.c static int insert_at(size_t value_size, struct btree_node *node, unsigned index,
node               89 drivers/md/persistent-data/dm-btree.c 	uint32_t nr_entries = le32_to_cpu(node->header.nr_entries);
node               93 drivers/md/persistent-data/dm-btree.c 	    index >= le32_to_cpu(node->header.max_entries)) {
node              101 drivers/md/persistent-data/dm-btree.c 	array_insert(node->keys, sizeof(*node->keys), nr_entries, index, &key_le);
node              102 drivers/md/persistent-data/dm-btree.c 	array_insert(value_base(node), value_size, nr_entries, index, value);
node              103 drivers/md/persistent-data/dm-btree.c 	node->header.nr_entries = cpu_to_le32(nr_entries + 1);
node              424 drivers/md/persistent-data/dm-btree.c 	struct dm_block *node;
node              427 drivers/md/persistent-data/dm-btree.c 	r = bn_read_lock(info, root, &node);
node              431 drivers/md/persistent-data/dm-btree.c 	n = dm_block_data(node);
node              466 drivers/md/persistent-data/dm-btree.c 	dm_tm_unlock(info->tm, node);
node              697 drivers/md/persistent-data/dm-btree.c 	struct btree_node *node;
node              704 drivers/md/persistent-data/dm-btree.c 		node = dm_block_data(shadow_current(s));
node              719 drivers/md/persistent-data/dm-btree.c 		node = dm_block_data(shadow_current(s));
node              721 drivers/md/persistent-data/dm-btree.c 		if (node->header.nr_entries == node->header.max_entries) {
node              731 drivers/md/persistent-data/dm-btree.c 		node = dm_block_data(shadow_current(s));
node              733 drivers/md/persistent-data/dm-btree.c 		i = lower_bound(node, key);
node              735 drivers/md/persistent-data/dm-btree.c 		if (le32_to_cpu(node->header.flags) & LEAF_NODE)
node              740 drivers/md/persistent-data/dm-btree.c 			node->keys[0] = cpu_to_le64(key);
node              744 drivers/md/persistent-data/dm-btree.c 		root = value64(node, i);
node              748 drivers/md/persistent-data/dm-btree.c 	if (i < 0 || le64_to_cpu(node->keys[i]) != key)
node              755 drivers/md/persistent-data/dm-btree.c static bool need_insert(struct btree_node *node, uint64_t *keys,
node              758 drivers/md/persistent-data/dm-btree.c         return ((index >= le32_to_cpu(node->header.nr_entries)) ||
node              759 drivers/md/persistent-data/dm-btree.c 		(le64_to_cpu(node->keys[index]) != keys[level]));
node              955 drivers/md/persistent-data/dm-btree.c 	struct dm_block *node;
node              959 drivers/md/persistent-data/dm-btree.c 	r = bn_read_lock(info, block, &node);
node              963 drivers/md/persistent-data/dm-btree.c 	n = dm_block_data(node);
node              980 drivers/md/persistent-data/dm-btree.c 	dm_tm_unlock(info->tm, node);
node             6777 drivers/md/raid5.c static int raid456_cpu_dead(unsigned int cpu, struct hlist_node *node)
node             6779 drivers/md/raid5.c 	struct r5conf *conf = hlist_entry_safe(node, struct r5conf, node);
node             6790 drivers/md/raid5.c 	cpuhp_state_remove_instance(CPUHP_MD_RAID5_PREPARE, &conf->node);
node             6814 drivers/md/raid5.c static int raid456_cpu_up_prepare(unsigned int cpu, struct hlist_node *node)
node             6816 drivers/md/raid5.c 	struct r5conf *conf = hlist_entry_safe(node, struct r5conf, node);
node             6835 drivers/md/raid5.c 	err = cpuhp_state_add_instance(CPUHP_MD_RAID5_PREPARE, &conf->node);
node              646 drivers/md/raid5.h 	struct hlist_node node;
node               21 drivers/media/cec/cec-priv.h #define to_cec_adapter(node) container_of(node, struct cec_adapter, devnode)
node              413 drivers/media/i2c/adp1653.c 			   struct device_node *node)
node              423 drivers/media/i2c/adp1653.c 	child = of_get_child_by_name(node, "flash");
node              444 drivers/media/i2c/adp1653.c 	child = of_get_child_by_name(node, "indicator");
node             1573 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	struct device_node *node = dev->of_node;
node             1579 drivers/media/i2c/s5c73m3/s5c73m3-core.c 		ret = of_get_named_gpio_flags(node, prop_names[i],
node             1596 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	struct device_node *node = dev->of_node;
node             1601 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	if (!node) {
node             1617 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	if (of_property_read_u32(node, "clock-frequency",
node             1628 drivers/media/i2c/s5c73m3/s5c73m3-core.c 	node_ep = of_graph_get_next_endpoint(node, NULL);
node             1630 drivers/media/i2c/s5c73m3/s5c73m3-core.c 		dev_warn(dev, "no endpoint defined for node: %pOF\n", node);
node             1820 drivers/media/i2c/s5k5baf.c 	struct device_node *node = dev->of_node;
node             1825 drivers/media/i2c/s5k5baf.c 		ret = of_get_named_gpio_flags(node, names[i], 0, &flags);
node             1839 drivers/media/i2c/s5k5baf.c 	struct device_node *node = dev->of_node;
node             1844 drivers/media/i2c/s5k5baf.c 	if (!node) {
node             1849 drivers/media/i2c/s5k5baf.c 	ret = of_property_read_u32(node, "clock-frequency",
node             1861 drivers/media/i2c/s5k5baf.c 	node_ep = of_graph_get_next_endpoint(node, NULL);
node             1863 drivers/media/i2c/s5k5baf.c 		dev_err(dev, "no endpoint defined at node %pOF\n", node);
node             1882 drivers/media/i2c/s5k5baf.c 			node);
node             1199 drivers/media/pci/cobalt/cobalt-v4l2.c static int cobalt_node_register(struct cobalt *cobalt, int node)
node             1203 drivers/media/pci/cobalt/cobalt-v4l2.c 	struct cobalt_stream *s = cobalt->streams + node;
node             1212 drivers/media/pci/cobalt/cobalt-v4l2.c 			"%s-%d", cobalt->v4l2_dev.name, node);
node             1232 drivers/media/pci/cobalt/cobalt-v4l2.c 			cobalt_warn("Setting up dummy video node %d\n", node);
node             1282 drivers/media/pci/cobalt/cobalt-v4l2.c 					node);
node             1285 drivers/media/pci/cobalt/cobalt-v4l2.c 	cobalt_info("registered node %d\n", node);
node             1292 drivers/media/pci/cobalt/cobalt-v4l2.c 	int node, ret;
node             1295 drivers/media/pci/cobalt/cobalt-v4l2.c 	for (node = 0; node < COBALT_NUM_STREAMS; node++) {
node             1296 drivers/media/pci/cobalt/cobalt-v4l2.c 		ret = cobalt_node_register(cobalt, node);
node             1306 drivers/media/pci/cobalt/cobalt-v4l2.c 	int node;
node             1309 drivers/media/pci/cobalt/cobalt-v4l2.c 	for (node = 0; node < COBALT_NUM_STREAMS; node++) {
node             1310 drivers/media/pci/cobalt/cobalt-v4l2.c 		struct cobalt_stream *s = cobalt->streams + node;
node             1050 drivers/media/pci/ivtv/ivtvfb.c 	oi->ivtvfb_info.node = -1;
node              350 drivers/media/pci/sta2x11/sta2x11_vip.c 	struct vip_buffer *vip_buf, *node;
node              359 drivers/media/pci/sta2x11/sta2x11_vip.c 	list_for_each_entry_safe(vip_buf, node, &vip->buffer_list, list) {
node               71 drivers/media/platform/atmel/atmel-isi.c 	struct device_node *node;
node              422 drivers/media/platform/atmel/atmel-isi.c 	struct frame_buffer *buf, *node;
node              464 drivers/media/platform/atmel/atmel-isi.c 	list_for_each_entry_safe(buf, node, &isi->video_buffer_list, list) {
node              477 drivers/media/platform/atmel/atmel-isi.c 	struct frame_buffer *buf, *node;
node              489 drivers/media/platform/atmel/atmel-isi.c 	list_for_each_entry_safe(buf, node, &isi->video_buffer_list, list) {
node             1101 drivers/media/platform/atmel/atmel-isi.c static int isi_graph_parse(struct atmel_isi *isi, struct device_node *node)
node             1106 drivers/media/platform/atmel/atmel-isi.c 	ep = of_graph_get_next_endpoint(node, ep);
node             1116 drivers/media/platform/atmel/atmel-isi.c 	isi->entity.node = remote;
node             1137 drivers/media/platform/atmel/atmel-isi.c 		of_node_put(isi->entity.node);
node              867 drivers/media/platform/exynos4-is/fimc-core.c 	struct device_node *node = dev->of_node;
node              874 drivers/media/platform/exynos4-is/fimc-core.c 	if (of_property_read_bool(node, "samsung,lcd-wb"))
node              881 drivers/media/platform/exynos4-is/fimc-core.c 	of_id = of_match_node(fimc_of_match, node);
node              885 drivers/media/platform/exynos4-is/fimc-core.c 	ret = of_property_read_u32_array(node, "samsung,pix-limits",
node              898 drivers/media/platform/exynos4-is/fimc-core.c 	ret = of_property_read_u32_array(node, "samsung,min-pix-sizes",
node              902 drivers/media/platform/exynos4-is/fimc-core.c 	ret = of_property_read_u32_array(node, "samsung,min-pix-alignment",
node              907 drivers/media/platform/exynos4-is/fimc-core.c 	ret = of_property_read_u32(node, "samsung,rotators", &args[1]);
node              910 drivers/media/platform/exynos4-is/fimc-core.c 	v->has_mainscaler_ext = of_property_read_bool(node,
node              913 drivers/media/platform/exynos4-is/fimc-core.c 	v->has_isp_wb = of_property_read_bool(node, "samsung,isp-wb");
node              914 drivers/media/platform/exynos4-is/fimc-core.c 	v->has_cam_if = of_property_read_bool(node, "samsung,cam-if");
node              915 drivers/media/platform/exynos4-is/fimc-core.c 	of_property_read_u32(node, "clock-frequency", clk_freq);
node              916 drivers/media/platform/exynos4-is/fimc-core.c 	fimc->id = of_alias_get_id(node, "fimc");
node              646 drivers/media/platform/exynos4-is/fimc-core.h static inline struct regmap * fimc_get_sysreg_regmap(struct device_node *node)
node              648 drivers/media/platform/exynos4-is/fimc-core.h 	return syscon_regmap_lookup_by_phandle(node, "samsung,sysreg");
node              651 drivers/media/platform/exynos4-is/fimc-core.h #define fimc_get_sysreg_regmap(node) (NULL)
node               39 drivers/media/platform/exynos4-is/fimc-is-i2c.c 	struct device_node *node = pdev->dev.of_node;
node               55 drivers/media/platform/exynos4-is/fimc-is-i2c.c 	i2c_adap->dev.of_node = node;
node               25 drivers/media/platform/exynos4-is/fimc-is-sensor.c 			struct device_node *node)
node               29 drivers/media/platform/exynos4-is/fimc-is-sensor.c 	of_id = of_match_node(fimc_is_sensor_of_ids, node);
node               51 drivers/media/platform/exynos4-is/fimc-is-sensor.h 				struct device_node *node);
node              165 drivers/media/platform/exynos4-is/fimc-is.c 						struct device_node *node)
node              172 drivers/media/platform/exynos4-is/fimc-is.c 	sensor->drvdata = fimc_is_sensor_get_drvdata(node);
node              175 drivers/media/platform/exynos4-is/fimc-is.c 							 node);
node              179 drivers/media/platform/exynos4-is/fimc-is.c 	ep = of_graph_get_next_endpoint(node, NULL);
node              781 drivers/media/platform/exynos4-is/fimc-is.c 	struct device_node *node;
node              803 drivers/media/platform/exynos4-is/fimc-is.c 	node = of_get_child_by_name(dev->of_node, "pmu");
node              804 drivers/media/platform/exynos4-is/fimc-is.c 	if (!node)
node              807 drivers/media/platform/exynos4-is/fimc-is.c 	is->pmu_regs = of_iomap(node, 0);
node              808 drivers/media/platform/exynos4-is/fimc-is.c 	of_node_put(node);
node              475 drivers/media/platform/exynos4-is/media-dev.c 	struct device_node *node;
node              493 drivers/media/platform/exynos4-is/media-dev.c 	for_each_available_child_of_node(parent, node) {
node              496 drivers/media/platform/exynos4-is/media-dev.c 		if (!of_node_name_eq(node, "csis"))
node              499 drivers/media/platform/exynos4-is/media-dev.c 		port = of_get_next_child(node, NULL);
node              506 drivers/media/platform/exynos4-is/media-dev.c 			of_node_put(node);
node              517 drivers/media/platform/exynos4-is/media-dev.c 	for_each_child_of_node(ports, node) {
node              518 drivers/media/platform/exynos4-is/media-dev.c 		ret = fimc_md_parse_port_node(fmd, node, index);
node              520 drivers/media/platform/exynos4-is/media-dev.c 			of_node_put(node);
node              617 drivers/media/platform/exynos4-is/media-dev.c 	struct device_node *node = pdev->dev.of_node;
node              620 drivers/media/platform/exynos4-is/media-dev.c 	id = node ? __of_get_csis_id(node) : max(0, pdev->id);
node              713 drivers/media/platform/exynos4-is/media-dev.c 	struct device_node *node;
node              716 drivers/media/platform/exynos4-is/media-dev.c 	for_each_available_child_of_node(parent, node) {
node              720 drivers/media/platform/exynos4-is/media-dev.c 		pdev = of_find_device_by_node(node);
node              725 drivers/media/platform/exynos4-is/media-dev.c 		if (of_node_name_eq(node, CSIS_OF_NODE_NAME))
node              727 drivers/media/platform/exynos4-is/media-dev.c 		else if (of_node_name_eq(node, FIMC_IS_OF_NODE_NAME))
node              729 drivers/media/platform/exynos4-is/media-dev.c 		else if (of_node_name_eq(node, FIMC_LITE_OF_NODE_NAME))
node              731 drivers/media/platform/exynos4-is/media-dev.c 		else if (of_node_name_eq(node, FIMC_OF_NODE_NAME) &&
node              732 drivers/media/platform/exynos4-is/media-dev.c 			 !of_property_read_bool(node, "samsung,lcd-wb"))
node              740 drivers/media/platform/exynos4-is/media-dev.c 			of_node_put(node);
node              186 drivers/media/platform/exynos4-is/media-dev.h static inline bool fimc_md_is_isp_available(struct device_node *node)
node              188 drivers/media/platform/exynos4-is/media-dev.h 	node = of_get_child_by_name(node, FIMC_IS_OF_NODE_NAME);
node              189 drivers/media/platform/exynos4-is/media-dev.h 	return node ? of_device_is_available(node) : false;
node              192 drivers/media/platform/exynos4-is/media-dev.h #define fimc_md_is_isp_available(node) (false)
node              717 drivers/media/platform/exynos4-is/mipi-csis.c 	struct device_node *node = pdev->dev.of_node;
node              721 drivers/media/platform/exynos4-is/mipi-csis.c 	if (of_property_read_u32(node, "clock-frequency",
node              724 drivers/media/platform/exynos4-is/mipi-csis.c 	if (of_property_read_u32(node, "bus-width",
node              728 drivers/media/platform/exynos4-is/mipi-csis.c 	node = of_graph_get_next_endpoint(node, NULL);
node              729 drivers/media/platform/exynos4-is/mipi-csis.c 	if (!node) {
node              735 drivers/media/platform/exynos4-is/mipi-csis.c 	ret = v4l2_fwnode_endpoint_parse(of_fwnode_handle(node), &endpoint);
node              746 drivers/media/platform/exynos4-is/mipi-csis.c 	of_property_read_u32(node, "samsung,csis-hs-settle",
node              748 drivers/media/platform/exynos4-is/mipi-csis.c 	state->wclk_ext = of_property_read_bool(node,
node              754 drivers/media/platform/exynos4-is/mipi-csis.c 	of_node_put(node);
node             1134 drivers/media/platform/marvell-ccic/mcam-core.c 	struct mcam_vb_buffer *buf, *node;
node             1139 drivers/media/platform/marvell-ccic/mcam-core.c 	list_for_each_entry_safe(buf, node, &cam->buffers, queue) {
node             1054 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 	struct device_node *node;
node             1057 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 	node = of_parse_phandle(jpeg->dev->of_node, "mediatek,larb", 0);
node             1058 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 	if (!node)
node             1060 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 	pdev = of_find_device_by_node(node);
node             1062 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 		of_node_put(node);
node             1065 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c 	of_node_put(node);
node               40 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c int mtk_mdp_comp_get_id(struct device *dev, struct device_node *node,
node               43 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 	int id = of_alias_get_id(node, mtk_mdp_comp_stem[comp_type]);
node               94 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c int mtk_mdp_comp_init(struct device *dev, struct device_node *node,
node              106 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 	comp->dev_node = of_node_get(node);
node              109 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 	comp->regs = of_iomap(node, 0);
node              112 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 		comp->clk[i] = of_clk_get(node, i);
node              126 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 	larb_node = of_parse_phandle(node, "mediatek,larb", 0);
node              129 drivers/media/platform/mtk-mdp/mtk_mdp_comp.c 			"Missing mediadek,larb phandle in %pOF node\n", node);
node               55 drivers/media/platform/mtk-mdp/mtk_mdp_comp.h int mtk_mdp_comp_init(struct device *dev, struct device_node *node,
node               58 drivers/media/platform/mtk-mdp/mtk_mdp_comp.h int mtk_mdp_comp_get_id(struct device *dev, struct device_node *node,
node               98 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 	struct device_node *node, *parent;
node              113 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 	node = of_get_next_child(dev->of_node, NULL);
node              114 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 	if (node) {
node              115 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 		of_node_put(node);
node              123 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 	for_each_child_of_node(parent, node) {
node              129 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 		of_id = of_match_node(mtk_mdp_comp_dt_ids, node);
node              133 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 		if (!of_device_is_available(node)) {
node              135 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 				node);
node              140 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 		comp_id = mtk_mdp_comp_get_id(dev, node, comp_type);
node              143 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 				 node);
node              150 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 			of_node_put(node);
node              155 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 		ret = mtk_mdp_comp_init(dev, node, comp, comp_id);
node              157 drivers/media/platform/mtk-mdp/mtk_mdp_core.c 			of_node_put(node);
node               19 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c 	struct device_node *node;
node               30 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c 	node = of_parse_phandle(pdev->dev.of_node, "mediatek,larb", 0);
node               31 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c 	if (!node) {
node               36 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c 	pdev = of_find_device_by_node(node);
node               37 drivers/media/platform/mtk-vcodec/mtk_vcodec_dec_pm.c 	of_node_put(node);
node               20 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	struct device_node *node;
node               36 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	node = of_parse_phandle(dev->of_node, "mediatek,larb", 0);
node               37 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	if (!node) {
node               41 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	pdev = of_find_device_by_node(node);
node               42 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	of_node_put(node);
node               49 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	node = of_parse_phandle(dev->of_node, "mediatek,larb", 1);
node               50 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	if (!node) {
node               55 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	pdev = of_find_device_by_node(node);
node               56 drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_pm.c 	of_node_put(node);
node              295 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	struct vdec_fb_node *node;
node              302 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_for_each_entry(node, &inst->fb_use_list, list) {
node              303 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 			struct vdec_fb *fb = (struct vdec_fb *)node->fb;
node              306 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 				list_move_tail(&node->list,
node              314 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	node = list_first_entry(&inst->available_fb_node_list,
node              316 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	node->fb = inst->cur_fb;
node              317 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	list_move_tail(&node->list, &inst->fb_use_list);
node              321 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		node = list_first_entry(&inst->available_fb_node_list,
node              323 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		node->fb = inst->cur_fb;
node              324 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_move_tail(&node->list, &inst->fb_disp_list);
node              330 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	struct vdec_fb_node *node, *tmp;
node              332 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	list_for_each_entry_safe(node, tmp, &inst->fb_use_list, list)
node              333 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_move_tail(&node->list, &inst->fb_free_list);
node              355 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	struct vdec_fb_node *node;
node              358 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		node = list_first_entry(&inst->available_fb_node_list,
node              360 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		node->fb = fb;
node              361 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_move_tail(&node->list, &inst->fb_free_list);
node              518 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	struct vdec_fb_node *node;
node              521 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	node = list_first_entry_or_null(&inst->fb_disp_list,
node              523 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	if (node) {
node              524 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_move_tail(&node->list, &inst->available_fb_node_list);
node              525 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		fb = (struct vdec_fb *)node->fb;
node              528 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 				 node->fb, fb->status);
node              539 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	struct vdec_fb_node *node;
node              542 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	node = list_first_entry_or_null(&inst->fb_free_list,
node              544 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 	if (node) {
node              545 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		list_move_tail(&node->list, &inst->available_fb_node_list);
node              546 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 		fb = (struct vdec_fb *)node->fb;
node              549 drivers/media/platform/mtk-vcodec/vdec/vdec_vp8_if.c 				 node->fb, fb->status);
node              218 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              220 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	list_for_each_entry(node, &inst->fb_use_list, list) {
node              221 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		fb = (struct vdec_fb *)node->fb;
node              223 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 			list_move_tail(&node->list,
node              234 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              237 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		node = list_first_entry_or_null(&inst->available_fb_node_list,
node              240 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		if (node) {
node              241 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 			node->fb = fb;
node              242 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 			list_move_tail(&node->list, &inst->fb_free_list);
node              432 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              439 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	node = list_first_entry_or_null(&inst->available_fb_node_list,
node              441 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	if (node) {
node              442 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		node->fb = fb;
node              443 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		list_move_tail(&node->list, &inst->fb_disp_list);
node              614 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              617 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	node = list_first_entry_or_null(&inst->fb_disp_list,
node              619 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	if (node) {
node              620 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		fb = (struct vdec_fb *)node->fb;
node              622 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		list_move_tail(&node->list, &inst->available_fb_node_list);
node              624 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 				 node->fb, fb->status);
node              634 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              641 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	node = list_first_entry_or_null(&inst->available_fb_node_list,
node              643 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	if (node) {
node              644 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		node->fb = fb;
node              645 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		list_move_tail(&node->list, &inst->fb_use_list);
node              655 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node, *tmp;
node              657 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	list_for_each_entry_safe(node, tmp, &inst->fb_use_list, list)
node              658 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		list_move_tail(&node->list, &inst->fb_free_list);
node              722 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	struct vdec_fb_node *node;
node              725 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	node = list_first_entry_or_null(&inst->fb_free_list,
node              727 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	if (node) {
node              728 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		list_move_tail(&node->list, &inst->available_fb_node_list);
node              729 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 		fb = (struct vdec_fb *)node->fb;
node              732 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 				 node->fb, fb->status);
node              427 drivers/media/platform/qcom/camss/camss.c 					struct device_node *node,
node              435 drivers/media/platform/qcom/camss/camss.c 	v4l2_fwnode_endpoint_parse(of_fwnode_handle(node), &vep);
node              468 drivers/media/platform/qcom/camss/camss.c 	struct device_node *node = NULL;
node              472 drivers/media/platform/qcom/camss/camss.c 	for_each_endpoint_of_node(dev->of_node, node) {
node              476 drivers/media/platform/qcom/camss/camss.c 		if (!of_device_is_available(node))
node              479 drivers/media/platform/qcom/camss/camss.c 		remote = of_graph_get_remote_port_parent(node);
node              497 drivers/media/platform/qcom/camss/camss.c 		ret = camss_of_parse_endpoint_node(dev, node, csd);
node              508 drivers/media/platform/qcom/camss/camss.c 	of_node_put(node);
node               68 drivers/media/platform/qcom/venus/firmware.c 	struct device_node *node;
node               79 drivers/media/platform/qcom/venus/firmware.c 	node = of_parse_phandle(dev->of_node, "memory-region", 0);
node               80 drivers/media/platform/qcom/venus/firmware.c 	if (!node) {
node               85 drivers/media/platform/qcom/venus/firmware.c 	ret = of_address_to_resource(node, 0, &r);
node              126 drivers/media/platform/qcom/venus/firmware.c 	of_node_put(node);
node              976 drivers/media/platform/rcar-vin/rcar-dma.c 	struct rvin_buffer *buf, *node;
node              987 drivers/media/platform/rcar-vin/rcar-dma.c 	list_for_each_entry_safe(buf, node, &vin->buf_list, list) {
node              290 drivers/media/platform/sh_vou.c 	struct sh_vou_buffer *buf, *node;
node              297 drivers/media/platform/sh_vou.c 		list_for_each_entry_safe(buf, node, &vou_dev->buf_list, list) {
node              338 drivers/media/platform/sh_vou.c 	struct sh_vou_buffer *buf, *node;
node              349 drivers/media/platform/sh_vou.c 	list_for_each_entry_safe(buf, node, &vou_dev->buf_list, list) {
node              322 drivers/media/platform/sti/bdisp/bdisp-debug.c 	struct bdisp_node *node;
node              331 drivers/media/platform/sti/bdisp/bdisp-debug.c 		node = bdisp->dbg.copy_node[i];
node              332 drivers/media/platform/sti/bdisp/bdisp-debug.c 		if (!node)
node              336 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "NIP\t0x%08X\n", node->nip);
node              337 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "CIC\t0x%08X\n", node->cic);
node              338 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_ins(s, node->ins);
node              339 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "ACK\t0x%08X\n", node->ack);
node              341 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "TBA\t0x%08X\n", node->tba);
node              342 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_tty(s, node->tty);
node              343 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_xy(s, node->txy, "TXY");
node              344 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sz(s, node->tsz, "TSZ");
node              347 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "S1BA\t0x%08X\n", node->s1ba);
node              348 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sty(s, node->s1ty, node->s1ba, "S1TY");
node              349 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_xy(s, node->s1xy, "S1XY");
node              351 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "S2BA\t0x%08X\n", node->s2ba);
node              352 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sty(s, node->s2ty, node->s2ba, "S2TY");
node              353 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_xy(s, node->s2xy, "S2XY");
node              354 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sz(s, node->s2sz, "S2SZ");
node              356 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "S3BA\t0x%08X\n", node->s3ba);
node              357 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sty(s, node->s3ty, node->s3ba, "S3TY");
node              358 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_xy(s, node->s3xy, "S3XY");
node              359 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_sz(s, node->s3sz, "S3SZ");
node              363 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_fctl(s, node->fctl);
node              366 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_rsf(s, node->rsf, "RSF");
node              367 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_rzi(s, node->rzi, "RZI");
node              368 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "HFP\t0x%08X\n", node->hfp);
node              369 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "VFP\t0x%08X\n", node->vfp);
node              371 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_rsf(s, node->y_rsf, "Y_RSF");
node              372 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_rzi(s, node->y_rzi, "Y_RZI");
node              373 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "Y_HFP\t0x%08X\n", node->y_hfp);
node              374 drivers/media/platform/sti/bdisp/bdisp-debug.c 		seq_printf(s, "Y_VFP\t0x%08X\n", node->y_vfp);
node              380 drivers/media/platform/sti/bdisp/bdisp-debug.c 		bdisp_dbg_dump_ivmx(s, node->ivmx0, node->ivmx1,
node              381 drivers/media/platform/sti/bdisp/bdisp-debug.c 				    node->ivmx2, node->ivmx3);
node              386 drivers/media/platform/sti/bdisp/bdisp-debug.c 	} while ((++i < MAX_NB_NODE) && node->nip);
node              394 drivers/media/platform/sti/bdisp/bdisp-debug.c 	struct bdisp_node *node;
node              404 drivers/media/platform/sti/bdisp/bdisp-debug.c 		node = bdisp->dbg.copy_node[i];
node              405 drivers/media/platform/sti/bdisp/bdisp-debug.c 		if (!node)
node              409 drivers/media/platform/sti/bdisp/bdisp-debug.c 		val = (u32 *)node;
node              412 drivers/media/platform/sti/bdisp/bdisp-debug.c 	} while ((++i < MAX_NB_NODE) && node->nip);
node              433 drivers/media/platform/sti/bdisp/bdisp-hw.c 	if (ctx && ctx->node[0])
node              436 drivers/media/platform/sti/bdisp/bdisp-hw.c 			       ctx->node[0], ctx->node_paddr[0],
node              467 drivers/media/platform/sti/bdisp/bdisp-hw.c 		ctx->node[i] = base;
node              469 drivers/media/platform/sti/bdisp/bdisp-hw.c 		dev_dbg(dev, "node[%d]=0x%p (paddr=%pad)\n", i, ctx->node[i],
node              759 drivers/media/platform/sti/bdisp/bdisp-hw.c 				struct bdisp_node *node,
node              774 drivers/media/platform/sti/bdisp/bdisp-hw.c 	memset(node, 0, sizeof(*node));
node              789 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->nip = 0;
node              790 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->cic = BLT_CIC_ALL_GRP;
node              791 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->ack = BLT_ACK_BYPASS_S2S3;
node              796 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ins = BLT_INS_S1_OFF | BLT_INS_S2_MEM | BLT_INS_S3_OFF;
node              802 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ins = BLT_INS_S1_OFF | BLT_INS_S3_MEM;
node              804 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->ins |= BLT_INS_S2_CF;
node              806 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->ins |= BLT_INS_S2_MEM;
node              813 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ins = BLT_INS_S3_MEM;
node              815 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->ins |= BLT_INS_S2_CF | BLT_INS_S1_CF;
node              817 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->ins |= BLT_INS_S2_MEM | BLT_INS_S1_MEM;
node              822 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->ins |= cfg->cconv ? BLT_INS_IVMX : 0;
node              824 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->ins |= (cfg->scale || cfg->src_420 || cfg->dst_420) ?
node              828 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tba = (t_plan == BDISP_CBCR) ? dst->paddr[1] : dst->paddr[0];
node              830 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty = dst->bytesperline;
node              831 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty |= bdisp_hw_color_format(dst_fmt);
node              832 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty |= BLT_TTY_DITHER;
node              833 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty |= (t_plan == BDISP_CBCR) ? BLT_TTY_CHROMA : 0;
node              834 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty |= cfg->hflip ? BLT_TTY_HSO : 0;
node              835 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tty |= cfg->vflip ? BLT_TTY_VSO : 0;
node              847 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->txy = cfg->vflip ? (dst_rect.height - 1) : dst_rect.top;
node              848 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->txy <<= 16;
node              849 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->txy |= cfg->hflip ? (dst_width - dst_x_offset - 1) :
node              852 drivers/media/platform/sti/bdisp/bdisp-hw.c 	node->tsz = dst_rect.height << 16 | dst_rect.width;
node              862 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ba = src->paddr[0];
node              864 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ty = src->bytesperline;
node              866 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s2ty *= 2;
node              868 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ty |= bdisp_hw_color_format(src_fmt);
node              870 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2xy = src_rect.top << 16 | src_rect.left;
node              871 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2sz = src_rect.height << 16 | src_rect.width;
node              882 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ba = src->paddr[1];
node              884 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ty = src->bytesperline;
node              886 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s2ty /= 2;
node              888 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s2ty *= 2;
node              890 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2ty |= bdisp_hw_color_format(src_fmt);
node              892 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2xy = src_rect.top << 16 | src_rect.left;
node              893 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s2sz = src_rect.height << 16 | src_rect.width;
node              897 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s1ba = src->paddr[2];
node              899 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s1ty = node->s2ty;
node              900 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s1xy = node->s2xy;
node              904 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s3ba = src->paddr[0];
node              906 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s3ty = src->bytesperline;
node              908 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3ty *= 2;
node              909 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->s3ty |= bdisp_hw_color_format(src_fmt);
node              913 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3xy = node->s2xy * 2;
node              914 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3sz = node->s2sz * 2;
node              917 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3ty |= BLT_S3TY_BLANK_ACC;
node              918 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3xy = node->s2xy;
node              919 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->s3sz = node->s2sz;
node              924 drivers/media/platform/sti/bdisp/bdisp-hw.c 	if (node->ins & BLT_INS_SCALE) {
node              930 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->fctl = BLT_FCTL_HV_SCALE;
node              932 drivers/media/platform/sti/bdisp/bdisp-hw.c 				node->fctl |= BLT_FCTL_Y_HV_SCALE;
node              934 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->fctl = BLT_FCTL_HV_SAMPLE;
node              936 drivers/media/platform/sti/bdisp/bdisp-hw.c 				node->fctl |= BLT_FCTL_Y_HV_SAMPLE;
node              951 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->rsf = v_inc << 16 | h_inc;
node              954 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->rzi = BLT_RZI_DEFAULT;
node              957 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->hfp = bdisp_hw_get_hf_addr(h_inc);
node              958 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->vfp = bdisp_hw_get_vf_addr(v_inc);
node              965 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->y_rsf = yv_inc << 16 | yh_inc;
node              966 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->y_rzi = BLT_RZI_DEFAULT;
node              967 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->y_hfp = bdisp_hw_get_hf_addr(yh_inc);
node              968 drivers/media/platform/sti/bdisp/bdisp-hw.c 			node->y_vfp = bdisp_hw_get_vf_addr(yv_inc);
node              976 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ivmx0 = ivmx[0];
node              977 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ivmx1 = ivmx[1];
node              978 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ivmx2 = ivmx[2];
node              979 drivers/media/platform/sti/bdisp/bdisp-hw.c 		node->ivmx3 = ivmx[3];
node              999 drivers/media/platform/sti/bdisp/bdisp-hw.c 		if (!ctx->node[i]) {
node             1011 drivers/media/platform/sti/bdisp/bdisp-hw.c 		bdisp_hw_build_node(ctx, &cfg, ctx->node[nid],
node             1015 drivers/media/platform/sti/bdisp/bdisp-hw.c 			ctx->node[nid - 1]->nip = ctx->node_paddr[nid];
node             1020 drivers/media/platform/sti/bdisp/bdisp-hw.c 			bdisp_hw_build_node(ctx, &cfg, ctx->node[nid],
node             1022 drivers/media/platform/sti/bdisp/bdisp-hw.c 			ctx->node[nid - 1]->nip = ctx->node_paddr[nid];
node             1033 drivers/media/platform/sti/bdisp/bdisp-hw.c 	ctx->node[nid - 1]->nip = 0;
node             1051 drivers/media/platform/sti/bdisp/bdisp-hw.c 	struct bdisp_node **node = ctx->node;
node             1071 drivers/media/platform/sti/bdisp/bdisp-hw.c 		*copy_node[i] = *node[i];
node             1112 drivers/media/platform/sti/bdisp/bdisp-hw.c 		if (!ctx->node[node_id]->nip)
node              120 drivers/media/platform/sti/bdisp/bdisp.h 	struct bdisp_node       *node[MAX_NB_NODE];
node              728 drivers/media/platform/stm32/stm32-dcmi.c 	struct dcmi_buf *buf, *node;
node              847 drivers/media/platform/stm32/stm32-dcmi.c 	list_for_each_entry_safe(buf, node, &dcmi->buffers, list) {
node              860 drivers/media/platform/stm32/stm32-dcmi.c 	struct dcmi_buf *buf, *node;
node              875 drivers/media/platform/stm32/stm32-dcmi.c 	list_for_each_entry_safe(buf, node, &dcmi->buffers, list) {
node             1778 drivers/media/platform/stm32/stm32-dcmi.c static int dcmi_graph_parse(struct stm32_dcmi *dcmi, struct device_node *node)
node             1783 drivers/media/platform/stm32/stm32-dcmi.c 	ep = of_graph_get_next_endpoint(node, ep);
node              207 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	struct sun4i_csi_buffer *buf, *node;
node              210 drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c 	list_for_each_entry_safe(buf, node, &csi->buf_list, list) {
node              222 drivers/media/platform/vimc/vimc-capture.c 	struct vimc_cap_buffer *vbuf, *node;
node              226 drivers/media/platform/vimc/vimc-capture.c 	list_for_each_entry_safe(vbuf, node, &vcap->buf_list, list) {
node             1228 drivers/media/platform/vivid/vivid-core.c 				dev->fb_info.node);
node             1625 drivers/media/platform/vivid/vivid-core.c 				dev->fb_info.node);
node              312 drivers/media/platform/vivid/vivid-osd.c 	dev->fb_info.node = -1;
node              711 drivers/media/platform/xilinx/xilinx-tpg.c 	struct device_node *node = xtpg->xvip.dev->of_node;
node              717 drivers/media/platform/xilinx/xilinx-tpg.c 	ports = of_get_child_by_name(node, "ports");
node              719 drivers/media/platform/xilinx/xilinx-tpg.c 		ports = node;
node              102 drivers/media/platform/xilinx/xilinx-vip.c const struct xvip_video_format *xvip_of_get_format(struct device_node *node)
node              110 drivers/media/platform/xilinx/xilinx-vip.c 	ret = of_property_read_u32(node, "xlnx,video-format", &vf_code);
node              114 drivers/media/platform/xilinx/xilinx-vip.c 	ret = of_property_read_u32(node, "xlnx,video-width", &width);
node              119 drivers/media/platform/xilinx/xilinx-vip.c 		of_property_read_string(node, "xlnx,cfa-pattern", &pattern);
node              124 drivers/media/platform/xilinx/xilinx-vip.h const struct xvip_video_format *xvip_of_get_format(struct device_node *node);
node              188 drivers/media/platform/xilinx/xilinx-vipp.c 	struct device_node *node = xdev->dev->of_node;
node              203 drivers/media/platform/xilinx/xilinx-vipp.c 		ep = of_graph_get_next_endpoint(node, ep);
node              432 drivers/media/platform/xilinx/xilinx-vipp.c 				   struct device_node *node)
node              440 drivers/media/platform/xilinx/xilinx-vipp.c 	ret = of_property_read_string(node, "direction", &direction);
node              451 drivers/media/platform/xilinx/xilinx-vipp.c 	of_property_read_u32(node, "reg", &index);
node              459 drivers/media/platform/xilinx/xilinx-vipp.c 		dev_err(xdev->dev, "%pOF initialization failed\n", node);
node              311 drivers/media/platform/xilinx/xilinx-vtc.c 	struct device_node *node = xvtc->xvip.dev->of_node;
node              313 drivers/media/platform/xilinx/xilinx-vtc.c 	xvtc->has_detector = of_property_read_bool(node, "xlnx,detector");
node              314 drivers/media/platform/xilinx/xilinx-vtc.c 	xvtc->has_generator = of_property_read_bool(node, "xlnx,generator");
node              214 drivers/media/rc/ir-hix5hd2.c 	struct device_node *node = pdev->dev.of_node;
node              222 drivers/media/rc/ir-hix5hd2.c 	priv->regmap = syscon_regmap_lookup_by_phandle(node,
node              258 drivers/media/rc/ir-hix5hd2.c 	map_name = of_get_property(node, "linux,rc-map-name", NULL);
node              104 drivers/media/rc/meson-ir.c 	struct device_node *node = dev->of_node;
node              133 drivers/media/rc/meson-ir.c 	map_name = of_get_property(node, "linux,rc-map-name", NULL);
node              162 drivers/media/rc/meson-ir.c 	if (of_device_is_compatible(node, "amlogic,meson6-ir"))
node              198 drivers/media/rc/meson-ir.c 	struct device_node *node = dev->of_node;
node              208 drivers/media/rc/meson-ir.c 	if (of_device_is_compatible(node, "amlogic,meson6-ir"))
node             1340 drivers/media/usb/gspca/gspca.c 	struct gspca_buffer *buf, *node;
node             1344 drivers/media/usb/gspca/gspca.c 	list_for_each_entry_safe(buf, node, &gspca_dev->buf_list, list) {
node              731 drivers/media/usb/hackrf/hackrf.c 	struct hackrf_buffer *buffer, *node;
node              743 drivers/media/usb/hackrf/hackrf.c 	list_for_each_entry_safe(buffer, node, buffer_list, list) {
node             1083 drivers/media/usb/s2255/s2255drv.c 	struct s2255_buffer *buf, *node;
node             1087 drivers/media/usb/s2255/s2255drv.c 	list_for_each_entry_safe(buf, node, &vc->buf_list, list) {
node             1243 drivers/media/usb/uvc/uvc_ctrl.c 	list_for_each_entry(sev, &mapping->ev_subs, node) {
node             1419 drivers/media/usb/uvc/uvc_ctrl.c 	list_add_tail(&sev->node, &mapping->ev_subs);
node             1446 drivers/media/usb/uvc/uvc_ctrl.c 	list_del(&sev->node);
node             1494 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(sev, &ctrl->ev_subs, node)
node             2111 drivers/media/v4l2-core/v4l2-ctrls.c static inline u32 node2id(struct list_head *node)
node             2113 drivers/media/v4l2-core/v4l2-ctrls.c 	return list_entry(node, struct v4l2_ctrl_ref, node)->ctrl->id;
node             2167 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry_safe(ref, next_ref, &hdl->ctrl_refs, node) {
node             2168 drivers/media/v4l2-core/v4l2-ctrls.c 		list_del(&ref->node);
node             2172 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry_safe(ctrl, next_ctrl, &hdl->ctrls, node) {
node             2173 drivers/media/v4l2-core/v4l2-ctrls.c 		list_del(&ctrl->node);
node             2174 drivers/media/v4l2-core/v4l2-ctrls.c 		list_for_each_entry_safe(sev, next_sev, &ctrl->ev_subs, node)
node             2175 drivers/media/v4l2-core/v4l2-ctrls.c 			list_del(&sev->node);
node             2199 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &hdl->ctrl_refs, node) {
node             2302 drivers/media/v4l2-core/v4l2-ctrls.c 	INIT_LIST_HEAD(&new_ref->node);
node             2311 drivers/media/v4l2-core/v4l2-ctrls.c 		list_add_tail(&new_ref->node, &hdl->ctrl_refs);
node             2316 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &hdl->ctrl_refs, node) {
node             2324 drivers/media/v4l2-core/v4l2-ctrls.c 		list_add(&new_ref->node, ref->node.prev);
node             2469 drivers/media/v4l2-core/v4l2-ctrls.c 	INIT_LIST_HEAD(&ctrl->node);
node             2516 drivers/media/v4l2-core/v4l2-ctrls.c 	list_add_tail(&ctrl->node, &hdl->ctrls);
node             2691 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &add->ctrl_refs, node) {
node             2864 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ctrl, &hdl->ctrls, node)
node             2889 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ctrl, &hdl->ctrls, node)
node             2892 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ctrl, &hdl->ctrls, node) {
node             2971 drivers/media/v4l2-core/v4l2-ctrls.c 			list_for_each_entry_continue(ref, &hdl->ctrl_refs, node) {
node             2978 drivers/media/v4l2-core/v4l2-ctrls.c 			if (&ref->node == &hdl->ctrl_refs)
node             2985 drivers/media/v4l2-core/v4l2-ctrls.c 			list_for_each_entry(ref, &hdl->ctrl_refs, node) {
node             2992 drivers/media/v4l2-core/v4l2-ctrls.c 			if (&ref->node == &hdl->ctrl_refs)
node             3121 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &from->ctrl_refs, node) {
node             3160 drivers/media/v4l2-core/v4l2-ctrls.c 					 struct v4l2_ctrl_ref, node);
node             3161 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref_ctrl, &hdl->ctrl_refs, node) {
node             3166 drivers/media/v4l2-core/v4l2-ctrls.c 			if (list_is_last(&ref_ctrl_prev->node,
node             3169 drivers/media/v4l2-core/v4l2-ctrls.c 			ref_ctrl_prev = list_next_entry(ref_ctrl_prev, node);
node             4108 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &hdl->ctrl_refs, node) {
node             4174 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &hdl->ctrl_refs, node)
node             4177 drivers/media/v4l2-core/v4l2-ctrls.c 	list_for_each_entry(ref, &hdl->ctrl_refs, node) {
node             4330 drivers/media/v4l2-core/v4l2-ctrls.c 	list_add_tail(&sev->node, &ctrl->ev_subs);
node             4356 drivers/media/v4l2-core/v4l2-ctrls.c 	list_del(&sev->node);
node               30 drivers/memory/atmel-ebi.c 	struct list_head node;
node              369 drivers/memory/atmel-ebi.c 	list_add_tail(&ebid->node, &ebi->devs);
node              617 drivers/memory/atmel-ebi.c 	list_for_each_entry(ebid, &ebi->devs, node) {
node              105 drivers/memory/da8xx-ddrctl.c 	struct device_node *node;
node              112 drivers/memory/da8xx-ddrctl.c 	node = dev->of_node;
node               60 drivers/memory/emif.c 	struct list_head		node;
node              326 drivers/memory/emif.c 	list_for_each_entry(emif, &device_list, node) {
node              337 drivers/memory/emif.c 	list_for_each_entry(emif, &device_list, node) {
node             1553 drivers/memory/emif.c 	list_add(&emif->node, &device_list);
node             1813 drivers/memory/emif.c 	list_for_each_entry(emif, &device_list, node)
node             1881 drivers/memory/emif.c 	list_for_each_entry(emif, &device_list, node)
node             1905 drivers/memory/emif.c 	list_for_each_entry(emif, &device_list, node)
node              158 drivers/memory/jz4780-nemc.c 				       struct device_node *node)
node              188 drivers/memory/jz4780-nemc.c 	if (!of_property_read_u32(node, "ingenic,nemc-bus-width", &val)) {
node              204 drivers/memory/jz4780-nemc.c 	if (of_property_read_u32(node, "ingenic,nemc-tAS", &val) == 0) {
node              216 drivers/memory/jz4780-nemc.c 	if (of_property_read_u32(node, "ingenic,nemc-tAH", &val) == 0) {
node              228 drivers/memory/jz4780-nemc.c 	if (of_property_read_u32(node, "ingenic,nemc-tBP", &val) == 0) {
node              240 drivers/memory/jz4780-nemc.c 	if (of_property_read_u32(node, "ingenic,nemc-tAW", &val) == 0) {
node              252 drivers/memory/jz4780-nemc.c 	if (of_property_read_u32(node, "ingenic,nemc-tSTRV", &val) == 0) {
node               87 drivers/memory/mvebu-devbus.c 			       struct device_node *node,
node               94 drivers/memory/mvebu-devbus.c 	err = of_property_read_u32(node, name, &time_ps);
node               97 drivers/memory/mvebu-devbus.c 			node, name);
node              109 drivers/memory/mvebu-devbus.c 				    struct device_node *node,
node              115 drivers/memory/mvebu-devbus.c 	err = of_property_read_u32(node, "devbus,bus-width", &r->bus_width);
node              119 drivers/memory/mvebu-devbus.c 			node);
node              136 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,badr-skew-ps",
node              141 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,turn-off-ps",
node              146 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,acc-first-ps",
node              151 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,acc-next-ps",
node              157 drivers/memory/mvebu-devbus.c 		err = get_timing_param_ps(devbus, node, "devbus,rd-setup-ps",
node              162 drivers/memory/mvebu-devbus.c 		err = get_timing_param_ps(devbus, node, "devbus,rd-hold-ps",
node              167 drivers/memory/mvebu-devbus.c 		err = of_property_read_u32(node, "devbus,sync-enable",
node              172 drivers/memory/mvebu-devbus.c 				node);
node              177 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,ale-wr-ps",
node              182 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,wr-low-ps",
node              187 drivers/memory/mvebu-devbus.c 	err = get_timing_param_ps(devbus, node, "devbus,wr-high-ps",
node              196 drivers/memory/mvebu-devbus.c 					  struct device_node *node,
node              229 drivers/memory/mvebu-devbus.c 					   struct device_node *node,
node              266 drivers/memory/mvebu-devbus.c 	struct device_node *node = pdev->dev.of_node;
node              301 drivers/memory/mvebu-devbus.c 	if (!of_property_read_bool(node, "devbus,keep-config")) {
node              303 drivers/memory/mvebu-devbus.c 		err = devbus_get_timing_params(devbus, node, &r, &w);
node              308 drivers/memory/mvebu-devbus.c 		if (of_device_is_compatible(node, "marvell,orion-devbus"))
node              309 drivers/memory/mvebu-devbus.c 			devbus_orion_set_timing_params(devbus, node, &r, &w);
node              311 drivers/memory/mvebu-devbus.c 			devbus_armada_set_timing_params(devbus, node, &r, &w);
node              319 drivers/memory/mvebu-devbus.c 	err = of_platform_populate(node, NULL, NULL, dev);
node              345 drivers/memory/tegra/mc.c 			   struct device_node *node)
node              350 drivers/memory/tegra/mc.c 	err = of_property_read_u32(node, "clock-frequency", &tmp);
node              353 drivers/memory/tegra/mc.c 			"timing %pOFn: failed to read rate\n", node);
node              363 drivers/memory/tegra/mc.c 	err = of_property_read_u32_array(node, "nvidia,emem-configuration",
node              369 drivers/memory/tegra/mc.c 			node);
node              376 drivers/memory/tegra/mc.c static int load_timings(struct tegra_mc *mc, struct device_node *node)
node              380 drivers/memory/tegra/mc.c 	int child_count = of_get_child_count(node);
node              390 drivers/memory/tegra/mc.c 	for_each_child_of_node(node, child) {
node              405 drivers/memory/tegra/mc.c 	struct device_node *node;
node              413 drivers/memory/tegra/mc.c 	for_each_child_of_node(mc->dev->of_node, node) {
node              414 drivers/memory/tegra/mc.c 		err = of_property_read_u32(node, "nvidia,ram-code",
node              419 drivers/memory/tegra/mc.c 		err = load_timings(mc, node);
node              420 drivers/memory/tegra/mc.c 		of_node_put(node);
node              876 drivers/memory/tegra/tegra124-emc.c 				   struct device_node *node)
node              881 drivers/memory/tegra/tegra124-emc.c 	err = of_property_read_u32(node, "clock-frequency", &value);
node              884 drivers/memory/tegra/tegra124-emc.c 			node, err);
node              890 drivers/memory/tegra/tegra124-emc.c 	err = of_property_read_u32_array(node, "nvidia,emc-configuration",
node              896 drivers/memory/tegra/tegra124-emc.c 			node, err);
node              901 drivers/memory/tegra/tegra124-emc.c 	err = of_property_read_u32(node, dtprop, &timing->prop); \
node              904 drivers/memory/tegra/tegra124-emc.c 			node, err); \
node              946 drivers/memory/tegra/tegra124-emc.c 					  struct device_node *node)
node              948 drivers/memory/tegra/tegra124-emc.c 	int child_count = of_get_child_count(node);
node              961 drivers/memory/tegra/tegra124-emc.c 	for_each_child_of_node(node, child) {
node              983 drivers/memory/tegra/tegra124-emc.c tegra_emc_find_node_by_ram_code(struct device_node *node, u32 ram_code)
node              988 drivers/memory/tegra/tegra124-emc.c 	for_each_child_of_node(node, np) {
node              280 drivers/memory/tegra/tegra20-emc.c 				   struct device_node *node)
node              285 drivers/memory/tegra/tegra20-emc.c 	if (!of_device_is_compatible(node, "nvidia,tegra20-emc-table")) {
node              286 drivers/memory/tegra/tegra20-emc.c 		dev_err(emc->dev, "incompatible DT node: %pOF\n", node);
node              290 drivers/memory/tegra/tegra20-emc.c 	err = of_property_read_u32(node, "clock-frequency", &rate);
node              293 drivers/memory/tegra/tegra20-emc.c 			node, err);
node              297 drivers/memory/tegra/tegra20-emc.c 	err = of_property_read_u32_array(node, "nvidia,emc-registers",
node              303 drivers/memory/tegra/tegra20-emc.c 			node, err);
node              314 drivers/memory/tegra/tegra20-emc.c 		__func__, node, timing->rate);
node              334 drivers/memory/tegra/tegra20-emc.c 					  struct device_node *node)
node              341 drivers/memory/tegra/tegra20-emc.c 	child_count = of_get_child_count(node);
node              343 drivers/memory/tegra/tegra20-emc.c 		dev_err(emc->dev, "no memory timings in DT node: %pOF\n", node);
node              355 drivers/memory/tegra/tegra20-emc.c 	for_each_child_of_node(node, child) {
node              576 drivers/mfd/88pm860x-core.c 	struct device_node *node = i2c->dev.of_node;
node              627 drivers/mfd/88pm860x-core.c 	irq_domain_add_legacy(node, nr_irqs, chip->irq_base, 0,
node             1135 drivers/mfd/88pm860x-core.c 	struct device_node *node = client->dev.of_node;
node             1139 drivers/mfd/88pm860x-core.c 	if (node && !pdata) {
node             1146 drivers/mfd/88pm860x-core.c 		ret = pm860x_dt_init(node, &client->dev, pdata);
node              160 drivers/mfd/ab8500-gpadc.c 	struct list_head node;
node              180 drivers/mfd/ab8500-gpadc.c 	list_for_each_entry(gpadc, &ab8500_gpadc_list, node) {
node              998 drivers/mfd/ab8500-gpadc.c 	list_add_tail(&gpadc->node, &ab8500_gpadc_list);
node             1016 drivers/mfd/ab8500-gpadc.c 	list_del(&gpadc->node);
node              341 drivers/mfd/cros_ec_dev.c 	struct device_node *node;
node              428 drivers/mfd/cros_ec_dev.c 	node = ec->ec_dev->dev->of_node;
node              429 drivers/mfd/cros_ec_dev.c 	if (of_property_read_bool(node, "google,has-vbc-nvram")) {
node               68 drivers/mfd/ipaq-micro.c 		list_add_tail(&msg->node, &micro->queue);
node              104 drivers/mfd/ipaq-micro.c 							node);
node              105 drivers/mfd/ipaq-micro.c 				list_del_init(&micro->msg->node);
node              666 drivers/mfd/max8925-core.c 	struct device_node *node = chip->dev->of_node;
node              692 drivers/mfd/max8925-core.c 	irq_domain_add_legacy(node, MAX8925_NR_IRQS, chip->irq_base, 0,
node              152 drivers/mfd/max8925-i2c.c 	struct device_node *node = client->dev.of_node;
node              154 drivers/mfd/max8925-i2c.c 	if (node && !pdata) {
node              162 drivers/mfd/max8925-i2c.c 		if (max8925_dt_init(node, &client->dev, pdata))
node              105 drivers/mfd/mfd-core.c 			list_for_each_entry(child, &parent->children, node) {
node              115 drivers/mfd/mfd-core.c 			list_for_each_entry(child, &parent->children, node) {
node              130 drivers/mfd/mxs-lradc.c 	struct device_node *node = dev->of_node;
node              159 drivers/mfd/mxs-lradc.c 	ret = of_property_read_u32(node, "fsl,lradc-touchscreen-wires",
node              476 drivers/mfd/omap-usb-host.c 	struct device_node *node = dev->of_node;
node              478 drivers/mfd/omap-usb-host.c 	ret = of_property_read_u32(node, "num-ports", &pdata->nports);
node              496 drivers/mfd/omap-usb-host.c 		ret = of_property_read_string(node, prop, &mode);
node              513 drivers/mfd/omap-usb-host.c 	pdata->single_ulpi_bypass = of_property_read_bool(node,
node              392 drivers/mfd/palmas.c 	struct device_node *node = i2c->dev.of_node;
node              396 drivers/mfd/palmas.c 	ret = of_property_read_u32(node, "ti,mux-pad1", &prop);
node              402 drivers/mfd/palmas.c 	ret = of_property_read_u32(node, "ti,mux-pad2", &prop);
node              409 drivers/mfd/palmas.c 	ret = of_property_read_u32(node, "ti,power-ctrl", &prop);
node              419 drivers/mfd/palmas.c 	pdata->pm_off = of_property_read_bool(node,
node              511 drivers/mfd/palmas.c 	struct device_node *node = i2c->dev.of_node;
node              519 drivers/mfd/palmas.c 	if (node && !pdata) {
node              560 drivers/mfd/palmas.c 			palmas->i2c_clients[i]->dev.of_node = of_node_get(node);
node              681 drivers/mfd/palmas.c 	if (node) {
node              127 drivers/mfd/ti_am335x_tscadc.c 	struct device_node	*node;
node              142 drivers/mfd/ti_am335x_tscadc.c 	node = of_get_child_by_name(pdev->dev.of_node, "tsc");
node              143 drivers/mfd/ti_am335x_tscadc.c 	of_property_read_u32(node, "ti,wires", &tsc_wires);
node              144 drivers/mfd/ti_am335x_tscadc.c 	of_property_read_u32(node, "ti,coordiante-readouts", &readouts);
node              146 drivers/mfd/ti_am335x_tscadc.c 	node = of_get_child_by_name(pdev->dev.of_node, "adc");
node              147 drivers/mfd/ti_am335x_tscadc.c 	of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) {
node             1071 drivers/mfd/twl-core.c 	struct device_node		*node = client->dev.of_node;
node             1078 drivers/mfd/twl-core.c 	if (!node && !pdata) {
node             1216 drivers/mfd/twl-core.c 	if (node) {
node             1219 drivers/mfd/twl-core.c 		status = of_platform_populate(node, NULL, twl_auxdata_lookup,
node              150 drivers/mfd/twl4030-audio.c 	struct device_node *node;
node              155 drivers/mfd/twl4030-audio.c 	node = of_get_child_by_name(parent, "codec");
node              156 drivers/mfd/twl4030-audio.c 	if (node) {
node              157 drivers/mfd/twl4030-audio.c 		of_node_put(node);
node              165 drivers/mfd/twl4030-audio.c 			      struct device_node *node)
node              172 drivers/mfd/twl4030-audio.c 	if (!of_property_read_u32(node, "ti,enable-vibra", &vibra) && vibra)
node              182 drivers/mfd/twl4030-audio.c 	struct device_node *node = pdev->dev.of_node;
node              187 drivers/mfd/twl4030-audio.c 	if (!pdata && !node) {
node              225 drivers/mfd/twl4030-audio.c 	if (twl4030_audio_has_codec(pdata, node)) {
node              234 drivers/mfd/twl4030-audio.c 	if (twl4030_audio_has_vibra(pdata, node)) {
node              678 drivers/mfd/twl4030-irq.c 	struct			device_node *node = dev->of_node;
node              693 drivers/mfd/twl4030-irq.c 	irq_domain_add_legacy(node, nr_irqs, irq_base, 0,
node              678 drivers/mfd/twl4030-power.c 					struct device_node *node)
node              683 drivers/mfd/twl4030-power.c 	if (of_property_read_bool(node, "ti,system-power-controller"))
node              686 drivers/mfd/twl4030-power.c 	if (of_property_read_bool(node, "ti,use_poweroff"))
node              885 drivers/mfd/twl4030-power.c 	struct device_node *node = pdev->dev.of_node;
node              891 drivers/mfd/twl4030-power.c 	if (!pdata && !node) {
node              926 drivers/mfd/twl4030-power.c 	if (twl4030_power_use_poweroff(pdata, node) && !pm_power_off) {
node              367 drivers/mfd/twl6030-irq.c 	struct			device_node *node = dev->of_node;
node              415 drivers/mfd/twl6030-irq.c 		irq_domain_add_linear(node, nr_irqs,
node               88 drivers/mfd/twl6040.c 	struct device_node *node;
node               90 drivers/mfd/twl6040.c 	node = of_get_child_by_name(parent, "vibra");
node               91 drivers/mfd/twl6040.c 	if (node) {
node               92 drivers/mfd/twl6040.c 		of_node_put(node);
node              639 drivers/mfd/twl6040.c 	struct device_node *node = client->dev.of_node;
node              644 drivers/mfd/twl6040.c 	if (!node) {
node              716 drivers/mfd/twl6040.c 		twl6040->audpwron = of_get_named_gpio(node,
node              771 drivers/mfd/twl6040.c 	if (twl6040_has_vibra(node)) {
node              588 drivers/mfd/ucb1x00-core.c 	list_add_tail(&ucb->node, &ucb1x00_devices);
node              589 drivers/mfd/ucb1x00-core.c 	list_for_each_entry(drv, &ucb1x00_drivers, node) {
node              618 drivers/mfd/ucb1x00-core.c 	list_del(&ucb->node);
node              642 drivers/mfd/ucb1x00-core.c 	list_add_tail(&drv->node, &ucb1x00_drivers);
node              643 drivers/mfd/ucb1x00-core.c 	list_for_each_entry(ucb, &ucb1x00_devices, node) {
node              655 drivers/mfd/ucb1x00-core.c 	list_del(&drv->node);
node               50 drivers/mfd/vexpress-sysreg.c 		struct device_node *node = of_find_compatible_node(NULL, NULL,
node               53 drivers/mfd/vexpress-sysreg.c 		base = of_iomap(node, 0);
node              220 drivers/mfd/vexpress-sysreg.c 	struct device_node *node;
node              223 drivers/mfd/vexpress-sysreg.c 	for_each_matching_node(node, vexpress_sysreg_match)
node              224 drivers/mfd/vexpress-sysreg.c 		of_platform_device_create(node, NULL, NULL);
node               50 drivers/misc/atmel_tclib.c 	list_for_each_entry(tc, &tc_list, node) {
node              173 drivers/misc/atmel_tclib.c 	list_add_tail(&tc->node, &tc_list);
node               49 drivers/misc/enclosure.c 	edev = list_prepare_entry(start, &container_list, node);
node               53 drivers/misc/enclosure.c 	list_for_each_entry_continue(edev, &container_list, node) {
node               91 drivers/misc/enclosure.c 	list_for_each_entry(edev, &container_list, node) {
node              142 drivers/misc/enclosure.c 	list_add_tail(&edev->node, &container_list);
node              166 drivers/misc/enclosure.c 	list_del(&edev->node);
node              131 drivers/misc/fastrpc.c 	struct list_head node;
node              135 drivers/misc/fastrpc.c 	struct list_head node;
node              159 drivers/misc/fastrpc.c 	struct list_head node; /* list of ctxs */
node              242 drivers/misc/fastrpc.c 	list_for_each_entry(map, &fl->maps, node) {
node              422 drivers/misc/fastrpc.c 	INIT_LIST_HEAD(&ctx->node);
node              458 drivers/misc/fastrpc.c 	list_add_tail(&ctx->node, &user->pending);
node              476 drivers/misc/fastrpc.c 	list_del(&ctx->node);
node              535 drivers/misc/fastrpc.c 	INIT_LIST_HEAD(&a->node);
node              539 drivers/misc/fastrpc.c 	list_add(&a->node, &buffer->attachments);
node              552 drivers/misc/fastrpc.c 	list_del(&a->node);
node              607 drivers/misc/fastrpc.c 	INIT_LIST_HEAD(&map->node);
node              637 drivers/misc/fastrpc.c 	list_add_tail(&map->node, &fl->maps);
node              959 drivers/misc/fastrpc.c 	list_del(&ctx->node);
node             1071 drivers/misc/fastrpc.c 		list_del(&map->node);
node             1145 drivers/misc/fastrpc.c 	list_for_each_entry_safe(ctx, n, &fl->pending, node) {
node             1146 drivers/misc/fastrpc.c 		list_del(&ctx->node);
node             1150 drivers/misc/fastrpc.c 	list_for_each_entry_safe(map, m, &fl->maps, node) {
node             1151 drivers/misc/fastrpc.c 		list_del(&map->node);
node             1458 drivers/misc/fastrpc.c 	list_for_each_entry(ctx, &user->pending, node)
node              191 drivers/misc/genwqe/card_dev.c 	struct list_head *node, *next;
node              196 drivers/misc/genwqe/card_dev.c 	list_for_each_safe(node, next, &cfile->map_list) {
node              197 drivers/misc/genwqe/card_dev.c 		dma_map = list_entry(node, struct dma_mapping, card_list);
node              229 drivers/misc/genwqe/card_dev.c 	struct list_head *node, *next;
node              233 drivers/misc/genwqe/card_dev.c 	list_for_each_safe(node, next, &cfile->pin_list) {
node              234 drivers/misc/genwqe/card_dev.c 		dma_map = list_entry(node, struct dma_mapping, pin_list);
node              253 drivers/misc/habanalabs/debugfs.c 		hash_for_each(ctx->mem_hash, i, hnode, node) {
node              912 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *node = inode->i_private;
node              914 drivers/misc/habanalabs/debugfs.c 	return single_open(file, node->info_ent->show, node);
node              920 drivers/misc/habanalabs/debugfs.c 	struct hl_debugfs_entry *node = file->f_inode->i_private;
node              922 drivers/misc/habanalabs/debugfs.c 	if (node->info_ent->write)
node              923 drivers/misc/habanalabs/debugfs.c 		return node->info_ent->write(file, buf, count, f_pos);
node               70 drivers/misc/habanalabs/habanalabs.h 	struct hlist_node	node;
node              818 drivers/misc/habanalabs/habanalabs.h 	struct hlist_node	node;
node              861 drivers/misc/habanalabs/habanalabs.h 	struct list_head	node;
node              359 drivers/misc/habanalabs/memory.c 	list_for_each_entry_safe(va_block, tmp, va_list, node) {
node              360 drivers/misc/habanalabs/memory.c 		list_del(&va_block->node);
node              384 drivers/misc/habanalabs/memory.c 	list_for_each_entry(va_block, va_list, node)
node              409 drivers/misc/habanalabs/memory.c 	prev = list_prev_entry(va_block, node);
node              410 drivers/misc/habanalabs/memory.c 	if (&prev->node != va_list && prev->end + 1 == va_block->start) {
node              413 drivers/misc/habanalabs/memory.c 		list_del(&va_block->node);
node              418 drivers/misc/habanalabs/memory.c 	next = list_next_entry(va_block, node);
node              419 drivers/misc/habanalabs/memory.c 	if (&next->node != va_list && va_block->end + 1 == next->start) {
node              422 drivers/misc/habanalabs/memory.c 		list_del(&va_block->node);
node              449 drivers/misc/habanalabs/memory.c 	list_for_each_entry(va_block, va_list, node) {
node              472 drivers/misc/habanalabs/memory.c 		list_add(&va_block->node, va_list);
node              474 drivers/misc/habanalabs/memory.c 		list_add(&va_block->node, &res->node);
node              548 drivers/misc/habanalabs/memory.c 	list_for_each_entry(va_block, &va_range->list, node) {
node              599 drivers/misc/habanalabs/memory.c 		list_del(&new_va_block->node);
node              930 drivers/misc/habanalabs/memory.c 	hash_add(ctx->mem_hash, &hnode->node, ret_vaddr);
node              989 drivers/misc/habanalabs/memory.c 	hash_for_each_possible(ctx->mem_hash, hnode, node, (unsigned long)vaddr)
node             1001 drivers/misc/habanalabs/memory.c 	hash_del(&hnode->node);
node             1087 drivers/misc/habanalabs/memory.c 	hash_add(ctx->mem_hash, &hnode->node, vaddr);
node             1576 drivers/misc/habanalabs/memory.c 	va_block = list_first_entry(&va_range->list, typeof(*va_block), node);
node             1629 drivers/misc/habanalabs/memory.c 	hash_for_each_safe(ctx->mem_hash, i, tmp_node, hnode, node) {
node               20 drivers/misc/habanalabs/mmu.c 	hash_for_each_possible(ctx->mmu_shadow_hash, pgt_info, node,
node               35 drivers/misc/habanalabs/mmu.c 	hash_del(&pgt_info->node);
node               67 drivers/misc/habanalabs/mmu.c 	hash_add(ctx->mmu_shadow_hash, &pgt_info->node, shadow_addr);
node              506 drivers/misc/habanalabs/mmu.c 	hash_for_each_safe(ctx->mmu_shadow_hash, i, tmp, pgt_info, node) {
node               30 drivers/misc/ibmasm/event.c 	list_for_each_entry(reader, &sp->event_buffer->readers, node)
node              123 drivers/misc/ibmasm/event.c 	list_add(&reader->node, &sp->event_buffer->readers);
node              132 drivers/misc/ibmasm/event.c 	list_del(&reader->node);
node              124 drivers/misc/ibmasm/ibmasm.h 	struct list_head	node;
node              140 drivers/misc/ibmasm/ibmasm.h 	struct list_head	node;
node              208 drivers/misc/ibmasm/ibmasmfs.c 	list_add(&sp->node, &service_processors);
node              586 drivers/misc/ibmasm/ibmasmfs.c 		sp = list_entry(entry, struct service_processor, node);
node               16 drivers/misc/lkdtm/bugs.c 	struct list_head node;
node              195 drivers/misc/lkdtm/bugs.c 	list_add(&good.node, &test_head);
node              204 drivers/misc/lkdtm/bugs.c 	list_add(&bad.node, &test_head);
node              219 drivers/misc/lkdtm/bugs.c 	list_add(&item.node, &test_head);
node              222 drivers/misc/lkdtm/bugs.c 	list_del(&item.node);
node              225 drivers/misc/lkdtm/bugs.c 	list_add(&item.node, &test_head);
node              228 drivers/misc/lkdtm/bugs.c 	item.node.next = redirection;
node              229 drivers/misc/lkdtm/bugs.c 	list_del(&item.node);
node              306 drivers/misc/mic/cosm/cosm_scif_server.c 		cdev = cosm_find_cdev_by_id(port_id.node - 1);
node              112 drivers/misc/mic/cosm_client/cosm_scif_client.c 	port_id.node = 0;
node              275 drivers/misc/mic/scif/scif_api.c 			msg.dst.node = conreq->msg.src.node;
node              284 drivers/misc/mic/scif/scif_api.c 			scif_nodeqp_send(&scif_dev[conreq->msg.src.node],
node              372 drivers/misc/mic/scif/scif_api.c 	ep->port.node = scif_info.nodeid;
node              600 drivers/misc/mic/scif/scif_api.c 	if (!scif_dev || dst->node > scif_info.maxid)
node              605 drivers/misc/mic/scif/scif_api.c 	remote_dev = &scif_dev[dst->node];
node              646 drivers/misc/mic/scif/scif_api.c 		ep->port.node = scif_info.nodeid;
node              683 drivers/misc/mic/scif/scif_api.c 	ep->remote_dev = &scif_dev[dst->node];
node              794 drivers/misc/mic/scif/scif_api.c 	peer->node = conreq->msg.src.node;
node              806 drivers/misc/mic/scif/scif_api.c 	cep->remote_dev = &scif_dev[peer->node];
node              845 drivers/misc/mic/scif/scif_api.c 	cep->port.node = lep->port.node;
node              847 drivers/misc/mic/scif/scif_api.c 	cep->peer.node = peer->node;
node              905 drivers/misc/mic/scif/scif_api.c 	msg.dst.node = conreq->msg.src.node;
node              909 drivers/misc/mic/scif/scif_api.c 	scif_nodeqp_send(&scif_dev[conreq->msg.src.node], &msg);
node             1422 drivers/misc/mic/scif/scif_api.c 	int node;
node             1430 drivers/misc/mic/scif/scif_api.c 	for (node = 0; node <= scif_info.maxid; node++) {
node             1431 drivers/misc/mic/scif/scif_api.c 		if (_scifdev_alive(&scif_dev[node])) {
node             1434 drivers/misc/mic/scif/scif_api.c 				nodes[offset++] = node;
node               20 drivers/misc/mic/scif/scif_debugfs.c 	int node;
node               31 drivers/misc/mic/scif/scif_debugfs.c 	for (node = 0; node <= scif_info.maxid; node++)
node               32 drivers/misc/mic/scif/scif_debugfs.c 		seq_printf(s, "%-16d\t%-16s\n", scif_dev[node].node,
node               33 drivers/misc/mic/scif/scif_debugfs.c 			   _scifdev_alive(&scif_dev[node]) ?
node              153 drivers/misc/mic/scif/scif_epd.c 	scif_nodeqp_send(&scif_dev[msg->src.node], msg);
node              171 drivers/misc/mic/scif/scif_epd.c 		ep->peer.node = msg->src.node;
node              299 drivers/misc/mic/scif/scif_epd.c 	scif_nodeqp_send(&scif_dev[msg->src.node], msg);
node              113 drivers/misc/mic/scif/scif_fd.c 		req.self.node = ep->port.node;
node               47 drivers/misc/mic/scif/scif_main.c 			 "SCIF INTR %d", scifdev->node);
node              118 drivers/misc/mic/scif/scif_main.c 		scifdev->node = i;
node              224 drivers/misc/mic/scif/scif_main.c 		scif_disconnect_node(scifdev->node, true);
node              169 drivers/misc/mic/scif/scif_main.h 	u8 node;
node              221 drivers/misc/mic/scif/scif_main.h 	return dev->node == scif_info.nodeid;
node              268 drivers/misc/mic/scif/scif_main.h void scif_handle_remove_node(int node);
node               90 drivers/misc/mic/scif/scif_mmap.c static void _scif_zap_mmaps(int node, struct list_head *head)
node               98 drivers/misc/mic/scif/scif_mmap.c 		if (ep->remote_dev->node == node)
node              108 drivers/misc/mic/scif/scif_mmap.c void scif_zap_mmaps(int node)
node              110 drivers/misc/mic/scif/scif_mmap.c 	_scif_zap_mmaps(node, &scif_info.connected);
node              111 drivers/misc/mic/scif/scif_mmap.c 	_scif_zap_mmaps(node, &scif_info.disconnected);
node              144 drivers/misc/mic/scif/scif_mmap.c void scif_cleanup_rma_for_zombies(int node)
node              152 drivers/misc/mic/scif/scif_mmap.c 		if (ep->remote_dev && ep->remote_dev->node == node)
node               18 drivers/misc/mic/scif/scif_nm.c static void scif_invalidate_ep(int node)
node               27 drivers/misc/mic/scif/scif_nm.c 		if (ep->remote_dev->node == node) {
node               36 drivers/misc/mic/scif/scif_nm.c 		if (ep->remote_dev->node == node) {
node               87 drivers/misc/mic/scif/scif_nm.c 		msg.src.node = scif_info.nodeid;
node               88 drivers/misc/mic/scif/scif_nm.c 		msg.dst.node = SCIF_MGMT_NODE;
node               89 drivers/misc/mic/scif/scif_nm.c 		msg.payload[0] = dev->node;
node               95 drivers/misc/mic/scif/scif_nm.c 		msg.src.node = scif_info.nodeid;
node               96 drivers/misc/mic/scif/scif_nm.c 		msg.dst.node = dev->node;
node              123 drivers/misc/mic/scif/scif_nm.c 	scif_invalidate_ep(dev->node);
node              124 drivers/misc/mic/scif/scif_nm.c 	scif_zap_mmaps(dev->node);
node              125 drivers/misc/mic/scif/scif_nm.c 	scif_cleanup_rma_for_zombies(dev->node);
node              128 drivers/misc/mic/scif/scif_nm.c 	if (!dev->node && scif_info.card_initiated_exit) {
node              144 drivers/misc/mic/scif/scif_nm.c void scif_handle_remove_node(int node)
node              146 drivers/misc/mic/scif/scif_nm.c 	struct scif_dev *scifdev = &scif_dev[node];
node              152 drivers/misc/mic/scif/scif_nm.c static int scif_send_rmnode_msg(int node, int remove_node)
node              155 drivers/misc/mic/scif/scif_nm.c 	struct scif_dev *dev = &scif_dev[node];
node              158 drivers/misc/mic/scif/scif_nm.c 	notif_msg.src.node = scif_info.nodeid;
node              159 drivers/misc/mic/scif/scif_nm.c 	notif_msg.dst.node = node;
node              219 drivers/misc/mic/scif/scif_nm.c 	msg.src.node = scif_info.nodeid;
node              220 drivers/misc/mic/scif/scif_nm.c 	msg.dst.node = SCIF_MGMT_NODE;
node              239 drivers/misc/mic/scif/scif_nodeqp.c 			scif_dev[scif_info.nodeid].node, scifdev->node);
node              298 drivers/misc/mic/scif/scif_nodeqp.c 		msg.dst.node = scifdev->node;
node              311 drivers/misc/mic/scif/scif_nodeqp.c 	msg.src.node = scif_info.nodeid;
node              312 drivers/misc/mic/scif/scif_nodeqp.c 	msg.dst.node = scifdev->node;
node              416 drivers/misc/mic/scif/scif_nodeqp.c 	p2p->ppi_peer_id = peerdev->node;
node              478 drivers/misc/mic/scif/scif_nodeqp.c 			if (p2p->ppi_peer_id == dev_j->node)
node              498 drivers/misc/mic/scif/scif_nodeqp.c 	msg.src.node = dev_j->node;
node              499 drivers/misc/mic/scif/scif_nodeqp.c 	msg.dst.node = dev_i->node;
node              515 drivers/misc/mic/scif/scif_nodeqp.c 	msg.src.node = dev_i->node;
node              516 drivers/misc/mic/scif/scif_nodeqp.c 	msg.dst.node = dev_j->node;
node              604 drivers/misc/mic/scif/scif_nodeqp.c 		 label, message_types[msg->uop], msg->src.node, msg->src.port,
node              605 drivers/misc/mic/scif/scif_nodeqp.c 		 msg->dst.node, msg->dst.port, msg->payload[0], msg->payload[1],
node              733 drivers/misc/mic/scif/scif_nodeqp.c 		scif_disconnect_node(scifdev->node, false);
node              777 drivers/misc/mic/scif/scif_nodeqp.c 		scifdev->node, msg->dst.node, msg->src.node);
node              781 drivers/misc/mic/scif/scif_nodeqp.c 	newdev = &scif_dev[msg->src.node];
node              782 drivers/misc/mic/scif/scif_nodeqp.c 	newdev->node = msg->src.node;
node              788 drivers/misc/mic/scif/scif_nodeqp.c 			"failed to setup interrupts for %d\n", msg->src.node);
node              794 drivers/misc/mic/scif/scif_nodeqp.c 			"failed to map mmio for %d\n", msg->src.node);
node              824 drivers/misc/mic/scif/scif_nodeqp.c 	msg->dst.node = msg->src.node;
node              825 drivers/misc/mic/scif/scif_nodeqp.c 	msg->src.node = scif_info.nodeid;
node              839 drivers/misc/mic/scif/scif_nodeqp.c 		"node add failed for node %d\n", msg->src.node);
node              841 drivers/misc/mic/scif/scif_nodeqp.c 	msg->dst.node = msg->src.node;
node              842 drivers/misc/mic/scif/scif_nodeqp.c 	msg->src.node = scif_info.nodeid;
node              870 drivers/misc/mic/scif/scif_nodeqp.c 		__func__, __LINE__, peerdev->node, qp->qp_state);
node              890 drivers/misc/mic/scif/scif_nodeqp.c 	struct scif_dev *dst_dev = &scif_dev[msg->dst.node];
node              894 drivers/misc/mic/scif/scif_nodeqp.c 		scifdev->node, msg->src.node, msg->dst.node);
node              911 drivers/misc/mic/scif/scif_nodeqp.c 	peerdev = &scif_dev[msg->src.node];
node              913 drivers/misc/mic/scif/scif_nodeqp.c 	peerdev->node = msg->src.node;
node              941 drivers/misc/mic/scif/scif_nodeqp.c 		struct scif_dev *dst_dev = &scif_dev[msg->dst.node];
node              944 drivers/misc/mic/scif/scif_nodeqp.c 			"SCIF_NODE_ADD_NACK received from %d\n", scifdev->node);
node              958 drivers/misc/mic/scif/scif_nodeqp.c 	int node = msg->payload[0];
node              959 drivers/misc/mic/scif/scif_nodeqp.c 	struct scif_dev *scdev = &scif_dev[node];
node              962 drivers/misc/mic/scif/scif_nodeqp.c 	scif_handle_remove_node(node);
node              990 drivers/misc/mic/scif/scif_nodeqp.c 		swap(msg->dst.node, msg->src.node);
node             1014 drivers/misc/mic/scif/scif_nodeqp.c 		msg->uop, scifdev->node);
node             1082 drivers/misc/mic/scif/scif_nodeqp.c 			msg->uop, scifdev->node);
node             1227 drivers/misc/mic/scif/scif_nodeqp.c 		 "SCIF LOOPB %d", scifdev->node);
node             1266 drivers/misc/mic/scif/scif_nodeqp.c 	scif_info.nodeid = scifdev->node;
node             1328 drivers/misc/mic/scif/scif_nodeqp.c 			if (p2p->ppi_peer_id == scifdev->node) {
node               45 drivers/misc/mic/scif/scif_peer_bus.c 	spdev->dnode = scifdev->node;
node               60 drivers/misc/mic/scif/scif_peer_bus.c 		"dnode %d: initialize_device rc %d\n", scifdev->node, ret);
node               74 drivers/misc/mic/scif/scif_peer_bus.c 			"dnode %d: peer device_add failed\n", scifdev->node);
node               84 drivers/misc/mic/scif/scif_peer_bus.c 			"dnode %d: dmam_pool_create failed\n", scifdev->node);
node              342 drivers/misc/mic/scif/scif_rma.h void scif_zap_mmaps(int node);
node              344 drivers/misc/mic/scif/scif_rma.h bool scif_rma_do_apps_have_mmaps(int node);
node              346 drivers/misc/mic/scif/scif_rma.h void scif_cleanup_rma_for_zombies(int node);
node               87 drivers/misc/qcom-coincell.c 	struct device_node *node = pdev->dev.of_node;
node              102 drivers/misc/qcom-coincell.c 	rc = of_property_read_u32(node, "reg", &chgr.base_addr);
node              106 drivers/misc/qcom-coincell.c 	enable = !of_property_read_bool(node, "qcom,charger-disable");
node              109 drivers/misc/qcom-coincell.c 		rc = of_property_read_u32(node, "qcom,rset-ohms", &rset);
node              116 drivers/misc/qcom-coincell.c 		rc = of_property_read_u32(node, "qcom,vset-millivolts", &vset);
node               45 drivers/misc/sgi-gru/gru.h 	int	node;
node              223 drivers/misc/vmw_vmci/vmci_context.c 		struct vmci_handle_list *node;
node              233 drivers/misc/vmw_vmci/vmci_context.c 		list_for_each_entry_rcu(node, &sub_ctx->notifier_list, node) {
node              234 drivers/misc/vmw_vmci/vmci_context.c 			if (!vmci_handle_is_equal(node->handle, context_handle))
node              472 drivers/misc/vmw_vmci/vmci_context.c 				 &context->notifier_list, node) {
node              473 drivers/misc/vmw_vmci/vmci_context.c 		list_del(&notifier->node);
node              627 drivers/misc/vmw_vmci/vmci_context.c 	INIT_LIST_HEAD(&notifier->node);
node              633 drivers/misc/vmw_vmci/vmci_context.c 		list_for_each_entry(n, &context->notifier_list, node) {
node              644 drivers/misc/vmw_vmci/vmci_context.c 			list_add_tail_rcu(&notifier->node,
node              680 drivers/misc/vmw_vmci/vmci_context.c 				 &context->notifier_list, node) {
node              682 drivers/misc/vmw_vmci/vmci_context.c 			list_del_rcu(&notifier->node);
node              724 drivers/misc/vmw_vmci/vmci_context.c 	list_for_each_entry(entry, &context->notifier_list, node)
node               36 drivers/misc/vmw_vmci/vmci_context.h 	struct list_head node;
node               35 drivers/misc/vmw_vmci/vmci_doorbell.c 	struct hlist_node node;
node              124 drivers/misc/vmw_vmci/vmci_doorbell.c 			     node) {
node              188 drivers/misc/vmw_vmci/vmci_doorbell.c 	hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]);
node              201 drivers/misc/vmw_vmci/vmci_doorbell.c 	hlist_del_init(&entry->node);
node              360 drivers/misc/vmw_vmci/vmci_doorbell.c 	hlist_for_each_entry(dbell, &vmci_doorbell_it.entries[bucket], node) {
node              466 drivers/misc/vmw_vmci/vmci_doorbell.c 	INIT_HLIST_NODE(&entry->node);
node              531 drivers/misc/vmw_vmci/vmci_doorbell.c 	if (!hlist_unhashed(&entry->node)) {
node               27 drivers/misc/vmw_vmci/vmci_event.c 	struct list_head node;	/* on one of subscriber lists */
node               50 drivers/misc/vmw_vmci/vmci_event.c 		list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) {
node               58 drivers/misc/vmw_vmci/vmci_event.c 			list_del(&cur->node);
node               73 drivers/misc/vmw_vmci/vmci_event.c 		list_for_each_entry(cur, &subscriber_array[e], node) {
node               92 drivers/misc/vmw_vmci/vmci_event.c 	list_for_each_entry_rcu(cur, subscriber_list, node) {
node              158 drivers/misc/vmw_vmci/vmci_event.c 	INIT_LIST_HEAD(&sub->node);
node              179 drivers/misc/vmw_vmci/vmci_event.c 		list_add_rcu(&sub->node, &subscriber_array[event]);
node              206 drivers/misc/vmw_vmci/vmci_event.c 		list_del_rcu(&s->node);
node               46 drivers/misc/vmw_vmci/vmci_resource.c 				 &vmci_resource_table.entries[idx], node) {
node              123 drivers/misc/vmw_vmci/vmci_resource.c 	INIT_HLIST_NODE(&resource->node);
node              128 drivers/misc/vmw_vmci/vmci_resource.c 	hlist_add_head_rcu(&resource->node, &vmci_resource_table.entries[idx]);
node              146 drivers/misc/vmw_vmci/vmci_resource.c 	hlist_for_each_entry(r, &vmci_resource_table.entries[idx], node) {
node              148 drivers/misc/vmw_vmci/vmci_resource.c 			hlist_del_init_rcu(&r->node);
node              196 drivers/misc/vmw_vmci/vmci_resource.c 	WARN_ON(!hlist_unhashed(&resource->node));
node               30 drivers/misc/vmw_vmci/vmci_resource.h 	struct hlist_node node;
node             1066 drivers/misc/xilinx_sdfec.c 	struct device_node *node = dev->of_node;
node             1074 drivers/misc/xilinx_sdfec.c 	rval = of_property_read_string(node, "xlnx,sdfec-code", &fec_code);
node             1085 drivers/misc/xilinx_sdfec.c 	rval = of_property_read_u32(node, "xlnx,sdfec-din-words",
node             1095 drivers/misc/xilinx_sdfec.c 	rval = of_property_read_u32(node, "xlnx,sdfec-din-width", &din_width);
node             1110 drivers/misc/xilinx_sdfec.c 	rval = of_property_read_u32(node, "xlnx,sdfec-dout-words",
node             1120 drivers/misc/xilinx_sdfec.c 	rval = of_property_read_u32(node, "xlnx,sdfec-dout-width", &dout_width);
node              161 drivers/mmc/core/block.c 	struct list_head node;
node             2547 drivers/mmc/core/block.c 	list_add(&rpmb->node, &md->rpmbs);
node             2647 drivers/mmc/core/block.c 		rpmb = list_entry(pos, struct mmc_rpmb_data, node);
node             1087 drivers/mmc/core/core.c static int mmc_of_get_func_num(struct device_node *node)
node             1092 drivers/mmc/core/core.c 	ret = of_property_read_u32(node, "reg", &reg);
node             1102 drivers/mmc/core/core.c 	struct device_node *node;
node             1107 drivers/mmc/core/core.c 	for_each_child_of_node(host->parent->of_node, node) {
node             1108 drivers/mmc/core/core.c 		if (mmc_of_get_func_num(node) == func_num)
node             1109 drivers/mmc/core/core.c 			return node;
node              149 drivers/mmc/host/cavium-octeon.c 	struct device_node *cn, *node = pdev->dev.of_node;
node              177 drivers/mmc/host/cavium-octeon.c 	if (of_device_is_compatible(node, "cavium,octeon-7890-mmc")) {
node              280 drivers/mmc/host/cavium-octeon.c 	for_each_child_of_node(node, cn) {
node               59 drivers/mmc/host/cavium-thunderx.c 	struct device_node *node = pdev->dev.of_node;
node              127 drivers/mmc/host/cavium-thunderx.c 	for_each_child_of_node(node, child_node) {
node              954 drivers/mmc/host/cavium.c 	struct device_node *node = dev->of_node;
node              959 drivers/mmc/host/cavium.c 	ret = of_property_read_u32(node, "reg", &id);
node              961 drivers/mmc/host/cavium.c 		dev_err(dev, "Missing or invalid reg property on %pOF\n", node);
node              966 drivers/mmc/host/cavium.c 		dev_err(dev, "Invalid reg property on %pOF\n", node);
node              987 drivers/mmc/host/cavium.c 		of_property_read_u32(node, "cavium,bus-max-width", &bus_width);
node              996 drivers/mmc/host/cavium.c 		of_property_read_u32(node, "spi-max-frequency", &mmc->f_max);
node             1003 drivers/mmc/host/cavium.c 	of_property_read_u32(node, "cavium,cmd-clk-skew", &cmd_skew);
node             1004 drivers/mmc/host/cavium.c 	of_property_read_u32(node, "cavium,dat-clk-skew", &dat_skew);
node              155 drivers/mmc/host/dw_mmc-zx.c 	struct device_node *node;
node              161 drivers/mmc/host/dw_mmc-zx.c 	node = of_parse_phandle(np, "zte,aon-syscon", 0);
node              162 drivers/mmc/host/dw_mmc-zx.c 	if (node) {
node              163 drivers/mmc/host/dw_mmc-zx.c 		sysc_base = syscon_node_to_regmap(node);
node              164 drivers/mmc/host/dw_mmc-zx.c 		of_node_put(node);
node              558 drivers/mmc/host/moxart-mmc.c 	struct device_node *node = dev->of_node;
node              575 drivers/mmc/host/moxart-mmc.c 	ret = of_address_to_resource(node, 0, &res_mmc);
node              581 drivers/mmc/host/moxart-mmc.c 	irq = irq_of_parse_and_map(node, 0);
node              701 drivers/mmc/host/sdhci-acpi.c 		list_for_each_entry(child, &device->children, node)
node              736 drivers/mmc/host/sdhci-of-arasan.c 	struct device_node *node;
node              757 drivers/mmc/host/sdhci-of-arasan.c 	node = of_parse_phandle(pdev->dev.of_node, "arasan,soc-ctl-syscon", 0);
node              758 drivers/mmc/host/sdhci-of-arasan.c 	if (node) {
node              759 drivers/mmc/host/sdhci-of-arasan.c 		sdhci_arasan->soc_ctl_base = syscon_node_to_regmap(node);
node              760 drivers/mmc/host/sdhci-of-arasan.c 		of_node_put(node);
node             1129 drivers/mmc/host/sdhci-pci-core.c 	list_for_each_entry(child, &device->children, node)
node              427 drivers/mmc/host/sdhci-s3c.c 	struct device_node *node = dev->of_node;
node              431 drivers/mmc/host/sdhci-s3c.c 	if (of_property_read_u32(node, "bus-width", &max_width))
node              436 drivers/mmc/host/sdhci-s3c.c 	if (of_get_property(node, "broken-cd", NULL)) {
node              441 drivers/mmc/host/sdhci-s3c.c 	if (of_get_property(node, "non-removable", NULL)) {
node              446 drivers/mmc/host/sdhci-s3c.c 	if (of_get_named_gpio(node, "cd-gpios", 0))
node              161 drivers/mtd/nand/raw/atmel/nand-controller.c 	struct list_head node;
node             1553 drivers/mtd/nand/raw/atmel/nand-controller.c 	list_del(&nand->node);
node             1692 drivers/mtd/nand/raw/atmel/nand-controller.c 	list_add_tail(&nand->node, &nc->chips);
node             1703 drivers/mtd/nand/raw/atmel/nand-controller.c 	list_for_each_entry_safe(nand, tmp, &nc->chips, node) {
node             2569 drivers/mtd/nand/raw/atmel/nand-controller.c 	list_for_each_entry(nand, &nc->chips, node) {
node              230 drivers/mtd/nand/raw/brcmnand/brcmnand.c 	struct list_head	node;
node             2470 drivers/mtd/nand/raw/brcmnand/brcmnand.c 	list_for_each_entry(host, &ctrl->host_list, node)
node             2504 drivers/mtd/nand/raw/brcmnand/brcmnand.c 	list_for_each_entry(host, &ctrl->host_list, node) {
node             2718 drivers/mtd/nand/raw/brcmnand/brcmnand.c 			list_add_tail(&host->node, &ctrl->host_list);
node             2742 drivers/mtd/nand/raw/brcmnand/brcmnand.c 	list_for_each_entry(host, &ctrl->host_list, node)
node             1185 drivers/mtd/nand/raw/denali.c 		list_for_each_entry(dchip2, &denali->chips, node) {
node             1238 drivers/mtd/nand/raw/denali.c 	list_add_tail(&dchip->node, &denali->chips);
node             1340 drivers/mtd/nand/raw/denali.c 	list_for_each_entry(dchip, &denali->chips, node)
node              329 drivers/mtd/nand/raw/denali.h 	struct list_head node;
node              858 drivers/mtd/nand/raw/fsl_elbc_nand.c 	struct device_node *node = pdev->dev.of_node;
node              867 drivers/mtd/nand/raw/fsl_elbc_nand.c 	ret = of_address_to_resource(node, 0, &res);
node              984 drivers/mtd/nand/raw/fsl_ifc_nand.c 	struct device_node *node = dev->dev.of_node;
node              992 drivers/mtd/nand/raw/fsl_ifc_nand.c 	ret = of_address_to_resource(node, 0, &res);
node              329 drivers/mtd/nand/raw/marvell_nand.c 	struct list_head node;
node             2671 drivers/mtd/nand/raw/marvell_nand.c 	list_add_tail(&marvell_nand->node, &nfc->chips);
node             2721 drivers/mtd/nand/raw/marvell_nand.c 	list_for_each_entry_safe(entry, temp, &nfc->chips, node) {
node             2723 drivers/mtd/nand/raw/marvell_nand.c 		list_del(&entry->node);
node             2959 drivers/mtd/nand/raw/marvell_nand.c 	list_for_each_entry(chip, &nfc->chips, node)
node              108 drivers/mtd/nand/raw/meson_nand.c 	struct list_head node;
node             1285 drivers/mtd/nand/raw/meson_nand.c 	list_add_tail(&meson_chip->node, &nfc->chips);
node             1298 drivers/mtd/nand/raw/meson_nand.c 					      struct meson_nfc_nand_chip, node);
node             1306 drivers/mtd/nand/raw/meson_nand.c 		list_del(&meson_chip->node);
node              126 drivers/mtd/nand/raw/mtk_nand.c 	struct list_head node;
node             1426 drivers/mtd/nand/raw/mtk_nand.c 	list_add_tail(&chip->node, &nfc->chips);
node             1586 drivers/mtd/nand/raw/mtk_nand.c 					node);
node             1588 drivers/mtd/nand/raw/mtk_nand.c 		list_del(&chip->node);
node             1622 drivers/mtd/nand/raw/mtk_nand.c 	list_for_each_entry(chip, &nfc->chips, node) {
node              133 drivers/mtd/nand/raw/ndfc.c 			  struct device_node *node)
node              160 drivers/mtd/nand/raw/ndfc.c 	flash_np = of_get_next_child(node, NULL);
node              276 drivers/mtd/nand/raw/qcom_nandc.c 	struct list_head node;
node              436 drivers/mtd/nand/raw/qcom_nandc.c 	struct list_head node;
node              792 drivers/mtd/nand/raw/qcom_nandc.c 	list_add_tail(&desc->node, &nandc->desc_list);
node              953 drivers/mtd/nand/raw/qcom_nandc.c 	list_add_tail(&desc->node, &nandc->desc_list);
node             1312 drivers/mtd/nand/raw/qcom_nandc.c 	list_for_each_entry(desc, &nandc->desc_list, node)
node             1343 drivers/mtd/nand/raw/qcom_nandc.c 	list_for_each_entry_safe(desc, n, &nandc->desc_list, node) {
node             1344 drivers/mtd/nand/raw/qcom_nandc.c 		list_del(&desc->node);
node             2863 drivers/mtd/nand/raw/qcom_nandc.c 		list_add_tail(&host->node, &nandc->host_list);
node             2986 drivers/mtd/nand/raw/qcom_nandc.c 	list_for_each_entry(host, &nandc->host_list, node)
node              192 drivers/mtd/nand/raw/sunxi_nand.c 	struct list_head node;
node             2010 drivers/mtd/nand/raw/sunxi_nand.c 	list_add_tail(&sunxi_nand->node, &nfc->chips);
node             2045 drivers/mtd/nand/raw/sunxi_nand.c 					      node);
node             2048 drivers/mtd/nand/raw/sunxi_nand.c 		list_del(&sunxi_nand->node);
node              373 drivers/mtd/ubi/fastmap.c 	struct rb_node *node, *node2;
node              376 drivers/mtd/ubi/fastmap.c 	ubi_rb_for_each_entry(node, av, &ai->volumes, rb) {
node              305 drivers/net/arcnet/arc-rimi.c static int node;
node              310 drivers/net/arcnet/arc-rimi.c module_param(node, int, 0);
node              326 drivers/net/arcnet/arc-rimi.c 	if (node && node != 0xff)
node              327 drivers/net/arcnet/arc-rimi.c 		dev->dev_addr[0] = node;
node              368 drivers/net/arcnet/arc-rimi.c 		node = ints[3];
node              122 drivers/net/arcnet/com20020-isa.c static int node = 0;
node              131 drivers/net/arcnet/com20020-isa.c module_param(node, int, 0);
node              153 drivers/net/arcnet/com20020-isa.c 	if (node && node != 0xff)
node              154 drivers/net/arcnet/com20020-isa.c 		dev->dev_addr[0] = node;
node              211 drivers/net/arcnet/com20020-isa.c 		node = ints[3];
node               51 drivers/net/arcnet/com20020-pci.c static int node;
node               58 drivers/net/arcnet/com20020-pci.c module_param(node, int, 0);
node              193 drivers/net/arcnet/com20020-pci.c 		dev->dev_addr[0] = node;
node               88 drivers/net/arcnet/com20020_cs.c static int node;
node               94 drivers/net/arcnet/com20020_cs.c module_param(node, int, 0);
node              136 drivers/net/arcnet/com20020_cs.c 	dev->dev_addr[0] = node;
node               67 drivers/net/caif/caif_serial.c 	struct list_head node;
node              310 drivers/net/caif/caif_serial.c 		list_for_each_entry_safe(ser, tmp, &list, node) {
node              361 drivers/net/caif/caif_serial.c 	list_add(&ser->node, &ser_list);
node              376 drivers/net/caif/caif_serial.c 	list_move(&ser->node, &ser_release_list);
node              212 drivers/net/ethernet/amazon/ena/ena_netdev.c 	int size, i, node;
node              221 drivers/net/ethernet/amazon/ena/ena_netdev.c 	node = cpu_to_node(ena_irq->cpu);
node              223 drivers/net/ethernet/amazon/ena/ena_netdev.c 	tx_ring->tx_buffer_info = vzalloc_node(size, node);
node              231 drivers/net/ethernet/amazon/ena/ena_netdev.c 	tx_ring->free_ids = vzalloc_node(size, node);
node              239 drivers/net/ethernet/amazon/ena/ena_netdev.c 	tx_ring->push_buf_intermediate_buf = vzalloc_node(size, node);
node              358 drivers/net/ethernet/amazon/ena/ena_netdev.c 	int size, node, i;
node              370 drivers/net/ethernet/amazon/ena/ena_netdev.c 	node = cpu_to_node(ena_irq->cpu);
node              372 drivers/net/ethernet/amazon/ena/ena_netdev.c 	rx_ring->rx_buffer_info = vzalloc_node(size, node);
node              380 drivers/net/ethernet/amazon/ena/ena_netdev.c 	rx_ring->free_ids = vzalloc_node(size, node);
node              188 drivers/net/ethernet/amd/xgbe/xgbe-desc.c static void *xgbe_alloc_node(size_t size, int node)
node              192 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	mem = kzalloc_node(size, GFP_KERNEL, node);
node              200 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 				 dma_addr_t *dma, int node)
node              205 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	set_dev_node(dev, node);
node              228 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 					  ring->node);
node              235 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	ring->rdata = xgbe_alloc_node(size, ring->node);
node              241 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		  ring->rdesc, &ring->rdesc_dma, ring->rdata, ring->node);
node              287 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 			    int node)
node              300 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		pages = alloc_pages_node(node, gfp, order);
node              308 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if (!pages && (node != NUMA_NO_NODE)) {
node              309 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		node = NUMA_NO_NODE;
node              363 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		ret = xgbe_alloc_pages(pdata, &ring->rx_hdr_pa, 0, ring->node);
node              370 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 				       PAGE_ALLOC_COSTLY_ORDER, ring->node);
node              161 drivers/net/ethernet/amd/xgbe/xgbe-drv.c static void *xgbe_alloc_node(size_t size, int node)
node              165 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	mem = kzalloc_node(size, GFP_KERNEL, node);
node              196 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	int node;
node              204 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		node = cpu_to_node(cpu);
node              206 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		channel = xgbe_alloc_node(sizeof(*channel), node);
node              216 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		channel->node = node;
node              223 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			ring = xgbe_alloc_node(sizeof(*ring), node);
node              228 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			ring->node = node;
node              234 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			ring = xgbe_alloc_node(sizeof(*ring), node);
node              239 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			ring->node = node;
node              245 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			  "%s: cpu=%u, node=%d\n", channel->name, cpu, node);
node              468 drivers/net/ethernet/amd/xgbe/xgbe.h 	int node;
node              522 drivers/net/ethernet/amd/xgbe/xgbe.h 	int node;
node              681 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 		rc =  rhashtable_remove_fast(&tc_info->l2_table, &l2_node->node,
node              709 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 		rc = rhashtable_insert_fast(l2_table, &l2_node->node,
node              827 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 		rc =  rhashtable_remove_fast(tunnel_table, &tunnel_node->node,
node              861 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 		rc = rhashtable_insert_fast(tunnel_table, &tunnel_node->node,
node              926 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 					     &decap_l2_node->node,
node             1211 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	rc = rhashtable_remove_fast(&tc_info->flow_table, &flow_node->node,
node             1309 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	rc = rhashtable_insert_fast(&tc_info->flow_table, &new_node->node,
node             1584 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	.head_offset = offsetof(struct bnxt_tc_flow_node, node),
node             1591 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	.head_offset = offsetof(struct bnxt_tc_l2_node, node),
node             1598 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	.head_offset = offsetof(struct bnxt_tc_l2_node, node),
node             1605 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c 	.head_offset = offsetof(struct bnxt_tc_tunnel_node, node),
node              134 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h 	struct rhash_head		node;
node              158 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h 	struct rhash_head	node;
node              172 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.h 	struct rhash_head		node;
node              270 drivers/net/ethernet/cavium/liquidio/octeon_iq.h 	struct list_head node;
node              613 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	struct list_head *node;
node              616 drivers/net/ethernet/cavium/liquidio/octeon_network.h 		node = NULL;
node              618 drivers/net/ethernet/cavium/liquidio/octeon_network.h 		node = root->next;
node              620 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	if (node)
node              621 drivers/net/ethernet/cavium/liquidio/octeon_network.h 		list_del(node);
node              623 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	return node;
node              413 drivers/net/ethernet/cavium/liquidio/request_manager.c 			list_add_tail(&sc->node, &oct->response_list
node              766 drivers/net/ethernet/cavium/liquidio/request_manager.c 		list_add_tail(&sc->node, &oct->sc_buf_pool.head);
node              790 drivers/net/ethernet/cavium/liquidio/request_manager.c 		sc = list_entry(tmp, struct octeon_soft_command, node);
node              793 drivers/net/ethernet/cavium/liquidio/request_manager.c 			list_del(&sc->node);
node              798 drivers/net/ethernet/cavium/liquidio/request_manager.c 				list_add_tail(&sc->node, &zombie_sc_list->head);
node              826 drivers/net/ethernet/cavium/liquidio/request_manager.c 		sc = list_entry(tmp, struct octeon_soft_command, node);
node              935 drivers/net/ethernet/cavium/liquidio/request_manager.c 	list_add_tail(&sc->node, &oct->sc_buf_pool.head);
node               85 drivers/net/ethernet/cavium/liquidio/response_manager.c 				      struct octeon_soft_command, node);
node              137 drivers/net/ethernet/cavium/liquidio/response_manager.c 			list_del(&sc->node);
node              146 drivers/net/ethernet/cavium/liquidio/response_manager.c 				list_add_tail(&sc->node,
node              186 drivers/net/ethernet/cavium/liquidio/response_manager.c 					list_add_tail(&sc->node,
node              312 drivers/net/ethernet/cavium/thunder/nic.h 	u8			node;
node               42 drivers/net/ethernet/cavium/thunder/nic_main.c 	u8			node;
node              173 drivers/net/ethernet/cavium/thunder/nic_main.c 		mac = bgx_get_lmac_mac(nic->node, bgx_idx, lmac);
node              178 drivers/net/ethernet/cavium/thunder/nic_main.c 	mbx.nic_cfg.node_id = nic->node;
node              244 drivers/net/ethernet/cavium/thunder/nic_main.c 		mbx.bgx_stats.stats = bgx_get_rx_stats(nic->node, bgx_idx,
node              247 drivers/net/ethernet/cavium/thunder/nic_main.c 		mbx.bgx_stats.stats = bgx_get_tx_stats(nic->node, bgx_idx,
node              268 drivers/net/ethernet/cavium/thunder/nic_main.c 	lmac_cnt = bgx_get_lmac_count(nic->node, bgx);
node              319 drivers/net/ethernet/cavium/thunder/nic_main.c 	unsigned bgx_map = bgx_get_map(nic->node);
node              329 drivers/net/ethernet/cavium/thunder/nic_main.c 		lmac_cnt = bgx_get_lmac_count(nic->node, bgx);
node              770 drivers/net/ethernet/cavium/thunder/nic_main.c 	bgx_lmac_internal_loopback(nic->node, bgx_idx, lmac_idx, lbk->enable);
node              858 drivers/net/ethernet/cavium/thunder/nic_main.c 	bgx_lmac_rx_tx_enable(nic->node, bgx, lmac, enable);
node              873 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_lmac_get_pfc(nic->node, bgx, lmac, &pfc);
node              880 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_lmac_set_pfc(nic->node, bgx, lmac, cfg);
node              909 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_config_timestamping(nic->node, bgx_idx, lmac, true);
node              915 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_config_timestamping(nic->node, bgx_idx, lmac, false);
node              940 drivers/net/ethernet/cavium/thunder/nic_main.c 	bgx_get_lmac_link_state(nic->node, bgx, lmac, &link);
node             1036 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_set_lmac_mac(nic->node, bgx, lmac, mbx.mac.mac_addr);
node             1097 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_reset_xcast_mode(nic->node, bgx, lmac,
node             1109 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_set_dmac_cam_filter(nic->node, bgx, lmac,
node             1122 drivers/net/ethernet/cavium/thunder/nic_main.c 		bgx_set_xcast_mode(nic->node, bgx, lmac, mbx.xcast.mode);
node             1345 drivers/net/ethernet/cavium/thunder/nic_main.c 	nic->node = nic_get_node_id(pdev);
node              223 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		nic->node = mbx.nic_cfg.node_id;
node             1114 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		cpumask_set_cpu(cpumask_local_spread(cpu, nic->node),
node              169 drivers/net/ethernet/cavium/thunder/thunder_bgx.c static struct bgx *get_bgx(int node, int bgx_idx)
node              171 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	int idx = (node * max_bgx_per_node) + bgx_idx;
node              177 drivers/net/ethernet/cavium/thunder/thunder_bgx.c unsigned bgx_get_map(int node)
node              183 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		if (bgx_vnic[(node * max_bgx_per_node) + i])
node              192 drivers/net/ethernet/cavium/thunder/thunder_bgx.c int bgx_get_lmac_count(int node, int bgx_idx)
node              196 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx = get_bgx(node, bgx_idx);
node              205 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_get_lmac_link_state(int node, int bgx_idx, int lmacid, void *status)
node              211 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx = get_bgx(node, bgx_idx);
node              223 drivers/net/ethernet/cavium/thunder/thunder_bgx.c const u8 *bgx_get_lmac_mac(int node, int bgx_idx, int lmacid)
node              225 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              234 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_lmac_mac(int node, int bgx_idx, int lmacid, const u8 *mac)
node              236 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              331 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_dmac_cam_filter(int node, int bgx_idx, int lmacid,
node              334 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              354 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_set_xcast_mode(int node, int bgx_idx, int lmacid, u8 mode)
node              356 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              388 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_reset_xcast_mode(int node, int bgx_idx, int lmacid, u8 vf_id)
node              390 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              397 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx_set_xcast_mode(node, bgx_idx, lmacid,
node              402 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_rx_tx_enable(int node, int bgx_idx, int lmacid, bool enable)
node              404 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              434 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_config_timestamping(int node, int bgx_idx, int lmacid, bool enable)
node              436 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              462 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_get_pfc(int node, int bgx_idx, int lmacid, void *pause)
node              465 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              482 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_set_pfc(int node, int bgx_idx, int lmacid, void *pause)
node              485 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct bgx *bgx = get_bgx(node, bgx_idx);
node              630 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx)
node              634 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx = get_bgx(node, bgx_idx);
node              644 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx)
node              648 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx = get_bgx(node, bgx_idx);
node              657 drivers/net/ethernet/cavium/thunder/thunder_bgx.c void bgx_lmac_internal_loopback(int node, int bgx_idx,
node              664 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	bgx = get_bgx(node, bgx_idx);
node             1471 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	struct device_node *node = NULL;
node             1482 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		node = to_of_node(fwn);
node             1483 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		if (!node)
node             1486 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		mac = of_get_mac_address(node);
node             1493 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		phy_np = of_parse_phandle(node, "phy-handle", 0);
node             1509 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 			of_node_put(node);
node             1526 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 	of_node_put(node);
node              219 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_dmac_cam_filter(int node, int bgx_idx, int lmacid, u64 mac, u8 vf);
node              220 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_reset_xcast_mode(int node, int bgx_idx, int lmacid, u8 vf);
node              221 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_xcast_mode(int node, int bgx_idx, int lmacid, u8 mode);
node              223 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_rx_tx_enable(int node, int bgx_idx, int lmacid, bool enable);
node              224 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_add_dmac_addr(u64 dmac, int node, int bgx_idx, int lmac);
node              225 drivers/net/ethernet/cavium/thunder/thunder_bgx.h unsigned bgx_get_map(int node);
node              226 drivers/net/ethernet/cavium/thunder/thunder_bgx.h int bgx_get_lmac_count(int node, int bgx);
node              227 drivers/net/ethernet/cavium/thunder/thunder_bgx.h const u8 *bgx_get_lmac_mac(int node, int bgx_idx, int lmacid);
node              228 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_set_lmac_mac(int node, int bgx_idx, int lmacid, const u8 *mac);
node              229 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_get_lmac_link_state(int node, int bgx_idx, int lmacid, void *status);
node              230 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_internal_loopback(int node, int bgx_idx,
node              232 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_config_timestamping(int node, int bgx_idx, int lmacid, bool enable);
node              233 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_get_pfc(int node, int bgx_idx, int lmacid, void *pause);
node              234 drivers/net/ethernet/cavium/thunder/thunder_bgx.h void bgx_lmac_set_pfc(int node, int bgx_idx, int lmacid, void *pause);
node              239 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx);
node              240 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx);
node             3564 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	int node, ret;
node             3611 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	node = dev_to_node(adapter->pdev_dev);
node             3613 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		newpage = alloc_pages_node(node, __GFP_NOWARN | GFP_KERNEL |
node              697 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c 	ret = rhashtable_insert_fast(&adap->flower_tbl, &ch_flower->node,
node              727 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c 	ret = rhashtable_remove_fast(&adap->flower_tbl, &ch_flower->node,
node              830 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c 	.head_offset = offsetof(struct ch_tc_flower_entry, node),
node               51 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.h 	struct rhash_head node;
node              600 drivers/net/ethernet/chelsio/cxgb4/sge.c 	int node;
node              608 drivers/net/ethernet/chelsio/cxgb4/sge.c 	node = dev_to_node(adap->pdev_dev);
node              617 drivers/net/ethernet/chelsio/cxgb4/sge.c 		pg = alloc_pages_node(node, gfp | __GFP_COMP, s->fl_pg_order);
node              648 drivers/net/ethernet/chelsio/cxgb4/sge.c 		pg = alloc_pages_node(node, gfp, 0);
node              715 drivers/net/ethernet/chelsio/cxgb4/sge.c 			size_t stat_size, int node)
node              724 drivers/net/ethernet/chelsio/cxgb4/sge.c 		s = kcalloc_node(sw_size, nelem, GFP_KERNEL, node);
node              117 drivers/net/ethernet/cisco/enic/enic.h 	struct hlist_node node;
node               97 drivers/net/ethernet/cisco/enic/enic_clsf.c 		hlist_for_each_entry_safe(n, tmp, hhead, node) {
node               99 drivers/net/ethernet/cisco/enic/enic_clsf.c 			hlist_del(&n->node);
node              117 drivers/net/ethernet/cisco/enic/enic_clsf.c 		hlist_for_each_entry_safe(n, tmp, hhead, node)
node              139 drivers/net/ethernet/cisco/enic/enic_clsf.c 		hlist_for_each_entry_safe(n, tmp, hhead, node) {
node              146 drivers/net/ethernet/cisco/enic/enic_clsf.c 				hlist_del(&n->node);
node              161 drivers/net/ethernet/cisco/enic/enic_clsf.c 	hlist_for_each_entry(tpos, h, node)
node              216 drivers/net/ethernet/cisco/enic/enic_clsf.c 				hlist_del(&n->node);
node              241 drivers/net/ethernet/cisco/enic/enic_clsf.c 					INIT_HLIST_NODE(&d->node);
node              242 drivers/net/ethernet/cisco/enic/enic_clsf.c 					hlist_add_head(&d->node, head);
node              277 drivers/net/ethernet/cisco/enic/enic_clsf.c 		INIT_HLIST_NODE(&n->node);
node              278 drivers/net/ethernet/cisco/enic/enic_clsf.c 		hlist_add_head(&n->node, &enic->rfs_h.ht_head[tbl_idx]);
node              427 drivers/net/ethernet/cisco/enic/enic_ethtool.c 		hlist_for_each_entry_safe(n, tmp, hhead, node) {
node              114 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c 	struct device_node *node;
node              140 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c 	node = of_find_compatible_node(NULL, NULL, "fsl,dpaa2-ptp");
node              141 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c 	if (!node) {
node              146 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c 	dev->of_node = node;
node              148 drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c 	base = of_iomap(node, 0);
node             2068 drivers/net/ethernet/freescale/fec_main.c 	struct device_node *node;
node             2149 drivers/net/ethernet/freescale/fec_main.c 	node = of_get_child_by_name(pdev->dev.of_node, "mdio");
node             2150 drivers/net/ethernet/freescale/fec_main.c 	err = of_mdiobus_register(fep->mii_bus, node);
node             2151 drivers/net/ethernet/freescale/fec_main.c 	of_node_put(node);
node             1109 drivers/net/ethernet/freescale/fman/fman_dtsec.c 	INIT_LIST_HEAD(&hash_entry->node);
node             1113 drivers/net/ethernet/freescale/fman/fman_dtsec.c 		list_add_tail(&hash_entry->node,
node             1116 drivers/net/ethernet/freescale/fman/fman_dtsec.c 		list_add_tail(&hash_entry->node,
node             1209 drivers/net/ethernet/freescale/fman/fman_dtsec.c 				list_del_init(&hash_entry->node);
node             1222 drivers/net/ethernet/freescale/fman/fman_dtsec.c 				list_del_init(&hash_entry->node);
node               76 drivers/net/ethernet/freescale/fman/fman_mac.h 	hlist_entry_safe(ptr, struct eth_hash_entry, node)
node              168 drivers/net/ethernet/freescale/fman/fman_mac.h 	struct list_head node;
node              219 drivers/net/ethernet/freescale/fman/fman_mac.h 		list_del_init(&hash_entry->node);
node              935 drivers/net/ethernet/freescale/fman/fman_memac.c 	INIT_LIST_HEAD(&hash_entry->node);
node              937 drivers/net/ethernet/freescale/fman/fman_memac.c 	list_add_tail(&hash_entry->node,
node              990 drivers/net/ethernet/freescale/fman/fman_memac.c 			list_del_init(&hash_entry->node);
node              560 drivers/net/ethernet/freescale/fman/fman_tgec.c 	INIT_LIST_HEAD(&hash_entry->node);
node              562 drivers/net/ethernet/freescale/fman/fman_tgec.c 	list_add_tail(&hash_entry->node,
node              634 drivers/net/ethernet/freescale/fman/fman_tgec.c 			list_del_init(&hash_entry->node);
node              196 drivers/net/ethernet/freescale/ucc_geth.c 		struct list_head *node = lh->next;
node              197 drivers/net/ethernet/freescale/ucc_geth.c 		list_del(node);
node              199 drivers/net/ethernet/freescale/ucc_geth.c 		return node;
node             3916 drivers/net/ethernet/freescale/ucc_geth.c 	ugeth->node = np;
node             1055 drivers/net/ethernet/freescale/ucc_geth.h 	struct list_head node;
node             1058 drivers/net/ethernet/freescale/ucc_geth.h #define ENET_ADDR_CONT_ENTRY(ptr) list_entry(ptr, struct enet_addr_container, node)
node             1223 drivers/net/ethernet/freescale/ucc_geth.h 	struct device_node *node;
node              904 drivers/net/ethernet/hisilicon/hip04_eth.c 	struct device_node *node = d->of_node;
node              935 drivers/net/ethernet/hisilicon/hip04_eth.c 	ret = of_parse_phandle_with_fixed_args(node, "port-handle", 3, 0, &arg);
node              964 drivers/net/ethernet/hisilicon/hip04_eth.c 	priv->phy_mode = of_get_phy_mode(node);
node              984 drivers/net/ethernet/hisilicon/hip04_eth.c 	priv->phy_node = of_parse_phandle(node, "phy-handle", 0);
node              783 drivers/net/ethernet/hisilicon/hisi_femac.c 	struct device_node *node = dev->of_node;
node              837 drivers/net/ethernet/hisilicon/hisi_femac.c 		ret = of_property_read_u32_array(node,
node              846 drivers/net/ethernet/hisilicon/hisi_femac.c 	phy = of_phy_get_and_connect(ndev, node, hisi_femac_adjust_link);
node              857 drivers/net/ethernet/hisilicon/hisi_femac.c 	mac_addr = of_get_mac_address(node);
node             1096 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	struct device_node *node = dev->of_node;
node             1169 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 		ret = of_property_read_u32_array(node,
node             1192 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	ret = of_mdiobus_register(bus, node);
node             1196 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	priv->phy_mode = of_get_phy_mode(node);
node             1203 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	priv->phy_node = of_parse_phandle(node, "phy-handle", 0);
node             1224 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	mac_addr = of_get_mac_address(node);
node               18 drivers/net/ethernet/hisilicon/hns/hnae.c hnae_list_add(spinlock_t *lock, struct list_head *node, struct list_head *head)
node               23 drivers/net/ethernet/hisilicon/hns/hnae.c 	list_add_tail_rcu(node, head);
node               27 drivers/net/ethernet/hisilicon/hns/hnae.c static void hnae_list_del(spinlock_t *lock, struct list_head *node)
node               32 drivers/net/ethernet/hisilicon/hns/hnae.c 	list_del_rcu(node);
node              356 drivers/net/ethernet/hisilicon/hns/hnae.c 	hnae_list_add(&dev->lock, &handle->node, &dev->handle_list);
node              381 drivers/net/ethernet/hisilicon/hns/hnae.c 	hnae_list_del(&dev->lock, &h->node);
node              540 drivers/net/ethernet/hisilicon/hns/hnae.h 	struct list_head node;
node              565 drivers/net/ethernet/hisilicon/hns/hnae.h 	struct list_head node;    /* list to hnae_ae_dev->handle_list */
node             2587 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c static char *hns_dsaf_get_node_stats_strings(char *data, int node,
node             2594 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_pad_drop_pkts", node);
node             2596 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_manage_pkts", node);
node             2598 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pkts", node);
node             2600 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pkt_id", node);
node             2602 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_rx_pause_frame", node);
node             2604 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_release_buf_num", node);
node             2606 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_sbm_drop_pkts", node);
node             2608 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_crc_false_pkts", node);
node             2610 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_bp_drop_pkts", node);
node             2612 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_lookup_rslt_drop_pkts", node);
node             2614 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_local_rslt_fail_pkts", node);
node             2616 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_vlan_drop_pkts", node);
node             2618 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "innod%d_stp_drop_pkts", node);
node             2620 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	if (node < DSAF_SERVICE_NW_NUM && !is_ver1) {
node             2624 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 				 node, i);
node             2627 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 				 node, i);
node             2632 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	snprintf(buff, ETH_GSTRING_LEN, "onnod%d_tx_pkts", node);
node             2719 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	int node = port;
node             2725 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	buff = hns_dsaf_get_node_stats_strings(buff, node, dsaf_dev);
node             2728 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	node = port + DSAF_PPE_INODE_BASE;
node             2729 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c 	(void)hns_dsaf_get_node_stats_strings(buff, node, dsaf_dev);
node              113 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(client_tmp, &hnae3_client_list, node) {
node              118 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_add_tail(&client->node, &hnae3_client_list);
node              121 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) {
node              150 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(client_tmp, &hnae3_client_list, node) {
node              164 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) {
node              168 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_del(&client->node);
node              189 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_add_tail(&ae_algo->node, &hnae3_ae_algo_list);
node              192 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) {
node              216 drivers/net/ethernet/hisilicon/hns3/hnae3.c 		list_for_each_entry(client, &hnae3_client_list, node) {
node              243 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_dev, &hnae3_ae_dev_list, node) {
node              254 drivers/net/ethernet/hisilicon/hns3/hnae3.c 		list_for_each_entry(client, &hnae3_client_list, node)
node              262 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_del(&ae_algo->node);
node              283 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_add_tail(&ae_dev->node, &hnae3_ae_dev_list);
node              286 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_algo, &hnae3_ae_algo_list, node) {
node              313 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(client, &hnae3_client_list, node) {
node              326 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_del(&ae_dev->node);
node              347 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_for_each_entry(ae_algo, &hnae3_ae_algo_list, node) {
node              355 drivers/net/ethernet/hisilicon/hns3/hnae3.c 		list_for_each_entry(client, &hnae3_client_list, node)
node              363 drivers/net/ethernet/hisilicon/hns3/hnae3.c 	list_del(&ae_dev->node);
node              211 drivers/net/ethernet/hisilicon/hns3/hnae3.h #define HNAE3_IS_TX_RING(node) \
node              212 drivers/net/ethernet/hisilicon/hns3/hnae3.h 	(((node)->flag & (1 << HNAE3_RING_TYPE_B)) == HNAE3_RING_TYPE_TX)
node              231 drivers/net/ethernet/hisilicon/hns3/hnae3.h 	struct list_head node;
node              237 drivers/net/ethernet/hisilicon/hns3/hnae3.h 	struct list_head node;
node              552 drivers/net/ethernet/hisilicon/hns3/hnae3.h 	struct list_head node;
node             4555 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	INIT_LIST_HEAD(&client.node);
node             4494 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	struct hnae3_ring_chain_node *node;
node             4508 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	for (node = ring_chain; node; node = node->next) {
node             4512 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 				hnae3_get_bit(node->flag, HNAE3_RING_TYPE_B));
node             4514 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 				HCLGE_TQP_ID_S, node->tqp_index);
node             4517 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 				hnae3_get_field(node->int_gl_idx,
node             5681 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	struct hlist_node *node;
node             5694 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 		hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list,
node             5713 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	struct hlist_node *node;
node             5728 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) {
node             6034 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	struct hlist_node *node;
node             6036 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) {
node             6153 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	struct hlist_node *node;
node             6161 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) {
node             6172 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	hlist_for_each_entry_safe(rule, node, &del_list, rule_node) {
node             7326 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_add_tail(&mac_cfg->node, list);
node             7343 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_for_each_entry_safe(mac_cfg, tmp, list, node) {
node             7351 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&mac_cfg->node);
node             7367 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_for_each_entry_safe(mac_cfg, tmp, list, node) {
node             7376 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&mac_cfg->node);
node             7391 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 		list_for_each_entry_safe(mac, tmp, &vport->uc_mac_list, node) {
node             7392 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&mac->node);
node             7396 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 		list_for_each_entry_safe(mac, tmp, &vport->mc_mac_list, node) {
node             7397 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&mac->node);
node             8016 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_add_tail(&vlan->node, &vport->vlan_list);
node             8025 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) {
node             8049 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) {
node             8058 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&vlan->node);
node             8070 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 	list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) {
node             8080 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&vlan->node);
node             8095 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 		list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) {
node             8096 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 			list_del(&vlan->node);
node             8126 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c 		list_for_each_entry_safe(vlan, tmp, &vport->vlan_list, node) {
node              637 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h 	struct list_head node;
node              648 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h 	struct list_head node;
node              997 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 	struct hnae3_ring_chain_node *node;
node             1008 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 	for (node = ring_chain; node; node = node->next) {
node             1021 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 				hnae3_get_bit(node->flag, HNAE3_RING_TYPE_B);
node             1022 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 		req->msg[idx_offset + 1] = node->tqp_index;
node             1023 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 		req->msg[idx_offset + 2] = hnae3_get_field(node->int_gl_idx,
node             1031 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c 		    !node->next) {
node              653 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	struct hinic_api_cmd_cell *node;
node              657 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	node = dma_alloc_coherent(&pdev->dev, chain->cell_size, &node_paddr,
node              659 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	if (!node) {
node              664 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	node->read.hw_wb_resp_paddr = 0;
node              667 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	cell_ctxt->cell_vaddr = node;
node              672 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 		chain->head_node = node;
node              680 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 		err = alloc_cmd_buf(chain, node, cell_idx);
node              693 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	*node_vaddr = node;
node              697 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	dma_free_coherent(&pdev->dev, chain->cell_size, node, node_paddr);
node              712 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	struct hinic_api_cmd_cell *node;
node              718 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	node = cell_ctxt->cell_vaddr;
node              732 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 		dma_free_coherent(&pdev->dev, node_size, node,
node              759 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	struct hinic_api_cmd_cell *node = NULL, *pre_node = NULL;
node              765 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 		err = api_cmd_create_cell(chain, cell_idx, pre_node, &node);
node              771 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 		pre_node = node;
node              775 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	node->next_cell_paddr = cpu_to_be64(chain->head_cell_paddr);
node              322 drivers/net/ethernet/huawei/hinic/hinic_main.c 	int i, node, err = 0;
node              351 drivers/net/ethernet/huawei/hinic/hinic_main.c 		node = cpu_to_node(i);
node              352 drivers/net/ethernet/huawei/hinic/hinic_main.c 		if (node == dev_to_node(&pdev->dev))
node             2340 drivers/net/ethernet/ibm/emac/core.c 	struct device_node	*node;
node             2373 drivers/net/ethernet/ibm/emac/core.c 			if (deps[i].node == NULL)
node             2374 drivers/net/ethernet/ibm/emac/core.c 				deps[i].node = of_node_get(np);
node             2376 drivers/net/ethernet/ibm/emac/core.c 		if (deps[i].node == NULL)
node             2377 drivers/net/ethernet/ibm/emac/core.c 			deps[i].node = of_find_node_by_phandle(deps[i].phandle);
node             2378 drivers/net/ethernet/ibm/emac/core.c 		if (deps[i].node == NULL)
node             2381 drivers/net/ethernet/ibm/emac/core.c 			deps[i].ofdev = of_find_device_by_node(deps[i].node);
node             2437 drivers/net/ethernet/ibm/emac/core.c 		of_node_put(deps[i].node);
node             3476 drivers/net/ethernet/intel/i40e/i40e_main.c 	struct hlist_node *node;
node             3487 drivers/net/ethernet/intel/i40e/i40e_main.c 	hlist_for_each_entry_safe(filter, node,
node             5560 drivers/net/ethernet/intel/i40e/i40e_main.c 	struct hlist_node *node;
node             5602 drivers/net/ethernet/intel/i40e/i40e_main.c 		hlist_for_each_entry_safe(cfilter, node,
node             8369 drivers/net/ethernet/intel/i40e/i40e_main.c 	struct hlist_node *node;
node             8371 drivers/net/ethernet/intel/i40e/i40e_main.c 	hlist_for_each_entry_safe(cfilter, node,
node             8846 drivers/net/ethernet/intel/i40e/i40e_main.c 	struct hlist_node *node;
node             8869 drivers/net/ethernet/intel/i40e/i40e_main.c 		hlist_for_each_entry_safe(filter, node,
node             9699 drivers/net/ethernet/intel/i40e/i40e_main.c 	struct hlist_node *node;
node             9703 drivers/net/ethernet/intel/i40e/i40e_main.c 	hlist_for_each_entry_safe(cfilter, node, &pf->cloud_filter_list,
node             3281 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	struct hlist_node *node;
node             3284 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	hlist_for_each_entry_safe(cfilter, node,
node             3327 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	struct hlist_node *node;
node             3410 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	hlist_for_each_entry_safe(cf, node,
node             3193 drivers/net/ethernet/intel/ice/ice_common.c 	struct ice_aqc_txsched_elem_data node = { 0 };
node             3229 drivers/net/ethernet/intel/ice/ice_common.c 	node.parent_teid = parent->info.node_teid;
node             3252 drivers/net/ethernet/intel/ice/ice_common.c 	node.node_teid = buf->txqs[0].q_teid;
node             3253 drivers/net/ethernet/intel/ice/ice_common.c 	node.data.elem_type = ICE_AQC_ELEM_TYPE_LEAF;
node             3257 drivers/net/ethernet/intel/ice/ice_common.c 	status = ice_sched_add_node(pi, hw->num_tx_sched_layers - 1, &node);
node             3307 drivers/net/ethernet/intel/ice/ice_common.c 		struct ice_sched_node *node;
node             3309 drivers/net/ethernet/intel/ice/ice_common.c 		node = ice_sched_find_node_by_teid(pi->root, q_teids[i]);
node             3310 drivers/net/ethernet/intel/ice/ice_common.c 		if (!node)
node             3323 drivers/net/ethernet/intel/ice/ice_common.c 		qg_list.parent_teid = node->info.parent_teid;
node             3332 drivers/net/ethernet/intel/ice/ice_common.c 		ice_free_sched_node(pi, node);
node             1364 drivers/net/ethernet/intel/ice/ice_dcb.c 	struct ice_sched_node *node, *tc_node;
node             1411 drivers/net/ethernet/intel/ice/ice_dcb.c 		node = ice_sched_find_node_by_teid(pi->root, teid2);
node             1412 drivers/net/ethernet/intel/ice/ice_dcb.c 		if (node)
node             1413 drivers/net/ethernet/intel/ice/ice_dcb.c 			node->tc_num = j;
node              154 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *node;
node              180 drivers/net/ethernet/intel/ice/ice_sched.c 	node = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*node), GFP_KERNEL);
node              181 drivers/net/ethernet/intel/ice/ice_sched.c 	if (!node)
node              185 drivers/net/ethernet/intel/ice/ice_sched.c 		node->children = devm_kcalloc(ice_hw_to_dev(hw),
node              187 drivers/net/ethernet/intel/ice/ice_sched.c 					      sizeof(*node), GFP_KERNEL);
node              188 drivers/net/ethernet/intel/ice/ice_sched.c 		if (!node->children) {
node              189 drivers/net/ethernet/intel/ice/ice_sched.c 			devm_kfree(ice_hw_to_dev(hw), node);
node              194 drivers/net/ethernet/intel/ice/ice_sched.c 	node->in_use = true;
node              195 drivers/net/ethernet/intel/ice/ice_sched.c 	node->parent = parent;
node              196 drivers/net/ethernet/intel/ice/ice_sched.c 	node->tx_sched_layer = layer;
node              197 drivers/net/ethernet/intel/ice/ice_sched.c 	parent->children[parent->num_children++] = node;
node              198 drivers/net/ethernet/intel/ice/ice_sched.c 	memcpy(&node->info, &elem.generic[0], sizeof(node->info));
node              304 drivers/net/ethernet/intel/ice/ice_sched.c void ice_free_sched_node(struct ice_port_info *pi, struct ice_sched_node *node)
node              314 drivers/net/ethernet/intel/ice/ice_sched.c 	while (node->num_children)
node              315 drivers/net/ethernet/intel/ice/ice_sched.c 		ice_free_sched_node(pi, node->children[0]);
node              318 drivers/net/ethernet/intel/ice/ice_sched.c 	if (node->tx_sched_layer >= hw->sw_entry_point_layer &&
node              319 drivers/net/ethernet/intel/ice/ice_sched.c 	    node->info.data.elem_type != ICE_AQC_ELEM_TYPE_TC &&
node              320 drivers/net/ethernet/intel/ice/ice_sched.c 	    node->info.data.elem_type != ICE_AQC_ELEM_TYPE_ROOT_PORT &&
node              321 drivers/net/ethernet/intel/ice/ice_sched.c 	    node->info.data.elem_type != ICE_AQC_ELEM_TYPE_LEAF) {
node              322 drivers/net/ethernet/intel/ice/ice_sched.c 		u32 teid = le32_to_cpu(node->info.node_teid);
node              324 drivers/net/ethernet/intel/ice/ice_sched.c 		ice_sched_remove_elems(hw, node->parent, 1, &teid);
node              326 drivers/net/ethernet/intel/ice/ice_sched.c 	parent = node->parent;
node              333 drivers/net/ethernet/intel/ice/ice_sched.c 			if (parent->children[i] == node) {
node              341 drivers/net/ethernet/intel/ice/ice_sched.c 		p = ice_sched_get_first_node(pi, node, node->tx_sched_layer);
node              343 drivers/net/ethernet/intel/ice/ice_sched.c 			if (p->sibling == node) {
node              344 drivers/net/ethernet/intel/ice/ice_sched.c 				p->sibling = node->sibling;
node              351 drivers/net/ethernet/intel/ice/ice_sched.c 		if (pi->sib_head[node->tc_num][node->tx_sched_layer] == node)
node              352 drivers/net/ethernet/intel/ice/ice_sched.c 			pi->sib_head[node->tc_num][node->tx_sched_layer] =
node              353 drivers/net/ethernet/intel/ice/ice_sched.c 				node->sibling;
node              357 drivers/net/ethernet/intel/ice/ice_sched.c 	if (node->children)
node              358 drivers/net/ethernet/intel/ice/ice_sched.c 		devm_kfree(ice_hw_to_dev(hw), node->children);
node              359 drivers/net/ethernet/intel/ice/ice_sched.c 	devm_kfree(ice_hw_to_dev(hw), node);
node              878 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *node;
node              880 drivers/net/ethernet/intel/ice/ice_sched.c 	node = pi->root;
node              881 drivers/net/ethernet/intel/ice/ice_sched.c 	while (node) {
node              882 drivers/net/ethernet/intel/ice/ice_sched.c 		if (!node->num_children)
node              884 drivers/net/ethernet/intel/ice/ice_sched.c 		node = node->children[0];
node              886 drivers/net/ethernet/intel/ice/ice_sched.c 	if (node && node->info.data.elem_type == ICE_AQC_ELEM_TYPE_LEAF) {
node              887 drivers/net/ethernet/intel/ice/ice_sched.c 		u32 teid = le32_to_cpu(node->info.node_teid);
node              891 drivers/net/ethernet/intel/ice/ice_sched.c 		status = ice_sched_remove_elems(pi->hw, node->parent, 1, &teid);
node              893 drivers/net/ethernet/intel/ice/ice_sched.c 			ice_free_sched_node(pi, node);
node              906 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *node;
node              911 drivers/net/ethernet/intel/ice/ice_sched.c 	node = pi->root;
node              912 drivers/net/ethernet/intel/ice/ice_sched.c 	while (node) {
node              913 drivers/net/ethernet/intel/ice/ice_sched.c 		if (node->tx_sched_layer >= pi->hw->sw_entry_point_layer &&
node              914 drivers/net/ethernet/intel/ice/ice_sched.c 		    node->info.data.elem_type != ICE_AQC_ELEM_TYPE_TC &&
node              915 drivers/net/ethernet/intel/ice/ice_sched.c 		    node->info.data.elem_type != ICE_AQC_ELEM_TYPE_ROOT_PORT) {
node              916 drivers/net/ethernet/intel/ice/ice_sched.c 			ice_free_sched_node(pi, node);
node              920 drivers/net/ethernet/intel/ice/ice_sched.c 		if (!node->num_children)
node              922 drivers/net/ethernet/intel/ice/ice_sched.c 		node = node->children[0];
node             1095 drivers/net/ethernet/intel/ice/ice_sched.c 			       struct ice_sched_node *node)
node             1102 drivers/net/ethernet/intel/ice/ice_sched.c 		if (node == child)
node             1105 drivers/net/ethernet/intel/ice/ice_sched.c 		if (child->tx_sched_layer > node->tx_sched_layer)
node             1111 drivers/net/ethernet/intel/ice/ice_sched.c 		if (ice_sched_find_node_in_subtree(hw, child, node))
node             1174 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *node;
node             1178 drivers/net/ethernet/intel/ice/ice_sched.c 	node = ice_sched_get_first_node(hw->port_info, tc_node, vsi_layer);
node             1181 drivers/net/ethernet/intel/ice/ice_sched.c 	while (node) {
node             1182 drivers/net/ethernet/intel/ice/ice_sched.c 		if (node->vsi_handle == vsi_handle)
node             1183 drivers/net/ethernet/intel/ice/ice_sched.c 			return node;
node             1184 drivers/net/ethernet/intel/ice/ice_sched.c 		node = node->sibling;
node             1187 drivers/net/ethernet/intel/ice/ice_sched.c 	return node;
node             1234 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *parent, *node;
node             1261 drivers/net/ethernet/intel/ice/ice_sched.c 			node = parent;
node             1262 drivers/net/ethernet/intel/ice/ice_sched.c 			while (node) {
node             1263 drivers/net/ethernet/intel/ice/ice_sched.c 				node->owner = owner;
node             1264 drivers/net/ethernet/intel/ice/ice_sched.c 				node = node->sibling;
node             1288 drivers/net/ethernet/intel/ice/ice_sched.c 	struct ice_sched_node *node;
node             1303 drivers/net/ethernet/intel/ice/ice_sched.c 			node = ice_sched_get_first_node(hw->port_info, tc_node,
node             1306 drivers/net/ethernet/intel/ice/ice_sched.c 			while (node) {
node             1307 drivers/net/ethernet/intel/ice/ice_sched.c 				if (node->num_children < hw->max_children[i])
node             1309 drivers/net/ethernet/intel/ice/ice_sched.c 				node = node->sibling;
node             1316 drivers/net/ethernet/intel/ice/ice_sched.c 			if (node)
node             1578 drivers/net/ethernet/intel/ice/ice_sched.c static bool ice_sched_is_leaf_node_present(struct ice_sched_node *node)
node             1582 drivers/net/ethernet/intel/ice/ice_sched.c 	for (i = 0; i < node->num_children; i++)
node             1583 drivers/net/ethernet/intel/ice/ice_sched.c 		if (ice_sched_is_leaf_node_present(node->children[i]))
node             1586 drivers/net/ethernet/intel/ice/ice_sched.c 	return (node->info.data.elem_type == ICE_AQC_ELEM_TYPE_LEAF);
node               42 drivers/net/ethernet/intel/ice/ice_sched.h void ice_free_sched_node(struct ice_port_info *pi, struct ice_sched_node *node);
node              837 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	int node = NUMA_NO_NODE;
node              850 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				node = cpu_to_node(cpu);
node              857 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				GFP_KERNEL, node);
node              867 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	q_vector->numa_node = node;
node             3728 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_online(unsigned int cpu, struct hlist_node *node)
node             3731 drivers/net/ethernet/marvell/mvneta.c 	struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port,
node             3786 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_down_prepare(unsigned int cpu, struct hlist_node *node)
node             3788 drivers/net/ethernet/marvell/mvneta.c 	struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port,
node             3808 drivers/net/ethernet/marvell/mvneta.c static int mvneta_cpu_dead(unsigned int cpu, struct hlist_node *node)
node             3810 drivers/net/ethernet/marvell/mvneta.c 	struct mvneta_port *pp = hlist_entry_safe(node, struct mvneta_port,
node              396 drivers/net/ethernet/marvell/mvneta_bm.c struct mvneta_bm *mvneta_bm_get(struct device_node *node)
node              398 drivers/net/ethernet/marvell/mvneta_bm.c 	struct platform_device *pdev = of_find_device_by_node(node);
node              134 drivers/net/ethernet/marvell/mvneta_bm.h struct mvneta_bm *mvneta_bm_get(struct device_node *node);
node              188 drivers/net/ethernet/marvell/mvneta_bm.h static inline struct mvneta_bm *mvneta_bm_get(struct device_node *node)
node              352 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	int node;
node              435 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	int node;
node               66 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	struct hlist_node	node;
node             1659 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	hlist_for_each_entry(mce, &mce_list->head, node) {
node             1669 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 		hlist_del(&mce->node);
node             1685 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 		hlist_add_head(&mce->node, &mce_list->head);
node             1687 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 		hlist_add_behind(&mce->node, &tail->node);
node             1742 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	hlist_for_each_entry(mce, &mce_list->head, node) {
node             1746 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 			next_mce = hlist_entry(mce->node.next,
node             1747 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 					       struct mce, node);
node               49 drivers/net/ethernet/mellanox/mlx4/en_cq.c 		      int node)
node               55 drivers/net/ethernet/mellanox/mlx4/en_cq.c 	cq = kzalloc_node(sizeof(*cq), GFP_KERNEL, node);
node               71 drivers/net/ethernet/mellanox/mlx4/en_cq.c 	set_dev_node(&mdev->dev->persist->pdev->dev, node);
node             2143 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 	int node;
node             2148 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 			node = cpu_to_node(i % num_online_cpus());
node             2150 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 					      prof->tx_ring_size, i, t, node))
node             2155 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 						   TXBB_SIZE, node, i))
node             2162 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 		node = cpu_to_node(i % num_online_cpus());
node             2164 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 				      prof->rx_ring_size, i, RX, node))
node             2169 drivers/net/ethernet/mellanox/mlx4/en_netdev.c 					   node, i))
node              265 drivers/net/ethernet/mellanox/mlx4/en_rx.c 			   u32 size, u16 stride, int node, int queue_index)
node              272 drivers/net/ethernet/mellanox/mlx4/en_rx.c 	ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node);
node              291 drivers/net/ethernet/mellanox/mlx4/en_rx.c 	ring->rx_info = kvzalloc_node(tmp, GFP_KERNEL, node);
node              301 drivers/net/ethernet/mellanox/mlx4/en_rx.c 	set_dev_node(&mdev->dev->persist->pdev->dev, node);
node               52 drivers/net/ethernet/mellanox/mlx4/en_tx.c 			   u16 stride, int node, int queue_index)
node               59 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node);
node               71 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	ring->tx_info = kvmalloc_node(tmp, GFP_KERNEL, node);
node               80 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	ring->bounce_buf = kmalloc_node(MAX_DESC_SIZE, GFP_KERNEL, node);
node               91 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	set_dev_node(&mdev->dev->persist->pdev->dev, node);
node              120 drivers/net/ethernet/mellanox/mlx4/en_tx.c 	err = mlx4_bf_alloc(mdev->dev, &ring->bf, node);
node               99 drivers/net/ethernet/mellanox/mlx4/icm.c 				gfp_t gfp_mask, int node)
node              103 drivers/net/ethernet/mellanox/mlx4/icm.c 	page = alloc_pages_node(node, gfp_mask, order);
node              691 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h 		      int entries, int ring, enum cq_type mode, int node);
node              714 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h 			   int node, int queue_index);
node              728 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h 			   u32 size, u16 stride, int node, int queue_index);
node              172 drivers/net/ethernet/mellanox/mlx4/pd.c int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf, int node)
node              190 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node);
node               73 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct rb_node		node;
node              237 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	struct rb_node *node = root->rb_node;
node              239 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	while (node) {
node              240 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		struct res_common *res = rb_entry(node, struct res_common,
node              241 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 						  node);
node              244 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			node = node->rb_left;
node              246 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			node = node->rb_right;
node              260 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 						   node);
node              272 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	rb_link_node(&res->node, parent, new);
node              273 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	rb_insert_color(&res->node, root);
node             1322 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		rb_erase(&res_arr[i]->node, root);
node             1492 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		rb_erase(&r->node, &tracker->res_tree[type]);
node             4688 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&qp->com.node,
node             4761 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&srq->com.node,
node             4826 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&cq->com.node,
node             4888 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&mpt->com.node,
node             4956 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&mtt->com.node,
node             5036 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		fs_rule = rb_entry(p, struct res_fs_rule, com.node);
node             5095 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&fs_rule->com.node,
node             5141 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					rb_erase(&eq->com.node,
node             5200 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 				rb_erase(&counter->com.node,
node             5237 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			rb_erase(&xrcd->com.node, &tracker->res_tree[RES_XRCD]);
node               57 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 					   int node)
node               66 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 	set_dev_node(device, node);
node               75 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 			struct mlx5_frag_buf *buf, int node)
node               88 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 							  &t, node);
node              122 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 			     struct mlx5_frag_buf *buf, int node)
node              139 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 							  &frag->map, node);
node              181 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 						 int node)
node              199 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 						       &pgdir->db_dma, node);
node              234 drivers/net/ethernet/mellanox/mlx5/core/alloc.c int mlx5_db_alloc_node(struct mlx5_core_dev *dev, struct mlx5_db *db, int node)
node              245 drivers/net/ethernet/mellanox/mlx5/core/alloc.c 	pgdir = mlx5_alloc_db_pgdir(dev, node);
node              115 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h 			   fs_get_obj(__entry->ft, fg->node.parent);
node              203 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h 			   fs_get_obj(__entry->fg, fte->node.parent);
node              286 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h 			   fs_get_obj(__entry->fte, rule->node.parent);
node              312 drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.h 			   fs_get_obj(__entry->fte, rule->node.parent);
node              113 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	struct rhash_head	node;
node             3484 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	.head_offset = offsetof(struct mlx5e_tc_flow, node),
node             3825 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	err = rhashtable_lookup_insert_fast(tc_ht, &flow->node, tc_ht_params);
node             3867 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	rhashtable_remove_fast(tc_ht, &flow->node, tc_ht_params);
node               52 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	struct l2addr_node     node;
node              497 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	u8 *mac = vaddr->node.addr;
node              529 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	u8 *mac = vaddr->node.addr;
node              558 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	u8 *mac = vaddr->node.addr;
node              606 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	u8 *mac = vaddr->node.addr;
node              646 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	u8 *mac = vaddr->node.addr;
node              692 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	struct l2addr_node *node;
node              703 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	for_each_l2hash_node(node, tmp, hash, hi) {
node              704 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 		addr = container_of(node, struct vport_addr, node);
node              724 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	struct l2addr_node *node;
node              742 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	for_each_l2hash_node(node, tmp, hash, hi) {
node              743 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 		addr = container_of(node, struct vport_addr, node);
node              809 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	struct l2addr_node *node;
node              817 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 	for_each_l2hash_node(node, tmp, esw->mc_table, hi) {
node              818 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c 		u8 *mac = node->addr;
node              198 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h 	struct l2addr_node     node;
node               80 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	struct rb_node			node;
node              764 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	fs_get_obj(fg, fte->node.parent);
node              862 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	struct rb_node *node = root->rb_node;
node              864 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	while (node) {
node              866 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 				container_of(node, struct mlx5_fpga_ipsec_rule,
node              867 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 					     node);
node              870 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 			node = node->rb_left;
node              872 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 			node = node->rb_right;
node              900 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 					     node);
node              912 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	rb_link_node(&rule->node, parent, new);
node              913 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	rb_insert_color(&rule->node, root);
node              936 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	rb_erase(&rule->node, root);
node             1320 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 	rbtree_postorder_for_each_entry_safe(r, tmp, root, node) {
node             1321 drivers/net/ethernet/mellanox/mlx5/core/fpga/ipsec.c 		rb_erase(&r->node, root);
node              376 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c 	list_for_each_entry(dst, &fte->node.children, node.list) {
node              487 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c 		list_for_each_entry(dst, &fte->node.children, node.list) {
node              542 drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c 		list_for_each_entry(dst, &fte->node.children, node.list) {
node              228 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_table(struct fs_node *node);
node              229 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_group(struct fs_node *node);
node              230 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_fte(struct fs_node *node);
node              231 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_table(struct fs_node *node);
node              232 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_group(struct fs_node *node);
node              233 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_fte(struct fs_node *node);
node              234 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_prio(struct fs_node *node);
node              235 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_ns(struct fs_node *node);
node              239 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_hw_rule(struct fs_node *node);
node              247 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_init_node(struct fs_node *node,
node              251 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	refcount_set(&node->refcount, 1);
node              252 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	INIT_LIST_HEAD(&node->list);
node              253 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	INIT_LIST_HEAD(&node->children);
node              254 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	init_rwsem(&node->lock);
node              255 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	node->del_hw_func = del_hw_func;
node              256 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	node->del_sw_func = del_sw_func;
node              257 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	node->active = false;
node              260 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_add_node(struct fs_node *node, struct fs_node *parent)
node              264 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	node->parent = parent;
node              268 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		node->root = node;
node              270 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		node->root = parent->root;
node              273 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static int tree_get_node(struct fs_node *node)
node              275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	return refcount_inc_not_zero(&node->refcount);
node              278 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void nested_down_read_ref_node(struct fs_node *node,
node              281 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node) {
node              282 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		down_read_nested(&node->lock, class);
node              283 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		refcount_inc(&node->refcount);
node              287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void nested_down_write_ref_node(struct fs_node *node,
node              290 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node) {
node              291 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		down_write_nested(&node->lock, class);
node              292 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		refcount_inc(&node->refcount);
node              296 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void down_write_ref_node(struct fs_node *node, bool locked)
node              298 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node) {
node              300 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			down_write(&node->lock);
node              301 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		refcount_inc(&node->refcount);
node              305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void up_read_ref_node(struct fs_node *node)
node              307 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	refcount_dec(&node->refcount);
node              308 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_read(&node->lock);
node              311 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void up_write_ref_node(struct fs_node *node, bool locked)
node              313 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	refcount_dec(&node->refcount);
node              315 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write(&node->lock);
node              318 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void tree_put_node(struct fs_node *node, bool locked)
node              320 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct fs_node *parent_node = node->parent;
node              322 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (refcount_dec_and_test(&node->refcount)) {
node              323 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		if (node->del_hw_func)
node              324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			node->del_hw_func(node);
node              330 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			list_del_init(&node->list);
node              331 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			if (node->del_sw_func)
node              332 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				node->del_sw_func(node);
node              335 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			kfree(node);
node              337 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		node = NULL;
node              339 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (!node && parent_node)
node              343 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static int tree_remove_node(struct fs_node *node, bool locked)
node              345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (refcount_read(&node->refcount) > 1) {
node              346 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		refcount_dec(&node->refcount);
node              349 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_put_node(node, locked);
node              379 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static struct mlx5_flow_root_namespace *find_root(struct fs_node *node)
node              384 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = node->root;
node              391 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	ns = container_of(root, struct mlx5_flow_namespace, node);
node              395 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static inline struct mlx5_flow_steering *get_steering(struct fs_node *node)
node              397 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(node);
node              404 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static inline struct mlx5_core_dev *get_dev(struct fs_node *node)
node              406 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(node);
node              413 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_ns(struct fs_node *node)
node              415 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	kfree(node);
node              418 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_prio(struct fs_node *node)
node              420 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	kfree(node);
node              423 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_table(struct fs_node *node)
node              430 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, node);
node              431 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	dev = get_dev(&ft->node);
node              432 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ft->node);
node              435 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node->active) {
node              442 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_table(struct fs_node *node)
node              447 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, node);
node              450 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(prio, ft->node.parent);
node              463 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, fte->node.parent);
node              464 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node              465 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	dev = get_dev(&fte->node);
node              467 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ft->node);
node              476 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_hw_rule(struct fs_node *node)
node              481 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(rule, node);
node              482 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fte, rule->node.parent);
node              508 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_fte(struct fs_node *node)
node              517 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fte, node);
node              518 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, fte->node.parent);
node              519 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node              522 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	dev = get_dev(&ft->node);
node              523 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ft->node);
node              524 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node->active) {
node              530 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		node->active = 0;
node              534 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_fte(struct fs_node *node)
node              536 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(node);
node              541 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fte, node);
node              542 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, fte->node.parent);
node              552 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_hw_flow_group(struct fs_node *node)
node              559 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, node);
node              560 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node              561 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	dev = get_dev(&ft->node);
node              564 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ft->node);
node              565 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (fg->node.active && root->cmds->destroy_flow_group(root, ft, fg))
node              570 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void del_sw_flow_group(struct fs_node *node)
node              572 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(node);
node              577 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, node);
node              578 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node              607 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&fte->node, &fg->node);
node              608 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add_tail(&fte->node.list, &fg->node.children);
node              620 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(&ft->node);
node              628 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fte->node.type =  FS_TYPE_FLOW_ENTRY;
node              632 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&fte->node, NULL, del_sw_fte);
node              667 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fg->node.type =  FS_TYPE_FLOW_GROUP;
node              681 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(&ft->node);
node              699 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&fg->node, del_hw_flow_group, del_sw_flow_group);
node              700 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&fg->node, &ft->node);
node              702 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add(&fg->node.list, prev);
node              703 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	atomic_inc(&ft->node.version);
node              727 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	ft->node.type = FS_TYPE_FLOW_TABLE;
node              784 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	parent = prio->node.parent;
node              785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	curr_node = &prio->node;
node              810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&prio->node);
node              841 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		fs_get_obj(prev_prio, prev_ft->node.parent);
node              850 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&prio->node);
node              896 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fte, rule->node.parent);
node              899 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	down_write_ref_node(&fte->node, false);
node              900 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fg, fte->node.parent);
node              901 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node              904 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ft->node);
node              907 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&fte->node, false);
node              972 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (list_empty(&prio->node.children)) {
node              992 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct list_head *prev = &prio->node.children;
node              998 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		prev = &iter->node.list;
node             1000 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add(&ft->node.list, prev);
node             1008 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ns->node);
node             1044 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&ft->node, del_hw_flow_table, del_sw_flow_table);
node             1055 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	ft->node.active = true;
node             1056 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	down_write_ref_node(&fs_prio->node, false);
node             1057 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&ft->node, &fs_prio->node);
node             1060 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&fs_prio->node, false);
node             1139 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ft->node);
node             1155 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	down_write_ref_node(&ft->node, false);
node             1158 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				     ft->node.children.prev);
node             1159 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&ft->node, false);
node             1165 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&fg->node, false);
node             1169 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fg->node.active = true;
node             1183 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	rule->node.type = FS_TYPE_FLOW_DEST;
node             1209 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		if (refcount_dec_and_test(&handle->rule[i]->node.refcount)) {
node             1211 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			list_del(&handle->rule[i]->node.list);
node             1240 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				refcount_inc(&rule->node.refcount);
node             1253 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_init_node(&rule->node, NULL, del_sw_hw_rule);
node             1256 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			list_add(&rule->node.list, &fte->node.children);
node             1258 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			list_add_tail(&rule->node.list, &fte->node.children);
node             1300 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(ft, fg->node.parent);
node             1301 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&fg->node);
node             1309 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fte->node.active = true;
node             1311 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	atomic_inc(&fte->node.version);
node             1324 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct list_head *prev = &ft->node.children;
node             1345 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		prev = &fg->node.list;
node             1370 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ft->node);
node             1402 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		fg->node.active = true;
node             1439 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_for_each_entry(rule, &fte->node.children, node.list) {
node             1473 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		mlx5_core_warn(get_dev(&fte->node),
node             1480 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		mlx5_core_warn(get_dev(&fte->node),
node             1517 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		if (refcount_read(&handle->rule[i]->node.refcount) == 1) {
node             1518 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			tree_add_node(&handle->rule[i]->node, &fte->node);
node             1564 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&head->first.g->node, ft_locked);
node             1567 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			tree_put_node(&iter->g->node, ft_locked);
node             1592 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			if (!tree_get_node(&g->node))
node             1606 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		if (!tree_get_node(&g->node)) {
node             1624 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		version += (u64)atomic_read(&iter->g->node.version);
node             1636 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		nested_down_write_ref_node(&g->node, FS_LOCK_PARENT);
node             1638 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		nested_down_read_ref_node(&g->node, FS_LOCK_PARENT);
node             1641 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (!fte_tmp || !tree_get_node(&fte_tmp->node)) {
node             1645 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (!fte_tmp->node.active) {
node             1646 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&fte_tmp->node, false);
node             1651 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	nested_down_write_ref_node(&fte_tmp->node, FS_LOCK_CHILD);
node             1654 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&g->node, false);
node             1656 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_read_ref_node(&g->node);
node             1669 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(&ft->node);
node             1695 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&fte_tmp->node, false);
node             1696 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&fte_tmp->node, false);
node             1709 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (atomic_read(&ft->node.version) != ft_version) {
node             1725 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		if (!g->node.active)
node             1728 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		nested_down_write_ref_node(&g->node, FS_LOCK_PARENT);
node             1732 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			up_write_ref_node(&g->node, false);
node             1739 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		nested_down_write_ref_node(&fte->node, FS_LOCK_CHILD);
node             1740 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&g->node, false);
node             1742 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&fte->node, false);
node             1743 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&fte->node, false);
node             1760 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_steering *steering = get_steering(&ft->node);
node             1777 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	nested_down_read_ref_node(&ft->node, FS_LOCK_GRANDPARENT);
node             1779 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	version = atomic_read(&ft->node.version);
node             1785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			up_write_ref_node(&ft->node, false);
node             1787 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			up_read_ref_node(&ft->node);
node             1792 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_read_ref_node(&ft->node);
node             1800 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			up_write_ref_node(&ft->node, false);
node             1805 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		nested_down_write_ref_node(&ft->node, FS_LOCK_GRANDPARENT);
node             1810 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	    version != atomic_read(&ft->node.version))
node             1816 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&ft->node, false);
node             1820 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	nested_down_write_ref_node(&g->node, FS_LOCK_PARENT);
node             1821 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&ft->node, false);
node             1839 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	nested_down_write_ref_node(&fte->node, FS_LOCK_CHILD);
node             1840 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&g->node, false);
node             1842 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&fte->node, false);
node             1843 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_put_node(&fte->node, false);
node             1844 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_put_node(&g->node, false);
node             1848 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	up_write_ref_node(&g->node, false);
node             1849 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_put_node(&g->node, false);
node             1856 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		(MLX5_CAP_FLOWTABLE(get_dev(&ft->node), nic_rx_multi_path_tirs)));
node             1866 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ft->node);
node             1873 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(prio, ft->node.parent);
node             1927 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(fte, handle->rule[0]->node.parent);
node             1928 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	down_write_ref_node(&fte->node, false);
node             1930 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_remove_node(&handle->rule[i]->node, true);
node             1933 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write_ref_node(&fte->node, false);
node             1935 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		del_hw_fte(&fte->node);
node             1936 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		up_write(&fte->node.lock);
node             1937 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(&fte->node, false);
node             1948 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(prio, ft->node.parent);
node             1950 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (!list_is_last(&ft->node.list, &prio->node.children))
node             1951 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		return list_next_entry(ft, node.list);
node             1957 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ft->node);
node             2003 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_core_dev *dev = get_dev(&ft->node);
node             2012 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_get_obj(prio, ft->node.parent);
node             2013 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if  (!(list_first_entry(&prio->node.children,
node             2015 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 				node.list) == ft))
node             2032 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct mlx5_flow_root_namespace *root = find_root(&ft->node);
node             2041 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (tree_remove_node(&ft->node, false))
node             2042 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		mlx5_core_warn(get_dev(&ft->node), "Flow table %d wasn't destroyed, refcount > 1\n",
node             2052 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (tree_remove_node(&fg->node, false))
node             2053 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		mlx5_core_warn(get_dev(&fg->node), "Flow group %d wasn't destroyed, refcount > 1\n",
node             2118 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	ns = list_first_entry(&fs_prio->node.children,
node             2120 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			      node.list);
node             2164 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	fs_prio->node.type = type;
node             2165 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&fs_prio->node, NULL, del_sw_prio);
node             2166 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&fs_prio->node, &ns->node);
node             2169 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add_tail(&fs_prio->node.list, &ns->node.children);
node             2190 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	ns->node.type = FS_TYPE_NAMESPACE;
node             2206 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&ns->node, NULL, del_sw_ns);
node             2207 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&ns->node, &prio->node);
node             2208 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	list_add_tail(&ns->node.list, &prio->node.children);
node             2269 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		base = &fs_prio->node;
node             2275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		base = &fs_ns->node;
node             2305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 					       &fs_ns->node,
node             2339 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_init_node(&ns->node, NULL, NULL);
node             2340 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	tree_add_node(&ns->node, NULL);
node             2418 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	err = init_root_tree(steering, &root_fs, &steering->root_ns->ns.node);
node             2435 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c static void clean_tree(struct fs_node *node)
node             2437 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	if (node) {
node             2441 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_get_node(node);
node             2442 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		list_for_each_entry_safe(iter, temp, &node->children, list)
node             2444 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_put_node(node, false);
node             2445 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		tree_remove_node(node, false);
node             2454 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	clean_tree(&root_ns->ns.node);
node             2543 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			     &steering->rdma_rx_root_ns->ns.node);
node             2722 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 			     &steering->egress_root_ns->ns.node);
node             2909 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	return find_root(&ns->node);
node             3021 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	root = find_root(&ns->node);
node              138 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct fs_node				node;
node              154 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct fs_node			node;
node              196 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct fs_node			node;
node              211 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct fs_node			node;
node              221 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct	fs_node			node;
node              232 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	struct fs_node			node;
node              275 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h #define fs_get_obj(v, _node)  {v = container_of((_node), typeof(*v), node); }
node              278 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	list_for_each_entry(pos, root, node.list)
node              281 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	list_for_each_entry_safe(pos, tmp, root, node.list)
node              284 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	list_for_each_entry_reverse(pos, &(prio)->node.children, list)
node              287 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	list_for_each_entry(pos, (&(prio)->node.children), list)
node              290 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(ns)->node.children)
node              293 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(prio)->node.children)
node              296 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(prio)->node.children)
node              299 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry_safe(pos, tmp, &(prio)->node.children)
node              302 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(ft)->node.children)
node              305 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(fg)->node.children)
node              308 drivers/net/ethernet/mellanox/mlx5/core/fs_core.h 	fs_list_for_each_entry(pos, &(fte)->node.children)
node               75 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	struct qpn_to_netdev *node;
node               77 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	hlist_for_each_entry(node, h, hlist) {
node               78 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 		if (node->underlay_qpn == qpn)
node               79 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 			return node;
node              110 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	struct qpn_to_netdev *node;
node              112 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn);
node              113 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	if (!node) {
node              119 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	hlist_del_init(&node->hlist);
node              121 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	kfree(node);
node              129 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	struct qpn_to_netdev *node;
node              131 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn);
node              132 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	if (!node)
node              135 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c 	return node->netdev;
node               68 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.c 	struct l2addr_node node;
node               57 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h 	hlist_for_each_entry(ptr, &(hash)[ix], node.hlist)  \
node               58 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h 		if (ether_addr_equal(ptr->node.addr, mac)) {\
node               73 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h 		ether_addr_copy(ptr->node.addr, mac);       \
node               74 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h 		hlist_add_head(&ptr->node.hlist, &(hash)[ix]);\
node               80 drivers/net/ethernet/mellanox/mlx5/core/lib/mpfs.h 	hlist_del(&(ptr)->node.hlist);                      \
node              353 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c 		list_for_each_entry(dst, &fte->node.children, node.list) {
node              391 drivers/net/ethernet/mellanox/mlx5/core/steering/fs_dr.c 		list_for_each_entry(dst, &fte->node.children, node.list) {
node               63 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	struct list_head node;
node               76 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	struct list_head node;
node              377 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_del(&mr_route->node);
node              419 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_add_tail(&mr_route->node, &mr_table->route_list);
node              436 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 		list_del(&mr_orig_route->node);
node              447 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_del(&mr_route->node);
node              930 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_add_tail(&mr_table->node, &mr->table_list);
node              946 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_del(&mr_table->node);
node              958 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_for_each_entry_safe(mr_route, tmp, &mr_table->route_list, node)
node             1004 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 	list_for_each_entry(mr_table, &mr->table_list, node)
node             1005 drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c 		list_for_each_entry(mr_route, &mr_table->route_list, node)
node              461 drivers/net/ethernet/moxa/moxart_ether.c 	struct device_node *node = p_dev->of_node;
node              472 drivers/net/ethernet/moxa/moxart_ether.c 	irq = irq_of_parse_and_map(node, 0);
node              741 drivers/net/ethernet/qlogic/qed/qed_fcoe.c 	struct hlist_node node;
node              782 drivers/net/ethernet/qlogic/qed/qed_fcoe.c 	hash_for_each_possible(cdev->connections, hash_con, node, handle) {
node              898 drivers/net/ethernet/qlogic/qed/qed_fcoe.c 	hash_add(cdev->connections, &hash_con->node, *handle);
node              918 drivers/net/ethernet/qlogic/qed/qed_fcoe.c 	hlist_del(&hash_con->node);
node             1078 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	struct hlist_node node;
node             1117 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	hash_for_each_possible(cdev->connections, hash_con, node, handle) {
node             1227 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	hash_add(cdev->connections, &hash_con->node, *handle);
node             1247 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	hlist_del(&hash_con->node);
node               94 drivers/net/ethernet/qlogic/qede/qede_filter.c 	struct hlist_node node;
node              175 drivers/net/ethernet/qlogic/qede/qede_filter.c 	INIT_HLIST_NODE(&fltr->node);
node              176 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_add_head(&fltr->node,
node              194 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_del(&fltr->node);
node              267 drivers/net/ethernet/qlogic/qede/qede_filter.c 		hlist_for_each_entry_safe(fltr, temp, head, node) {
node              415 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_for_each_entry(tpos, h, node)
node             1358 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_for_each_entry(fltr, head, node)
node             1383 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_for_each_entry(fltr, head, node) {
node             1678 drivers/net/ethernet/qlogic/qede/qede_filter.c 	hlist_for_each_entry_safe(fltr, temp, head, node) {
node             1149 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct list_head *node;
node             1155 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		list_for_each(node, &vf->rcv_pend.wait_list) {
node             1156 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 			trans = list_entry(node, struct qlcnic_bc_trans, list);
node              678 drivers/net/ethernet/qualcomm/emac/emac-mac.c 	int node = dev_to_node(adpt->netdev->dev.parent);
node              682 drivers/net/ethernet/qualcomm/emac/emac-mac.c 	tx_q->tpd.tpbuff = kzalloc_node(size, GFP_KERNEL, node);
node              719 drivers/net/ethernet/qualcomm/emac/emac-mac.c 	int node = dev_to_node(adpt->netdev->dev.parent);
node              724 drivers/net/ethernet/qualcomm/emac/emac-mac.c 	rx_q->rfd.rfbuff = kzalloc_node(size, GFP_KERNEL, node);
node             5540 drivers/net/ethernet/realtek/r8169_main.c 	int node = dev_to_node(d);
node             5544 drivers/net/ethernet/realtek/r8169_main.c 	data = alloc_pages_node(node, GFP_KERNEL, get_order(R8169_RX_BUF_SIZE));
node               87 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 	struct device_node *node = dev->of_node;
node              115 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 	priv->irq  = irq_of_parse_and_map(node, 0);
node              127 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 		priv->txq[i]->irq_no = irq_of_parse_and_map(node, chan++);
node              135 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 		priv->rxq[i]->irq_no = irq_of_parse_and_map(node, chan++);
node              142 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 	priv->lpi_irq = irq_of_parse_and_map(node, chan);
node             1228 drivers/net/ethernet/sfc/efx.c 		list_add_tail(&efx->node, &efx_primary_list);
node             1231 drivers/net/ethernet/sfc/efx.c 					 node) {
node             1233 drivers/net/ethernet/sfc/efx.c 				list_del(&other->node);
node             1238 drivers/net/ethernet/sfc/efx.c 				list_add_tail(&other->node,
node             1246 drivers/net/ethernet/sfc/efx.c 		list_for_each_entry(other, &efx_primary_list, node) {
node             1252 drivers/net/ethernet/sfc/efx.c 				list_add_tail(&efx->node,
node             1261 drivers/net/ethernet/sfc/efx.c 		list_add_tail(&efx->node, &efx_unassociated_list);
node             1269 drivers/net/ethernet/sfc/efx.c 	list_del(&efx->node);
node             1272 drivers/net/ethernet/sfc/efx.c 	list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
node             1273 drivers/net/ethernet/sfc/efx.c 		list_del(&other->node);
node             1276 drivers/net/ethernet/sfc/efx.c 		list_add_tail(&other->node, &efx_unassociated_list);
node             3009 drivers/net/ethernet/sfc/efx.c 	INIT_LIST_HEAD(&efx->node);
node             3179 drivers/net/ethernet/sfc/efx.c 	struct hlist_node *node;
node             3184 drivers/net/ethernet/sfc/efx.c 	hlist_for_each(node, head) {
node             3185 drivers/net/ethernet/sfc/efx.c 		rule = container_of(node, struct efx_arfs_rule, node);
node             3198 drivers/net/ethernet/sfc/efx.c 	struct hlist_node *node;
node             3203 drivers/net/ethernet/sfc/efx.c 	hlist_for_each(node, head) {
node             3204 drivers/net/ethernet/sfc/efx.c 		rule = container_of(node, struct efx_arfs_rule, node);
node             3214 drivers/net/ethernet/sfc/efx.c 		hlist_add_head(&rule->node, head);
node             3223 drivers/net/ethernet/sfc/efx.c 	struct hlist_node *node;
node             3228 drivers/net/ethernet/sfc/efx.c 	hlist_for_each(node, head) {
node             3229 drivers/net/ethernet/sfc/efx.c 		rule = container_of(node, struct efx_arfs_rule, node);
node             3239 drivers/net/ethernet/sfc/efx.c 			hlist_del(node);
node             1168 drivers/net/ethernet/sfc/falcon/efx.c 		list_add_tail(&efx->node, &ef4_primary_list);
node             1171 drivers/net/ethernet/sfc/falcon/efx.c 					 node) {
node             1173 drivers/net/ethernet/sfc/falcon/efx.c 				list_del(&other->node);
node             1178 drivers/net/ethernet/sfc/falcon/efx.c 				list_add_tail(&other->node,
node             1186 drivers/net/ethernet/sfc/falcon/efx.c 		list_for_each_entry(other, &ef4_primary_list, node) {
node             1192 drivers/net/ethernet/sfc/falcon/efx.c 				list_add_tail(&efx->node,
node             1201 drivers/net/ethernet/sfc/falcon/efx.c 		list_add_tail(&efx->node, &ef4_unassociated_list);
node             1209 drivers/net/ethernet/sfc/falcon/efx.c 	list_del(&efx->node);
node             1212 drivers/net/ethernet/sfc/falcon/efx.c 	list_for_each_entry_safe(other, next, &efx->secondary_list, node) {
node             1213 drivers/net/ethernet/sfc/falcon/efx.c 		list_del(&other->node);
node             1216 drivers/net/ethernet/sfc/falcon/efx.c 		list_add_tail(&other->node, &ef4_unassociated_list);
node             2641 drivers/net/ethernet/sfc/falcon/efx.c 	INIT_LIST_HEAD(&efx->node);
node               51 drivers/net/ethernet/sfc/falcon/mtd.c 	list_del(&part->node);
node               80 drivers/net/ethernet/sfc/falcon/mtd.c 		list_add_tail(&part->node, &efx->mtd_list);
node              105 drivers/net/ethernet/sfc/falcon/mtd.c 				 node);
node              107 drivers/net/ethernet/sfc/falcon/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
node              119 drivers/net/ethernet/sfc/falcon/mtd.c 	list_for_each_entry(part, &efx->mtd_list, node)
node              767 drivers/net/ethernet/sfc/falcon/net_driver.h 	struct list_head node;
node              915 drivers/net/ethernet/sfc/falcon/net_driver.h 	struct list_head node;
node               51 drivers/net/ethernet/sfc/mtd.c 	list_del(&part->node);
node               83 drivers/net/ethernet/sfc/mtd.c 		list_add_tail(&part->node, &efx->mtd_list);
node              108 drivers/net/ethernet/sfc/mtd.c 				 node);
node              110 drivers/net/ethernet/sfc/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
node              122 drivers/net/ethernet/sfc/mtd.c 	list_for_each_entry(part, &efx->mtd_list, node)
node              756 drivers/net/ethernet/sfc/net_driver.h 	struct hlist_node node;
node              937 drivers/net/ethernet/sfc/net_driver.h 	struct list_head node;
node             1106 drivers/net/ethernet/sfc/net_driver.h 	struct list_head node;
node              873 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c 	struct device_node *node = priv->device->of_node;
node              890 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c 		if (of_property_read_bool(node, "allwinner,leds-active-low"))
node              914 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c 	if (!of_property_read_u32(node, "allwinner,tx-delay-ps", &val)) {
node              932 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c 	if (!of_property_read_u32(node, "allwinner,rx-delay-ps", &val)) {
node             1077 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c static struct regmap *sun8i_dwmac_get_syscon_from_dev(struct device_node *node)
node             1083 drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c 	syscon_node = of_parse_phandle(node, "syscon", 0);
node             1012 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct device_node *node;
node             1015 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	node = priv->plat->phylink_node;
node             1017 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	if (node)
node             1018 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		ret = phylink_of_phy_connect(priv->phylink, node, 0);
node             1023 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	if (!node || ret) {
node              156 drivers/net/ethernet/ti/cpsw-phy-sel.c 	const struct device_node *node = (const struct device_node *)data;
node              157 drivers/net/ethernet/ti/cpsw-phy-sel.c 	return dev->of_node == node &&
node              163 drivers/net/ethernet/ti/cpsw-phy-sel.c 	struct device_node *node;
node              166 drivers/net/ethernet/ti/cpsw-phy-sel.c 	node = of_parse_phandle(dev->of_node, "cpsw-phy-sel", 0);
node              167 drivers/net/ethernet/ti/cpsw-phy-sel.c 	if (!node) {
node              168 drivers/net/ethernet/ti/cpsw-phy-sel.c 		node = of_get_child_by_name(dev->of_node, "cpsw-phy-sel");
node              169 drivers/net/ethernet/ti/cpsw-phy-sel.c 		if (!node) {
node              175 drivers/net/ethernet/ti/cpsw-phy-sel.c 	dev = bus_find_device(&platform_bus_type, NULL, node, match);
node              177 drivers/net/ethernet/ti/cpsw-phy-sel.c 		dev_err(dev, "unable to find platform device for %pOF\n", node);
node              187 drivers/net/ethernet/ti/cpsw-phy-sel.c 	of_node_put(node);
node             2494 drivers/net/ethernet/ti/cpsw.c 	struct device_node *node = pdev->dev.of_node;
node             2499 drivers/net/ethernet/ti/cpsw.c 	if (!node)
node             2502 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "slaves", &prop)) {
node             2508 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "active_slave", &prop)) {
node             2521 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "cpdma_channels", &prop)) {
node             2527 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "ale_entries", &prop)) {
node             2533 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "bd_ram_size", &prop)) {
node             2539 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_u32(node, "mac_control", &prop)) {
node             2545 drivers/net/ethernet/ti/cpsw.c 	if (of_property_read_bool(node, "dual_emac"))
node             2551 drivers/net/ethernet/ti/cpsw.c 	ret = of_platform_populate(node, NULL, NULL, &pdev->dev);
node             2556 drivers/net/ethernet/ti/cpsw.c 	for_each_available_child_of_node(node, slave_node) {
node             2670 drivers/net/ethernet/ti/cpsw.c 	struct device_node *node = pdev->dev.of_node;
node             2674 drivers/net/ethernet/ti/cpsw.c 	for_each_available_child_of_node(node, slave_node) {
node              536 drivers/net/ethernet/ti/cpts.c static int cpts_of_mux_clk_setup(struct cpts *cpts, struct device_node *node)
node              545 drivers/net/ethernet/ti/cpts.c 	refclk_np = of_get_child_by_name(node, "cpts-refclk-mux");
node              612 drivers/net/ethernet/ti/cpts.c static int cpts_of_parse(struct cpts *cpts, struct device_node *node)
node              617 drivers/net/ethernet/ti/cpts.c 	if (!of_property_read_u32(node, "cpts_clock_mult", &prop))
node              620 drivers/net/ethernet/ti/cpts.c 	if (!of_property_read_u32(node, "cpts_clock_shift", &prop))
node              627 drivers/net/ethernet/ti/cpts.c 	return cpts_of_mux_clk_setup(cpts, node);
node              635 drivers/net/ethernet/ti/cpts.c 			 struct device_node *node)
node              648 drivers/net/ethernet/ti/cpts.c 	ret = cpts_of_parse(cpts, node);
node              652 drivers/net/ethernet/ti/cpts.c 	cpts->refclk = devm_get_clk_from_child(dev, node, "cpts");
node              124 drivers/net/ethernet/ti/cpts.h 			 struct device_node *node);
node              149 drivers/net/ethernet/ti/cpts.h 			 struct device_node *node)
node              312 drivers/net/ethernet/ti/davinci_mdio.c 	struct device_node *node = pdev->dev.of_node;
node              315 drivers/net/ethernet/ti/davinci_mdio.c 	if (!node)
node              318 drivers/net/ethernet/ti/davinci_mdio.c 	if (of_property_read_u32(node, "bus_freq", &prop)) {
node               63 drivers/net/ethernet/ti/netcp.h 	struct list_head	node;
node              194 drivers/net/ethernet/ti/netcp.h 			 struct device *device, struct device_node *node,
node              200 drivers/net/ethernet/ti/netcp.h 			  struct device_node *node, void **intf_priv);
node              235 drivers/net/ethernet/ti/netcp_core.c 	struct device_node *devices, *interface, *node = dev->of_node;
node              244 drivers/net/ethernet/ti/netcp_core.c 	devices = of_get_child_by_name(node, "netcp-devices");
node             1388 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry(naddr, &netcp->addr_list, node) {
node             1416 drivers/net/ethernet/ti/netcp_core.c 	list_add_tail(&naddr->node, &netcp->addr_list);
node             1423 drivers/net/ethernet/ti/netcp_core.c 	list_del(&naddr->node);
node             1431 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry(naddr, &netcp->addr_list, node)
node             1458 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry_safe(naddr, tmp, &netcp->addr_list, node) {
node             1482 drivers/net/ethernet/ti/netcp_core.c 	list_for_each_entry_safe(naddr, tmp, &netcp->addr_list, node) {
node             1962 drivers/net/ethernet/ti/netcp_core.c 	struct device_node *node = dev->of_node;
node             2008 drivers/net/ethernet/ti/netcp_core.c 		if (of_address_to_resource(node, NETCP_EFUSE_REG_INDEX, &res)) {
node             2151 drivers/net/ethernet/ti/netcp_core.c 	struct device_node *node = pdev->dev.of_node;
node             2163 drivers/net/ethernet/ti/netcp_core.c 	if (!node) {
node             2188 drivers/net/ethernet/ti/netcp_core.c 	interfaces = of_get_child_by_name(node, "netcp-interfaces");
node              703 drivers/net/ethernet/ti/netcp_ethss.c 	struct device_node		*node;
node             2323 drivers/net/ethernet/ti/netcp_ethss.c 		phy_mode = of_get_phy_mode(slave->node);
node             2998 drivers/net/ethernet/ti/netcp_ethss.c 		      struct device_node *node)
node             3004 drivers/net/ethernet/ti/netcp_ethss.c 	if (of_property_read_u32(node, "slave-port", &slave->slave_num)) {
node             3009 drivers/net/ethernet/ti/netcp_ethss.c 	if (of_property_read_u32(node, "link-interface",
node             3016 drivers/net/ethernet/ti/netcp_ethss.c 	slave->node = node;
node             3021 drivers/net/ethernet/ti/netcp_ethss.c 		slave->phy_node = of_parse_phandle(node, "phy-handle", 0);
node             3122 drivers/net/ethernet/ti/netcp_ethss.c 				 struct device_node *node)
node             3131 drivers/net/ethernet/ti/netcp_ethss.c 	for_each_child_of_node(node, port) {
node             3227 drivers/net/ethernet/ti/netcp_ethss.c 				 struct device_node *node)
node             3233 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, XGBE_SS_REG_INDEX, &res);
node             3237 drivers/net/ethernet/ti/netcp_ethss.c 			node, XGBE_SS_REG_INDEX);
node             3248 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, XGBE_SM_REG_INDEX, &res);
node             3252 drivers/net/ethernet/ti/netcp_ethss.c 			node, XGBE_SM_REG_INDEX);
node             3263 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, XGBE_SERDES_REG_INDEX, &res);
node             3267 drivers/net/ethernet/ti/netcp_ethss.c 			node, XGBE_SERDES_REG_INDEX);
node             3335 drivers/net/ethernet/ti/netcp_ethss.c 				    struct device_node *node)
node             3341 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, GBE_SS_REG_INDEX, &res);
node             3345 drivers/net/ethernet/ti/netcp_ethss.c 			node, GBE_SS_REG_INDEX);
node             3360 drivers/net/ethernet/ti/netcp_ethss.c 				struct device_node *node)
node             3366 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, GBE_SGMII34_REG_INDEX, &res);
node             3370 drivers/net/ethernet/ti/netcp_ethss.c 			node, GBE_SGMII34_REG_INDEX);
node             3382 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, GBE_SM_REG_INDEX, &res);
node             3386 drivers/net/ethernet/ti/netcp_ethss.c 			node, GBE_SM_REG_INDEX);
node             3458 drivers/net/ethernet/ti/netcp_ethss.c 				struct device_node *node)
node             3492 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_address_to_resource(node, GBENU_SM_REG_INDEX, &res);
node             3496 drivers/net/ethernet/ti/netcp_ethss.c 			node, GBENU_SM_REG_INDEX);
node             3555 drivers/net/ethernet/ti/netcp_ethss.c 		     struct device_node *node, void **inst_priv)
node             3564 drivers/net/ethernet/ti/netcp_ethss.c 	if (!node) {
node             3573 drivers/net/ethernet/ti/netcp_ethss.c 	if (of_device_is_compatible(node, "ti,netcp-gbe-5") ||
node             3574 drivers/net/ethernet/ti/netcp_ethss.c 	    of_device_is_compatible(node, "ti,netcp-gbe")) {
node             3576 drivers/net/ethernet/ti/netcp_ethss.c 	} else if (of_device_is_compatible(node, "ti,netcp-gbe-9")) {
node             3578 drivers/net/ethernet/ti/netcp_ethss.c 	} else if (of_device_is_compatible(node, "ti,netcp-gbe-2")) {
node             3581 drivers/net/ethernet/ti/netcp_ethss.c 	} else if (of_device_is_compatible(node, "ti,netcp-xgbe")) {
node             3596 drivers/net/ethernet/ti/netcp_ethss.c 	if (of_find_property(node, "enable-ale", NULL)) {
node             3604 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_property_read_u32(node, "tx-queue",
node             3611 drivers/net/ethernet/ti/netcp_ethss.c 	ret = of_property_read_string(node, "tx-channel",
node             3618 drivers/net/ethernet/ti/netcp_ethss.c 	if (of_node_name_eq(node, "gbe")) {
node             3619 drivers/net/ethernet/ti/netcp_ethss.c 		ret = get_gbe_resource_version(gbe_dev, node);
node             3626 drivers/net/ethernet/ti/netcp_ethss.c 			ret = set_gbe_ethss14_priv(gbe_dev, node);
node             3628 drivers/net/ethernet/ti/netcp_ethss.c 			ret = set_gbenu_ethss_priv(gbe_dev, node);
node             3632 drivers/net/ethernet/ti/netcp_ethss.c 	} else if (of_node_name_eq(node, "xgbe")) {
node             3633 drivers/net/ethernet/ti/netcp_ethss.c 		ret = set_xgbe_ethss10_priv(gbe_dev, node);
node             3639 drivers/net/ethernet/ti/netcp_ethss.c 		dev_err(dev, "unknown GBE node(%pOFn)\n", node);
node             3646 drivers/net/ethernet/ti/netcp_ethss.c 	interfaces = of_get_child_by_name(node, "interfaces");
node             3684 drivers/net/ethernet/ti/netcp_ethss.c 	secondary_ports = of_get_child_by_name(node, "secondary-slave-ports");
node             3716 drivers/net/ethernet/ti/netcp_ethss.c 	cpts_node = of_get_child_by_name(node, "cpts");
node             3718 drivers/net/ethernet/ti/netcp_ethss.c 		cpts_node = of_node_get(node);
node             3751 drivers/net/ethernet/ti/netcp_ethss.c 		      struct device_node *node, void **intf_priv)
node             3757 drivers/net/ethernet/ti/netcp_ethss.c 	if (!node) {
node             3778 drivers/net/ethernet/ti/netcp_ethss.c 	if (init_slave(gbe_dev, gbe_intf->slave, node)) {
node               72 drivers/net/fddi/skfp/h/sba.h 	struct s_sba_node_vars	node[MAX_NODES] ;
node              128 drivers/net/geneve.c 	struct geneve_dev_node *node;
node              134 drivers/net/geneve.c 	hlist_for_each_entry_rcu(node, vni_list_head, hlist) {
node              135 drivers/net/geneve.c 		if (eq_tun_id_and_vni((u8 *)&node->geneve->info.key.tun_id, vni) &&
node              136 drivers/net/geneve.c 		    addr == node->geneve->info.key.u.ipv4.dst)
node              137 drivers/net/geneve.c 			return node->geneve;
node              147 drivers/net/geneve.c 	struct geneve_dev_node *node;
node              153 drivers/net/geneve.c 	hlist_for_each_entry_rcu(node, vni_list_head, hlist) {
node              154 drivers/net/geneve.c 		if (eq_tun_id_and_vni((u8 *)&node->geneve->info.key.tun_id, vni) &&
node              155 drivers/net/geneve.c 		    ipv6_addr_equal(&addr6, &node->geneve->info.key.u.ipv6.dst))
node              156 drivers/net/geneve.c 			return node->geneve;
node              652 drivers/net/geneve.c 	struct geneve_dev_node *node;
node              673 drivers/net/geneve.c 		node = &geneve->hlist6;
node              678 drivers/net/geneve.c 		node = &geneve->hlist4;
node              680 drivers/net/geneve.c 	node->geneve = geneve;
node              684 drivers/net/geneve.c 	hlist_add_head_rcu(&node->hlist, &gs->vni_list[hash]);
node              284 drivers/net/hyperv/netvsc.c 	int node = cpu_to_node(nvchan->channel->target_cpu);
node              288 drivers/net/hyperv/netvsc.c 	nvchan->mrc.slots = vzalloc_node(size, node);
node               41 drivers/net/phy/fixed_phy.c 	struct list_head node;
node               59 drivers/net/phy/fixed_phy.c 	list_for_each_entry(fp, &fmb->phys, node) {
node               80 drivers/net/phy/fixed_phy.c 	list_for_each_entry(fp, &fmb->phys, node) {
node              125 drivers/net/phy/fixed_phy.c 	list_for_each_entry(fp, &fmb->phys, node) {
node              164 drivers/net/phy/fixed_phy.c 	list_add_tail(&fp->node, &fmb->phys);
node              183 drivers/net/phy/fixed_phy.c 	list_for_each_entry_safe(fp, tmp, &fmb->phys, node) {
node              185 drivers/net/phy/fixed_phy.c 			list_del(&fp->node);
node              386 drivers/net/phy/fixed_phy.c 	list_for_each_entry_safe(fp, tmp, &fmb->phys, node) {
node              387 drivers/net/phy/fixed_phy.c 		list_del(&fp->node);
node               25 drivers/net/phy/mdio-thunder.c 	struct device_node *node;
node               66 drivers/net/phy/mdio-thunder.c 		node = to_of_node(fwn);
node               67 drivers/net/phy/mdio-thunder.c 		if (!node)
node               70 drivers/net/phy/mdio-thunder.c 		err = of_address_to_resource(node, 0, &r);
node               74 drivers/net/phy/mdio-thunder.c 				node);
node               99 drivers/net/phy/mdio-thunder.c 		err = of_mdiobus_register(bus->mii_bus, node);
node              246 drivers/net/phy/phy-core.c 	struct device_node *node = phydev->mdio.dev.of_node;
node              252 drivers/net/phy/phy-core.c 	if (!node)
node              255 drivers/net/phy/phy-core.c 	if (!of_property_read_u32(node, "max-speed", &max_speed))
node              261 drivers/net/phy/phy-core.c 	struct device_node *node = phydev->mdio.dev.of_node;
node              267 drivers/net/phy/phy-core.c 	if (!node)
node              270 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-100tx"))
node              272 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-1000t"))
node              274 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-10gt"))
node              276 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-1000kx"))
node              278 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-10gkx4"))
node              280 drivers/net/phy/phy-core.c 	if (of_property_read_bool(node, "eee-broken-10gkr"))
node               18 drivers/net/phy/sfp-bus.c 	struct list_head node;
node              299 drivers/net/phy/sfp-bus.c 	list_for_each_entry(sfp, &sfp_buses, node) {
node              310 drivers/net/phy/sfp-bus.c 		list_add(&new->node, &sfp_buses);
node              326 drivers/net/phy/sfp-bus.c 	list_del(&bus->node);
node             1888 drivers/net/phy/sfp.c 		struct device_node *node = pdev->dev.of_node;
node             1892 drivers/net/phy/sfp.c 		id = of_match_node(sfp_of_match, node);
node             1898 drivers/net/phy/sfp.c 		np = of_parse_phandle(node, "i2c-bus", 0);
node               66 drivers/net/rionet.c 	struct list_head node;
node              240 drivers/net/rionet.c 			list_for_each_entry(peer, &nets[netid].peers, node) {
node              354 drivers/net/rionet.c 	list_for_each_entry(peer, &nets[netid].peers, node) {
node              384 drivers/net/rionet.c 	list_for_each_entry(peer, &nets[netid].peers, node) {
node              414 drivers/net/rionet.c 	list_for_each_entry(peer, &nets[netid].peers, node) {
node              416 drivers/net/rionet.c 			list_del(&peer->node);
node              620 drivers/net/rionet.c 		list_add_tail(&peer->node, &nets[netid].peers);
node              649 drivers/net/rionet.c 		list_for_each_entry(peer, &nets[i].peers, node) {
node              838 drivers/net/team/team.c 	struct list_head *node;
node              843 drivers/net/team/team.c 	node = qom_list;
node              847 drivers/net/team/team.c 		node = &cur->qom_list;
node              849 drivers/net/team/team.c 	list_add_tail_rcu(&port->qom_list, node);
node             1799 drivers/net/usb/lan78xx.c 	struct device_node *node;
node             1829 drivers/net/usb/lan78xx.c 	node = of_get_child_by_name(dev->udev->dev.of_node, "mdio");
node             1830 drivers/net/usb/lan78xx.c 	ret = of_mdiobus_register(dev->mdiobus, node);
node             1831 drivers/net/usb/lan78xx.c 	of_node_put(node);
node             1539 drivers/net/usb/r8152.c 	int node = netdev->dev.parent ? dev_to_node(netdev->dev.parent) : -1;
node             1544 drivers/net/usb/r8152.c 	rx_agg = kmalloc_node(sizeof(*rx_agg), mflags, node);
node             1614 drivers/net/usb/r8152.c 	int node, i;
node             1616 drivers/net/usb/r8152.c 	node = netdev->dev.parent ? dev_to_node(netdev->dev.parent) : -1;
node             1636 drivers/net/usb/r8152.c 		buf = kmalloc_node(agg_buf_sz, GFP_KERNEL, node);
node             1643 drivers/net/usb/r8152.c 					   node);
node              218 drivers/net/virtio_net.c 	struct hlist_node node;
node             1971 drivers/net/virtio_net.c static int virtnet_cpu_online(unsigned int cpu, struct hlist_node *node)
node             1973 drivers/net/virtio_net.c 	struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info,
node             1974 drivers/net/virtio_net.c 						   node);
node             1979 drivers/net/virtio_net.c static int virtnet_cpu_dead(unsigned int cpu, struct hlist_node *node)
node             1981 drivers/net/virtio_net.c 	struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info,
node             1987 drivers/net/virtio_net.c static int virtnet_cpu_down_prep(unsigned int cpu, struct hlist_node *node)
node             1989 drivers/net/virtio_net.c 	struct virtnet_info *vi = hlist_entry_safe(node, struct virtnet_info,
node             1990 drivers/net/virtio_net.c 						   node);
node             2002 drivers/net/virtio_net.c 	ret = cpuhp_state_add_instance_nocalls(virtionet_online, &vi->node);
node             2009 drivers/net/virtio_net.c 	cpuhp_state_remove_instance_nocalls(virtionet_online, &vi->node);
node             2015 drivers/net/virtio_net.c 	cpuhp_state_remove_instance_nocalls(virtionet_online, &vi->node);
node              208 drivers/net/vxlan.c 	struct vxlan_dev_node *node;
node              214 drivers/net/vxlan.c 	hlist_for_each_entry_rcu(node, vni_head(vs, vni), hlist) {
node              215 drivers/net/vxlan.c 		if (node->vxlan->default_dst.remote_vni != vni)
node              219 drivers/net/vxlan.c 			const struct vxlan_config *cfg = &node->vxlan->cfg;
node              226 drivers/net/vxlan.c 		return node->vxlan;
node             2772 drivers/net/vxlan.c 			     struct vxlan_dev_node *node)
node             2777 drivers/net/vxlan.c 	node->vxlan = vxlan;
node             2779 drivers/net/vxlan.c 	hlist_add_head_rcu(&node->hlist, vni_head(vs, vni));
node             3277 drivers/net/vxlan.c 	struct vxlan_dev_node *node;
node             3304 drivers/net/vxlan.c 		node = &vxlan->hlist6;
node             3309 drivers/net/vxlan.c 		node = &vxlan->hlist4;
node             3311 drivers/net/vxlan.c 	vxlan_vs_add_dev(vs, vxlan, node);
node               46 drivers/net/wan/hd64570.c #define SCA_INTR_MSCI(node)    (node ? 0x10 : 0x01)
node               47 drivers/net/wan/hd64570.c #define SCA_INTR_DMAC_RX(node) (node ? 0x20 : 0x02)
node               48 drivers/net/wan/hd64570.c #define SCA_INTR_DMAC_TX(node) (node ? 0x40 : 0x04)
node              131 drivers/net/wan/hd64570.h #define DSR_RX(node) (DSR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET))
node              132 drivers/net/wan/hd64570.h #define DSR_TX(node) (DSR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET))
node              134 drivers/net/wan/hd64570.h #define DMR_RX(node) (DMR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET))
node              135 drivers/net/wan/hd64570.h #define DMR_TX(node) (DMR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET))
node              137 drivers/net/wan/hd64570.h #define FCT_RX(node) (FCT + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET))
node              138 drivers/net/wan/hd64570.h #define FCT_TX(node) (FCT + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET))
node              140 drivers/net/wan/hd64570.h #define DIR_RX(node) (DIR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET))
node              141 drivers/net/wan/hd64570.h #define DIR_TX(node) (DIR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET))
node              143 drivers/net/wan/hd64570.h #define DCR_RX(node) (DCR + (node ? DMAC1RX_OFFSET : DMAC0RX_OFFSET))
node              144 drivers/net/wan/hd64570.h #define DCR_TX(node) (DCR + (node ? DMAC1TX_OFFSET : DMAC0TX_OFFSET))
node               51 drivers/net/wan/lapbether.c 	struct list_head	node;
node               67 drivers/net/wan/lapbether.c 	list_for_each_entry_rcu(lapbeth, &lapbeth_devices, node) {
node              337 drivers/net/wan/lapbether.c 	list_add_rcu(&lapbeth->node, &lapbeth_devices);
node              353 drivers/net/wan/lapbether.c 	list_del_rcu(&lapbeth->node);
node              433 drivers/net/wan/lapbether.c 		lapbeth = list_entry(entry, struct lapbethdev, node);
node               58 drivers/net/wan/wanxl.c         int node;		/* physical port #0 - 3 */
node               94 drivers/net/wan/wanxl.c 	return &port->card->status->port_status[port->node];
node              298 drivers/net/wan/wanxl.c 	writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
node              409 drivers/net/wan/wanxl.c 	writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
node              421 drivers/net/wan/wanxl.c 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
node              435 drivers/net/wan/wanxl.c 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
node              782 drivers/net/wan/wanxl.c 		port->node = i;
node              987 drivers/net/wireless/ath/ath10k/core.c 	struct device_node *node;
node              990 drivers/net/wireless/ath/ath10k/core.c 	node = ar->dev->of_node;
node              991 drivers/net/wireless/ath/ath10k/core.c 	if (!node)
node              994 drivers/net/wireless/ath/ath10k/core.c 	of_property_read_string(node, "qcom,ath10k-calibration-variant",
node             1681 drivers/net/wireless/ath/ath10k/core.c 	struct device_node *node;
node             1686 drivers/net/wireless/ath/ath10k/core.c 	node = ar->dev->of_node;
node             1687 drivers/net/wireless/ath/ath10k/core.c 	if (!node)
node             1693 drivers/net/wireless/ath/ath10k/core.c 	if (!of_get_property(node, dt_name, &data_len)) {
node             1711 drivers/net/wireless/ath/ath10k/core.c 	ret = of_property_read_u8_array(node, dt_name, data, data_len);
node             4730 drivers/net/wireless/ath/ath10k/mac.c 	struct device_node *node;
node             4734 drivers/net/wireless/ath/ath10k/mac.c 	node = ar->dev->of_node;
node             4735 drivers/net/wireless/ath/ath10k/mac.c 	if (!node)
node             4738 drivers/net/wireless/ath/ath10k/mac.c 	ret = of_property_read_string_index(node, "ext-fem-name", 0, &fem_name);
node              885 drivers/net/wireless/ath/ath10k/qmi.c 	sq->sq_node = service->node;
node              959 drivers/net/wireless/ath/ath10k/qmi.c 	struct device_node *node;
node              963 drivers/net/wireless/ath/ath10k/qmi.c 	node = of_parse_phandle(dev->of_node, "memory-region", 0);
node              964 drivers/net/wireless/ath/ath10k/qmi.c 	if (node) {
node              965 drivers/net/wireless/ath/ath10k/qmi.c 		ret = of_address_to_resource(node, 0, &r);
node              970 drivers/net/wireless/ath/ath10k/qmi.c 		of_node_put(node);
node              705 drivers/net/wireless/ath/ath6kl/init.c 	struct device_node *node;
node              710 drivers/net/wireless/ath/ath6kl/init.c 	for_each_compatible_node(node, NULL, "atheros,ath6kl") {
node              711 drivers/net/wireless/ath/ath6kl/init.c 		board_id = of_get_property(node, board_id_prop, NULL);
node              714 drivers/net/wireless/ath/ath6kl/init.c 				    board_id_prop, node);
node              727 drivers/net/wireless/ath/ath6kl/init.c 		of_node_put(node);
node              642 drivers/net/wireless/ath/ath6kl/sdio.c 	struct hif_scatter_req *node = NULL;
node              647 drivers/net/wireless/ath/ath6kl/sdio.c 		node = list_first_entry(&ar_sdio->scat_req,
node              649 drivers/net/wireless/ath/ath6kl/sdio.c 		list_del(&node->list);
node              651 drivers/net/wireless/ath/ath6kl/sdio.c 		node->scat_q_depth = get_queue_depth(&ar_sdio->scat_req);
node              656 drivers/net/wireless/ath/ath6kl/sdio.c 	return node;
node             1051 drivers/net/wireless/ath/ath6kl/txrx.c 	struct skb_hold_q *node;
node             1078 drivers/net/wireless/ath/ath6kl/txrx.c 		node = &rxtid->hold_q[idx];
node             1079 drivers/net/wireless/ath/ath6kl/txrx.c 		if ((order == 1) && (!node->skb))
node             1082 drivers/net/wireless/ath/ath6kl/txrx.c 		if (node->skb) {
node             1083 drivers/net/wireless/ath/ath6kl/txrx.c 			if (node->is_amsdu)
node             1085 drivers/net/wireless/ath/ath6kl/txrx.c 						 node->skb);
node             1087 drivers/net/wireless/ath/ath6kl/txrx.c 				skb_queue_tail(&rxtid->q, node->skb);
node             1088 drivers/net/wireless/ath/ath6kl/txrx.c 			node->skb = NULL;
node             1112 drivers/net/wireless/ath/ath6kl/txrx.c 	struct skb_hold_q *node;
node             1175 drivers/net/wireless/ath/ath6kl/txrx.c 	node = &rxtid->hold_q[idx];
node             1191 drivers/net/wireless/ath/ath6kl/txrx.c 	dev_kfree_skb(node->skb);
node             1194 drivers/net/wireless/ath/ath6kl/txrx.c 	node->skb = frame;
node             1196 drivers/net/wireless/ath/ath6kl/txrx.c 	node->is_amsdu = is_amsdu;
node             1197 drivers/net/wireless/ath/ath6kl/txrx.c 	node->seq_no = seq_no;
node             1199 drivers/net/wireless/ath/ath6kl/txrx.c 	if (node->is_amsdu)
node             2126 drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c 	.node = {},
node               72 drivers/net/wireless/marvell/mwifiex/11n.h 	struct mwifiex_sta_node *node = mwifiex_get_sta_entry(priv, ptr->ra);
node               74 drivers/net/wireless/marvell/mwifiex/11n.h 	if (unlikely(!node))
node               77 drivers/net/wireless/marvell/mwifiex/11n.h 	return (node->ampdu_sta[tid] != BA_STREAM_NOT_ALLOWED) ? true : false;
node              171 drivers/net/wireless/marvell/mwifiex/11n.h 					     struct mwifiex_sta_node *node)
node              173 drivers/net/wireless/marvell/mwifiex/11n.h 	if (!node || ((priv->bss_role == MWIFIEX_BSS_ROLE_UAP) &&
node              179 drivers/net/wireless/marvell/mwifiex/11n.h 	return node->is_11n_enabled;
node              185 drivers/net/wireless/marvell/mwifiex/11n.h 	struct mwifiex_sta_node *node = mwifiex_get_sta_entry(priv, ra);
node              186 drivers/net/wireless/marvell/mwifiex/11n.h 	if (node)
node              187 drivers/net/wireless/marvell/mwifiex/11n.h 		return node->is_11n_enabled;
node              347 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 	struct mwifiex_sta_node *node;
node              373 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 			node = mwifiex_get_sta_entry(priv, ta);
node              374 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 			if (node)
node              375 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 				last_seq = node->rx_seq[tid];
node              378 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 		node = mwifiex_get_sta_entry(priv, ta);
node              379 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 		if (node)
node              380 drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c 			last_seq = node->rx_seq[tid];
node             1382 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			  struct mwifiex_sta_node *node,
node             1393 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		if (!node)
node             1399 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			jiffies_to_msecs(jiffies - node->stats.last_rx);
node             1401 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->signal = node->stats.rssi;
node             1402 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->signal_avg = node->stats.rssi;
node             1403 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->rx_bytes = node->stats.rx_bytes;
node             1404 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->tx_bytes = node->stats.tx_bytes;
node             1405 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->rx_packets = node->stats.rx_packets;
node             1406 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->tx_packets = node->stats.tx_packets;
node             1407 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->tx_failed = node->stats.tx_failed;
node             1410 drivers/net/wireless/marvell/mwifiex/cfg80211.c 				     node->stats.last_tx_htinfo,
node             1412 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		sinfo->txrate.legacy = node->stats.last_tx_rate * 5;
node             1499 drivers/net/wireless/marvell/mwifiex/cfg80211.c 	static struct mwifiex_sta_node *node;
node             1509 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		if (node && (&node->list == &priv->sta_list)) {
node             1510 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			node = NULL;
node             1514 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		node = list_prepare_entry(node, &priv->sta_list, list);
node             1515 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		list_for_each_entry_continue(node, &priv->sta_list, list) {
node             1516 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			ether_addr_copy(mac, node->mac_addr);
node             1517 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			return mwifiex_dump_station_info(priv, node, sinfo);
node             1585 drivers/net/wireless/marvell/mwifiex/main.h 			      struct mwifiex_sta_node *node);
node             1598 drivers/net/wireless/marvell/mwifiex/main.h 			    struct device_node *node, const char *prefix);
node             1607 drivers/net/wireless/marvell/mwifiex/main.h 		       int ies_len, struct mwifiex_sta_node *node);
node             1488 drivers/net/wireless/marvell/mwifiex/sta_cmd.c 			    struct device_node *node, const char *prefix)
node             1496 drivers/net/wireless/marvell/mwifiex/sta_cmd.c 	for_each_property_of_node(node, prop) {
node              878 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 	struct mwifiex_sta_node *node =
node              884 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 			if (!node || reason == TDLS_ERR_LINK_NONEXISTENT)
node              903 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 			if (node && reason != TDLS_ERR_LINK_EXISTS)
node              904 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 				node->tdls_status = TDLS_SETUP_FAILURE;
node              916 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 			if (node)
node              917 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c 				node->tdls_status = TDLS_SETUP_FAILURE;
node              113 drivers/net/wireless/marvell/mwifiex/uap_event.c 	struct mwifiex_sta_node *node;
node              148 drivers/net/wireless/marvell/mwifiex/uap_event.c 		node = mwifiex_add_sta_entry(priv, event->sta_addr);
node              149 drivers/net/wireless/marvell/mwifiex/uap_event.c 		if (!node) {
node              162 drivers/net/wireless/marvell/mwifiex/uap_event.c 				       sinfo->assoc_req_ies_len, node);
node              165 drivers/net/wireless/marvell/mwifiex/uap_event.c 			if (node->is_11n_enabled)
node              166 drivers/net/wireless/marvell/mwifiex/uap_event.c 				node->ampdu_sta[i] =
node              169 drivers/net/wireless/marvell/mwifiex/uap_event.c 				node->ampdu_sta[i] = BA_STREAM_NOT_ALLOWED;
node              171 drivers/net/wireless/marvell/mwifiex/uap_event.c 		memset(node->rx_seq, 0xff, sizeof(node->rx_seq));
node              347 drivers/net/wireless/marvell/mwifiex/uap_event.c 			      struct mwifiex_sta_node *node)
node              349 drivers/net/wireless/marvell/mwifiex/uap_event.c 	if (priv->ap_11n_enabled && node->is_11n_enabled) {
node              350 drivers/net/wireless/marvell/mwifiex/uap_event.c 		mwifiex_11n_del_rx_reorder_tbl_by_ta(priv, node->mac_addr);
node              351 drivers/net/wireless/marvell/mwifiex/uap_event.c 		mwifiex_del_tx_ba_stream_tbl_by_ra(priv, node->mac_addr);
node              353 drivers/net/wireless/marvell/mwifiex/uap_event.c 	mwifiex_del_sta_entry(priv, node->mac_addr);
node              380 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 	struct mwifiex_sta_node *node;
node              396 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 		node = mwifiex_get_sta_entry(priv, ta);
node              397 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 		if (node)
node              398 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 			node->stats.tx_failed++;
node              415 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 		node = mwifiex_get_sta_entry(priv, ta);
node              416 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 		if (node)
node              417 drivers/net/wireless/marvell/mwifiex/uap_txrx.c 			node->rx_seq[uap_rx_pd->priority] =
node              529 drivers/net/wireless/marvell/mwifiex/util.c 	struct mwifiex_sta_node *node;
node              534 drivers/net/wireless/marvell/mwifiex/util.c 	list_for_each_entry(node, &priv->sta_list, list) {
node              535 drivers/net/wireless/marvell/mwifiex/util.c 		if (!memcmp(node->mac_addr, mac, ETH_ALEN))
node              536 drivers/net/wireless/marvell/mwifiex/util.c 			return node;
node              545 drivers/net/wireless/marvell/mwifiex/util.c 	struct mwifiex_sta_node *node;
node              547 drivers/net/wireless/marvell/mwifiex/util.c 	list_for_each_entry(node, &priv->sta_list, list) {
node              548 drivers/net/wireless/marvell/mwifiex/util.c 		if (node->tdls_status == status)
node              549 drivers/net/wireless/marvell/mwifiex/util.c 			return node;
node              609 drivers/net/wireless/marvell/mwifiex/util.c 	struct mwifiex_sta_node *node;
node              615 drivers/net/wireless/marvell/mwifiex/util.c 	node = mwifiex_get_sta_entry(priv, mac);
node              616 drivers/net/wireless/marvell/mwifiex/util.c 	if (node)
node              619 drivers/net/wireless/marvell/mwifiex/util.c 	node = kzalloc(sizeof(*node), GFP_ATOMIC);
node              620 drivers/net/wireless/marvell/mwifiex/util.c 	if (!node)
node              623 drivers/net/wireless/marvell/mwifiex/util.c 	memcpy(node->mac_addr, mac, ETH_ALEN);
node              624 drivers/net/wireless/marvell/mwifiex/util.c 	list_add_tail(&node->list, &priv->sta_list);
node              628 drivers/net/wireless/marvell/mwifiex/util.c 	return node;
node              636 drivers/net/wireless/marvell/mwifiex/util.c 		       int ies_len, struct mwifiex_sta_node *node)
node              648 drivers/net/wireless/marvell/mwifiex/util.c 		node->is_11n_enabled = 1;
node              649 drivers/net/wireless/marvell/mwifiex/util.c 		node->max_amsdu = le16_to_cpu(ht_cap->cap_info) &
node              654 drivers/net/wireless/marvell/mwifiex/util.c 		node->is_11n_enabled = 0;
node              663 drivers/net/wireless/marvell/mwifiex/util.c 	struct mwifiex_sta_node *node;
node              667 drivers/net/wireless/marvell/mwifiex/util.c 	node = mwifiex_get_sta_entry(priv, mac);
node              668 drivers/net/wireless/marvell/mwifiex/util.c 	if (node) {
node              669 drivers/net/wireless/marvell/mwifiex/util.c 		list_del(&node->list);
node              670 drivers/net/wireless/marvell/mwifiex/util.c 		kfree(node);
node              680 drivers/net/wireless/marvell/mwifiex/util.c 	struct mwifiex_sta_node *node, *tmp;
node              684 drivers/net/wireless/marvell/mwifiex/util.c 	list_for_each_entry_safe(node, tmp, &priv->sta_list, list) {
node              685 drivers/net/wireless/marvell/mwifiex/util.c 		list_del(&node->list);
node              686 drivers/net/wireless/marvell/mwifiex/util.c 		kfree(node);
node              140 drivers/net/wireless/marvell/mwifiex/wmm.c 	struct mwifiex_sta_node *node;
node              166 drivers/net/wireless/marvell/mwifiex/wmm.c 			node = mwifiex_get_sta_entry(priv, ra);
node              167 drivers/net/wireless/marvell/mwifiex/wmm.c 			if (node)
node              168 drivers/net/wireless/marvell/mwifiex/wmm.c 				ra_list->tx_paused = node->tx_pause;
node              170 drivers/net/wireless/marvell/mwifiex/wmm.c 				      mwifiex_is_sta_11n_enabled(priv, node);
node              172 drivers/net/wireless/marvell/mwifiex/wmm.c 				ra_list->max_amsdu = node->max_amsdu;
node               19 drivers/net/wireless/quantenna/qtnfmac/util.c 	struct qtnf_sta_node *node;
node               24 drivers/net/wireless/quantenna/qtnfmac/util.c 	list_for_each_entry(node, &list->head, list) {
node               25 drivers/net/wireless/quantenna/qtnfmac/util.c 		if (ether_addr_equal(node->mac_addr, mac))
node               26 drivers/net/wireless/quantenna/qtnfmac/util.c 			return node;
node               35 drivers/net/wireless/quantenna/qtnfmac/util.c 	struct qtnf_sta_node *node;
node               40 drivers/net/wireless/quantenna/qtnfmac/util.c 	list_for_each_entry(node, &list->head, list) {
node               42 drivers/net/wireless/quantenna/qtnfmac/util.c 			return node;
node               52 drivers/net/wireless/quantenna/qtnfmac/util.c 	struct qtnf_sta_node *node;
node               57 drivers/net/wireless/quantenna/qtnfmac/util.c 	node = qtnf_sta_list_lookup(list, mac);
node               59 drivers/net/wireless/quantenna/qtnfmac/util.c 	if (node)
node               62 drivers/net/wireless/quantenna/qtnfmac/util.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node               63 drivers/net/wireless/quantenna/qtnfmac/util.c 	if (unlikely(!node))
node               66 drivers/net/wireless/quantenna/qtnfmac/util.c 	ether_addr_copy(node->mac_addr, mac);
node               67 drivers/net/wireless/quantenna/qtnfmac/util.c 	list_add_tail(&node->list, &list->head);
node               72 drivers/net/wireless/quantenna/qtnfmac/util.c 	return node;
node               78 drivers/net/wireless/quantenna/qtnfmac/util.c 	struct qtnf_sta_node *node;
node               81 drivers/net/wireless/quantenna/qtnfmac/util.c 	node = qtnf_sta_list_lookup(list, mac);
node               83 drivers/net/wireless/quantenna/qtnfmac/util.c 	if (node) {
node               84 drivers/net/wireless/quantenna/qtnfmac/util.c 		list_del(&node->list);
node               86 drivers/net/wireless/quantenna/qtnfmac/util.c 		kfree(node);
node               96 drivers/net/wireless/quantenna/qtnfmac/util.c 	struct qtnf_sta_node *node, *tmp;
node              100 drivers/net/wireless/quantenna/qtnfmac/util.c 	list_for_each_entry_safe(node, tmp, &list->head, list) {
node              101 drivers/net/wireless/quantenna/qtnfmac/util.c 		list_del(&node->list);
node              102 drivers/net/wireless/quantenna/qtnfmac/util.c 		kfree(node);
node              705 drivers/net/xen-netback/xenbus.c 	char *node;
node              708 drivers/net/xen-netback/xenbus.c 	if (vif->credit_watch.node)
node              711 drivers/net/xen-netback/xenbus.c 	node = kmalloc(maxlen, GFP_KERNEL);
node              712 drivers/net/xen-netback/xenbus.c 	if (!node)
node              714 drivers/net/xen-netback/xenbus.c 	snprintf(node, maxlen, "%s/rate", dev->nodename);
node              715 drivers/net/xen-netback/xenbus.c 	vif->credit_watch.node = node;
node              719 drivers/net/xen-netback/xenbus.c 		pr_err("Failed to set watcher %s\n", vif->credit_watch.node);
node              720 drivers/net/xen-netback/xenbus.c 		kfree(node);
node              721 drivers/net/xen-netback/xenbus.c 		vif->credit_watch.node = NULL;
node              729 drivers/net/xen-netback/xenbus.c 	if (vif->credit_watch.node) {
node              731 drivers/net/xen-netback/xenbus.c 		kfree(vif->credit_watch.node);
node              732 drivers/net/xen-netback/xenbus.c 		vif->credit_watch.node = NULL;
node              751 drivers/net/xen-netback/xenbus.c 	char *node;
node              755 drivers/net/xen-netback/xenbus.c 	if (vif->mcast_ctrl_watch.node) {
node              760 drivers/net/xen-netback/xenbus.c 	node = kmalloc(maxlen, GFP_KERNEL);
node              761 drivers/net/xen-netback/xenbus.c 	if (!node) {
node              765 drivers/net/xen-netback/xenbus.c 	snprintf(node, maxlen, "%s/request-multicast-control",
node              767 drivers/net/xen-netback/xenbus.c 	vif->mcast_ctrl_watch.node = node;
node              772 drivers/net/xen-netback/xenbus.c 		       vif->mcast_ctrl_watch.node);
node              773 drivers/net/xen-netback/xenbus.c 		kfree(node);
node              774 drivers/net/xen-netback/xenbus.c 		vif->mcast_ctrl_watch.node = NULL;
node              782 drivers/net/xen-netback/xenbus.c 	if (vif->mcast_ctrl_watch.node) {
node              784 drivers/net/xen-netback/xenbus.c 		kfree(vif->mcast_ctrl_watch.node);
node              785 drivers/net/xen-netback/xenbus.c 		vif->mcast_ctrl_watch.node = NULL;
node              806 drivers/net/xen-netback/xenbus.c 		kfree(be->hotplug_status_watch.node);
node              172 drivers/nfc/nfcmrvl/i2c.c static int nfcmrvl_i2c_parse_dt(struct device_node *node,
node              177 drivers/nfc/nfcmrvl/i2c.c 	ret = nfcmrvl_parse_dt(node, pdata);
node              183 drivers/nfc/nfcmrvl/i2c.c 	if (of_find_property(node, "i2c-int-falling", NULL))
node              188 drivers/nfc/nfcmrvl/i2c.c 	ret = irq_of_parse_and_map(node, 0);
node              262 drivers/nfc/nfcmrvl/main.c int nfcmrvl_parse_dt(struct device_node *node,
node              267 drivers/nfc/nfcmrvl/main.c 	reset_n_io = of_get_named_gpio(node, "reset-n-io", 0);
node              276 drivers/nfc/nfcmrvl/main.c 	if (of_find_property(node, "hci-muxed", NULL))
node              114 drivers/nfc/nfcmrvl/nfcmrvl.h int nfcmrvl_parse_dt(struct device_node *node,
node              120 drivers/nfc/nfcmrvl/spi.c static int nfcmrvl_spi_parse_dt(struct device_node *node,
node              125 drivers/nfc/nfcmrvl/spi.c 	ret = nfcmrvl_parse_dt(node, pdata);
node              131 drivers/nfc/nfcmrvl/spi.c 	ret = irq_of_parse_and_map(node, 0);
node               70 drivers/nfc/nfcmrvl/uart.c static int nfcmrvl_uart_parse_dt(struct device_node *node,
node               76 drivers/nfc/nfcmrvl/uart.c 	matched_node = of_get_compatible_child(node, "marvell,nfc-uart");
node               78 drivers/nfc/nfcmrvl/uart.c 		matched_node = of_get_compatible_child(node, "mrvl,nfc-uart");
node              586 drivers/ntb/hw/amd/ntb_hw_amd.c 	int rc, i, msix_count, node;
node              590 drivers/ntb/hw/amd/ntb_hw_amd.c 	node = dev_to_node(&pdev->dev);
node              596 drivers/ntb/hw/amd/ntb_hw_amd.c 				 GFP_KERNEL, node);
node              601 drivers/ntb/hw/amd/ntb_hw_amd.c 				  GFP_KERNEL, node);
node             1062 drivers/ntb/hw/amd/ntb_hw_amd.c 	int rc, node;
node             1064 drivers/ntb/hw/amd/ntb_hw_amd.c 	node = dev_to_node(&pdev->dev);
node             1066 drivers/ntb/hw/amd/ntb_hw_amd.c 	ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node);
node              367 drivers/ntb/hw/intel/ntb_hw_gen1.c 	int rc, i, msix_count, node;
node              371 drivers/ntb/hw/intel/ntb_hw_gen1.c 	node = dev_to_node(&pdev->dev);
node              382 drivers/ntb/hw/intel/ntb_hw_gen1.c 				 GFP_KERNEL, node);
node              387 drivers/ntb/hw/intel/ntb_hw_gen1.c 				  GFP_KERNEL, node);
node             1858 drivers/ntb/hw/intel/ntb_hw_gen1.c 	int rc, node;
node             1860 drivers/ntb/hw/intel/ntb_hw_gen1.c 	node = dev_to_node(&pdev->dev);
node             1863 drivers/ntb/hw/intel/ntb_hw_gen1.c 		ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node);
node             1880 drivers/ntb/hw/intel/ntb_hw_gen1.c 		ndev = kzalloc_node(sizeof(*ndev), GFP_KERNEL, node);
node              387 drivers/ntb/ntb_transport.c 	int node;
node              396 drivers/ntb/ntb_transport.c 		node = dev_to_node(&nt->ndev->dev);
node              399 drivers/ntb/ntb_transport.c 					  GFP_KERNEL, node);
node              625 drivers/ntb/ntb_transport.c 	int node;
node              657 drivers/ntb/ntb_transport.c 	node = dev_to_node(&ndev->dev);
node              659 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
node             1249 drivers/ntb/ntb_transport.c 	int node;
node             1269 drivers/ntb/ntb_transport.c 	node = dev_to_node(&ndev->dev);
node             1271 drivers/ntb/ntb_transport.c 	nt = kzalloc_node(sizeof(*nt), GFP_KERNEL, node);
node             1305 drivers/ntb/ntb_transport.c 				  GFP_KERNEL, node);
node             1352 drivers/ntb/ntb_transport.c 				  GFP_KERNEL, node);
node             1957 drivers/ntb/ntb_transport.c static bool ntb_dma_filter_fn(struct dma_chan *chan, void *node)
node             1959 drivers/ntb/ntb_transport.c 	return dev_to_node(&chan->dev->device) == (int)(unsigned long)node;
node             1988 drivers/ntb/ntb_transport.c 	int node;
node             1995 drivers/ntb/ntb_transport.c 	node = dev_to_node(&ndev->dev);
node             2020 drivers/ntb/ntb_transport.c 					    (void *)(unsigned long)node);
node             2026 drivers/ntb/ntb_transport.c 					    (void *)(unsigned long)node);
node             2054 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
node             2065 drivers/ntb/ntb_transport.c 		entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node);
node              857 drivers/ntb/test/ntb_perf.c 	int node;
node              859 drivers/ntb/test/ntb_perf.c 	node = dev_to_node(&perf->ntb->dev);
node              861 drivers/ntb/test/ntb_perf.c 	return node == NUMA_NO_NODE || node == dev_to_node(chan->device->dev);
node             3486 drivers/nvme/host/core.c 	int node = ctrl->numa_node, flags = GENHD_FL_EXT_DEVT, ret;
node             3488 drivers/nvme/host/core.c 	ns = kzalloc_node(sizeof(*ns), GFP_KERNEL, node);
node             3530 drivers/nvme/host/core.c 	disk = alloc_disk_node(0, node);
node             3546 drivers/nvme/host/core.c 		ret = nvme_nvm_register(ns, disk_name, node);
node              948 drivers/nvme/host/lightnvm.c int nvme_nvm_register(struct nvme_ns *ns, char *disk_name, int node)
node              956 drivers/nvme/host/lightnvm.c 	dev = nvm_alloc_dev(node);
node              142 drivers/nvme/host/multipath.c 	int node;
node              147 drivers/nvme/host/multipath.c 	for_each_node(node) {
node              148 drivers/nvme/host/multipath.c 		if (ns == rcu_access_pointer(head->current_path[node])) {
node              149 drivers/nvme/host/multipath.c 			rcu_assign_pointer(head->current_path[node], NULL);
node              177 drivers/nvme/host/multipath.c static struct nvme_ns *__nvme_find_path(struct nvme_ns_head *head, int node)
node              187 drivers/nvme/host/multipath.c 			distance = node_distance(node, ns->ctrl->numa_node);
node              212 drivers/nvme/host/multipath.c 		rcu_assign_pointer(head->current_path[node], found);
node              227 drivers/nvme/host/multipath.c 		int node, struct nvme_ns *old)
node              255 drivers/nvme/host/multipath.c 	rcu_assign_pointer(head->current_path[node], found);
node              267 drivers/nvme/host/multipath.c 	int node = numa_node_id();
node              270 drivers/nvme/host/multipath.c 	ns = srcu_dereference(head->current_path[node], &head->srcu);
node              272 drivers/nvme/host/multipath.c 		ns = nvme_round_robin_path(head, node, ns);
node              274 drivers/nvme/host/multipath.c 		ns = __nvme_find_path(head, node);
node              426 drivers/nvme/host/multipath.c 		int node, srcu_idx;
node              429 drivers/nvme/host/multipath.c 		for_each_node(node)
node              430 drivers/nvme/host/multipath.c 			__nvme_find_path(head, node);
node              631 drivers/nvme/host/nvme.h int nvme_nvm_register(struct nvme_ns *ns, char *disk_name, int node);
node              637 drivers/nvme/host/nvme.h 				    int node)
node             2784 drivers/nvme/host/pci.c 	int node, result = -ENOMEM;
node             2789 drivers/nvme/host/pci.c 	node = dev_to_node(&pdev->dev);
node             2790 drivers/nvme/host/pci.c 	if (node == NUMA_NO_NODE)
node             2793 drivers/nvme/host/pci.c 	dev = kzalloc_node(sizeof(*dev), GFP_KERNEL, node);
node             2798 drivers/nvme/host/pci.c 					GFP_KERNEL, node);
node             2830 drivers/nvme/host/pci.c 						GFP_KERNEL, node);
node              443 drivers/nvme/target/tcp.c 	struct llist_node *node;
node              445 drivers/nvme/target/tcp.c 	node = llist_del_all(&queue->resp_list);
node              446 drivers/nvme/target/tcp.c 	if (!node)
node              449 drivers/nvme/target/tcp.c 	while (node) {
node              450 drivers/nvme/target/tcp.c 		struct nvmet_tcp_cmd *cmd = llist_entry(node,
node              454 drivers/nvme/target/tcp.c 		node = node->next;
node               30 drivers/nvmem/core.c 	struct list_head	node;
node              110 drivers/nvmem/core.c 	list_del(&cell->node);
node              121 drivers/nvmem/core.c 	list_for_each_entry_safe(cell, p, &nvmem->cells, node)
node              128 drivers/nvmem/core.c 	list_add_tail(&cell->node, &cell->nvmem->cells);
node              244 drivers/nvmem/core.c 	list_for_each_entry(table, &nvmem_cell_tables, node) {
node              279 drivers/nvmem/core.c 	list_for_each_entry(iter, &nvmem->cells, node) {
node              711 drivers/nvmem/core.c 	list_for_each_entry(lookup, &nvmem_lookup_list, node) {
node              743 drivers/nvmem/core.c 	list_for_each_entry(iter, &nvmem->cells, node) {
node             1281 drivers/nvmem/core.c 	list_add_tail(&table->node, &nvmem_cell_tables);
node             1294 drivers/nvmem/core.c 	list_del(&table->node);
node             1311 drivers/nvmem/core.c 		list_add_tail(&entries[i].node, &nvmem_lookup_list);
node             1329 drivers/nvmem/core.c 		list_del(&entries[i].node);
node               91 drivers/nvmem/snvs_lpgpr.c 	struct device_node *node = dev->of_node;
node               98 drivers/nvmem/snvs_lpgpr.c 	if (!node)
node              109 drivers/nvmem/snvs_lpgpr.c 	syscon_node = of_get_parent(node);
node              236 drivers/of/address.c 			struct device_node *node, const char *name)
node              241 drivers/of/address.c 	parser->node = node;
node              242 drivers/of/address.c 	parser->pna = of_n_addr_cells(node);
node              245 drivers/of/address.c 	parser->range = of_get_property(node, name, &rlen);
node              255 drivers/of/address.c 				struct device_node *node)
node              257 drivers/of/address.c 	return parser_init(parser, node, "ranges");
node              262 drivers/of/address.c 				struct device_node *node)
node              264 drivers/of/address.c 	return parser_init(parser, node, "dma-ranges");
node              282 drivers/of/address.c 	range->cpu_addr = of_translate_address(parser->node,
node              295 drivers/of/address.c 		cpu_addr = of_translate_address(parser->node,
node              923 drivers/of/address.c 	struct device_node *node = of_node_get(np);
node              929 drivers/of/address.c 	if (!node)
node              935 drivers/of/address.c 		naddr = of_n_addr_cells(node);
node              936 drivers/of/address.c 		nsize = of_n_size_cells(node);
node              938 drivers/of/address.c 		parent = __of_get_dma_parent(node);
node              939 drivers/of/address.c 		of_node_put(node);
node              941 drivers/of/address.c 		node = parent;
node              942 drivers/of/address.c 		if (!node)
node              945 drivers/of/address.c 		ranges = of_get_property(node, "dma-ranges", &len);
node              967 drivers/of/address.c 	pna = of_n_addr_cells(node);
node              990 drivers/of/address.c 	of_node_put(node);
node             1006 drivers/of/address.c 	struct device_node *node = of_node_get(np);
node             1011 drivers/of/address.c 	while (node) {
node             1012 drivers/of/address.c 		if (of_property_read_bool(node, "dma-coherent")) {
node             1013 drivers/of/address.c 			of_node_put(node);
node             1016 drivers/of/address.c 		node = of_get_next_parent(node);
node             1018 drivers/of/address.c 	of_node_put(node);
node              686 drivers/of/base.c struct device_node *of_get_parent(const struct device_node *node)
node              691 drivers/of/base.c 	if (!node)
node              695 drivers/of/base.c 	np = of_node_get(node->parent);
node              712 drivers/of/base.c struct device_node *of_get_next_parent(struct device_node *node)
node              717 drivers/of/base.c 	if (!node)
node              721 drivers/of/base.c 	parent = of_node_get(node->parent);
node              722 drivers/of/base.c 	of_node_put(node);
node              728 drivers/of/base.c static struct device_node *__of_get_next_child(const struct device_node *node,
node              733 drivers/of/base.c 	if (!node)
node              736 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
node              756 drivers/of/base.c struct device_node *of_get_next_child(const struct device_node *node,
node              763 drivers/of/base.c 	next = __of_get_next_child(node, prev);
node              777 drivers/of/base.c struct device_node *of_get_next_available_child(const struct device_node *node,
node              783 drivers/of/base.c 	if (!node)
node              787 drivers/of/base.c 	next = prev ? prev->sibling : node->child;
node              812 drivers/of/base.c 	struct device_node *node;
node              815 drivers/of/base.c 		node = of_find_node_by_path("/cpus");
node              820 drivers/of/base.c 	else if (node) {
node              821 drivers/of/base.c 		next = node->child;
node              822 drivers/of/base.c 		of_node_put(node);
node              873 drivers/of/base.c struct device_node *of_get_child_by_name(const struct device_node *node,
node              878 drivers/of/base.c 	for_each_child_of_node(node, child)
node              903 drivers/of/base.c struct device_node *__of_find_node_by_full_path(struct device_node *node,
node              908 drivers/of/base.c 	while (node && *path == '/') {
node              909 drivers/of/base.c 		struct device_node *tmp = node;
node              912 drivers/of/base.c 		node = __of_find_node_by_path(node, path);
node              918 drivers/of/base.c 	return node;
node             1109 drivers/of/base.c 					   const struct device_node *node)
node             1118 drivers/of/base.c 		score = __of_device_is_compatible(node, matches->compatible,
node             1137 drivers/of/base.c 					 const struct device_node *node)
node             1143 drivers/of/base.c 	match = __of_match_node(matches, node);
node             1201 drivers/of/base.c int of_modalias_node(struct device_node *node, char *modalias, int len)
node             1206 drivers/of/base.c 	compatible = of_get_property(node, "compatible", &cplen);
node             1315 drivers/of/base.c 	if (it->node) {
node             1316 drivers/of/base.c 		of_node_put(it->node);
node             1317 drivers/of/base.c 		it->node = NULL;
node             1334 drivers/of/base.c 		it->node = of_find_node_by_phandle(it->phandle);
node             1337 drivers/of/base.c 			if (!it->node) {
node             1343 drivers/of/base.c 			if (of_property_read_u32(it->node, it->cells_name,
node             1356 drivers/of/base.c 					       it->node);
node             1382 drivers/of/base.c 	if (it->node) {
node             1383 drivers/of/base.c 		of_node_put(it->node);
node             1384 drivers/of/base.c 		it->node = NULL;
node             1436 drivers/of/base.c 				out_args->np = it.node;
node             1439 drivers/of/base.c 				of_node_put(it.node);
node             1456 drivers/of/base.c 	of_node_put(it.node);
node               32 drivers/of/dynamic.c struct device_node *of_node_get(struct device_node *node)
node               34 drivers/of/dynamic.c 	if (node)
node               35 drivers/of/dynamic.c 		kobject_get(&node->kobj);
node               36 drivers/of/dynamic.c 	return node;
node               45 drivers/of/dynamic.c void of_node_put(struct device_node *node)
node               47 drivers/of/dynamic.c 	if (node)
node               48 drivers/of/dynamic.c 		kobject_put(&node->kobj);
node              328 drivers/of/dynamic.c 	struct device_node *node = kobj_to_device_node(kobj);
node              331 drivers/of/dynamic.c 	if (!of_node_check_flag(node, OF_DETACHED)) {
node              332 drivers/of/dynamic.c 		pr_err("ERROR: Bad of_node_put() on %pOF\n", node);
node              336 drivers/of/dynamic.c 	if (!of_node_check_flag(node, OF_DYNAMIC))
node              339 drivers/of/dynamic.c 	if (of_node_check_flag(node, OF_OVERLAY)) {
node              341 drivers/of/dynamic.c 		if (!of_node_check_flag(node, OF_OVERLAY_FREE_CSET)) {
node              344 drivers/of/dynamic.c 			       node);
node              353 drivers/of/dynamic.c 		if (node->properties)
node              355 drivers/of/dynamic.c 			       __func__, node);
node              358 drivers/of/dynamic.c 	property_list_free(node->properties);
node              359 drivers/of/dynamic.c 	property_list_free(node->deadprops);
node              361 drivers/of/dynamic.c 	kfree(node->full_name);
node              362 drivers/of/dynamic.c 	kfree(node->data);
node              363 drivers/of/dynamic.c 	kfree(node);
node              423 drivers/of/dynamic.c 	struct device_node *node;
node              425 drivers/of/dynamic.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              426 drivers/of/dynamic.c 	if (!node)
node              428 drivers/of/dynamic.c 	node->full_name = kstrdup(full_name, GFP_KERNEL);
node              429 drivers/of/dynamic.c 	if (!node->full_name) {
node              430 drivers/of/dynamic.c 		kfree(node);
node              434 drivers/of/dynamic.c 	of_node_set_flag(node, OF_DYNAMIC);
node              435 drivers/of/dynamic.c 	of_node_set_flag(node, OF_DETACHED);
node              436 drivers/of/dynamic.c 	of_node_init(node);
node              445 drivers/of/dynamic.c 			if (__of_add_property(node, new_pp)) {
node              453 drivers/of/dynamic.c 	return node;
node              456 drivers/of/dynamic.c 	of_node_put(node); /* Frees the node and properties */
node              473 drivers/of/dynamic.c 	list_del(&ce->node);
node              692 drivers/of/dynamic.c 	list_for_each_entry_safe_reverse(ce, cen, &ocs->entries, node)
node              713 drivers/of/dynamic.c 	list_for_each_entry(ce, &ocs->entries, node) {
node              718 drivers/of/dynamic.c 							     node) {
node              745 drivers/of/dynamic.c 	list_for_each_entry(ce, &ocs->entries, node) {
node              816 drivers/of/dynamic.c 	list_for_each_entry_reverse(ce, &ocs->entries, node) {
node              820 drivers/of/dynamic.c 			list_for_each_entry_continue(ce, &ocs->entries, node) {
node              845 drivers/of/dynamic.c 	list_for_each_entry_reverse(ce, &ocs->entries, node) {
node              926 drivers/of/dynamic.c 	list_add_tail(&ce->node, &ocs->entries);
node               82 drivers/of/fdt.c static bool of_fdt_device_is_available(const void *blob, unsigned long node)
node               84 drivers/of/fdt.c 	const char *status = fdt_getprop(blob, node, "status", NULL);
node              476 drivers/of/fdt.c static int __init __reserved_mem_reserve_reg(unsigned long node,
node              486 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "reg", &len);
node              496 drivers/of/fdt.c 	nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL;
node              512 drivers/of/fdt.c 			fdt_reserved_mem_save_node(node, uname, base, size);
node              524 drivers/of/fdt.c static int __init __reserved_mem_check_root(unsigned long node)
node              528 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "#size-cells", NULL);
node              532 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "#address-cells", NULL);
node              536 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "ranges", NULL);
node              545 drivers/of/fdt.c static int __init __fdt_scan_reserved_mem(unsigned long node, const char *uname,
node              552 drivers/of/fdt.c 		if (__reserved_mem_check_root(node) != 0) {
node              568 drivers/of/fdt.c 	if (!of_fdt_device_is_available(initial_boot_params, node))
node              571 drivers/of/fdt.c 	err = __reserved_mem_reserve_reg(node, uname);
node              572 drivers/of/fdt.c 	if (err == -ENOENT && of_get_flat_dt_prop(node, "size", NULL))
node              573 drivers/of/fdt.c 		fdt_reserved_mem_save_node(node, uname, 0, 0);
node              629 drivers/of/fdt.c int __init of_scan_flat_dt(int (*it)(unsigned long node,
node              661 drivers/of/fdt.c 				    int (*it)(unsigned long node,
node              667 drivers/of/fdt.c 	int node;
node              669 drivers/of/fdt.c 	fdt_for_each_subnode(node, blob, parent) {
node              673 drivers/of/fdt.c 		pathp = fdt_get_name(blob, node, NULL);
node              676 drivers/of/fdt.c 		rc = it(node, pathp, data);
node              691 drivers/of/fdt.c int __init of_get_flat_dt_subnode_by_name(unsigned long node, const char *uname)
node              693 drivers/of/fdt.c 	return fdt_subnode_offset(initial_boot_params, node, uname);
node              710 drivers/of/fdt.c const void *__init of_get_flat_dt_prop(unsigned long node, const char *name,
node              713 drivers/of/fdt.c 	return fdt_getprop(initial_boot_params, node, name, size);
node              727 drivers/of/fdt.c 		      unsigned long node, const char *compat)
node              733 drivers/of/fdt.c 	cp = fdt_getprop(blob, node, "compatible", &cplen);
node              753 drivers/of/fdt.c int __init of_flat_dt_is_compatible(unsigned long node, const char *compat)
node              755 drivers/of/fdt.c 	return of_fdt_is_compatible(initial_boot_params, node, compat);
node              761 drivers/of/fdt.c static int __init of_flat_dt_match(unsigned long node, const char *const *compat)
node              769 drivers/of/fdt.c 		tmp = of_fdt_is_compatible(initial_boot_params, node, *compat);
node              781 drivers/of/fdt.c uint32_t __init of_get_flat_dt_phandle(unsigned long node)
node              783 drivers/of/fdt.c 	return fdt_get_phandle(initial_boot_params, node);
node              791 drivers/of/fdt.c 	int (*iterator)(unsigned long node, const char *uname, int depth, void *data);
node              875 drivers/of/fdt.c static void __init early_init_dt_check_for_initrd(unsigned long node)
node              883 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "linux,initrd-start", &len);
node              888 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "linux,initrd-end", &len);
node              901 drivers/of/fdt.c static inline void early_init_dt_check_for_initrd(unsigned long node)
node              960 drivers/of/fdt.c int __init early_init_dt_scan_root(unsigned long node, const char *uname,
node              971 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "#size-cells", NULL);
node              976 drivers/of/fdt.c 	prop = of_get_flat_dt_prop(node, "#address-cells", NULL);
node              996 drivers/of/fdt.c int __init early_init_dt_scan_memory(unsigned long node, const char *uname,
node              999 drivers/of/fdt.c 	const char *type = of_get_flat_dt_prop(node, "device_type", NULL);
node             1008 drivers/of/fdt.c 	reg = of_get_flat_dt_prop(node, "linux,usable-memory", &l);
node             1010 drivers/of/fdt.c 		reg = of_get_flat_dt_prop(node, "reg", &l);
node             1015 drivers/of/fdt.c 	hotpluggable = of_get_flat_dt_prop(node, "hotpluggable", NULL);
node             1043 drivers/of/fdt.c int __init early_init_dt_scan_chosen(unsigned long node, const char *uname,
node             1056 drivers/of/fdt.c 	early_init_dt_check_for_initrd(node);
node             1059 drivers/of/fdt.c 	p = of_get_flat_dt_prop(node, "bootargs", &l);
node             1083 drivers/of/fdt.c 	rng_seed = of_get_flat_dt_prop(node, "rng-seed", &l);
node             1088 drivers/of/fdt.c 		fdt_nop_property(initial_boot_params, node, "rng-seed");
node              246 drivers/of/fdt_address.c u64 __init of_flat_dt_translate_address(unsigned long node)
node              248 drivers/of/fdt_address.c 	return fdt_translate_address(initial_boot_params, node);
node                8 drivers/of/kobj.c static int of_node_is_initialized(struct device_node *node)
node               10 drivers/of/kobj.c 	return node && node->kobj.state_initialized;
node               14 drivers/of/kobj.c int of_node_is_attached(struct device_node *node)
node               16 drivers/of/kobj.c 	return node && node->kobj.state_in_sysfs;
node               61 drivers/of/of_private.h int of_node_is_attached(struct device_node *node);
node              126 drivers/of/of_private.h struct device_node *__of_find_node_by_full_path(struct device_node *node,
node              155 drivers/of/of_private.h 	list_for_each_entry(_te, &(_oft)->te_list, node)
node              159 drivers/of/of_private.h 	list_for_each_entry_reverse(_te, &(_oft)->te_list, node)
node               51 drivers/of/of_reserved_mem.c void __init fdt_reserved_mem_save_node(unsigned long node, const char *uname,
node               61 drivers/of/of_reserved_mem.c 	rmem->fdt_node = node;
node               74 drivers/of/of_reserved_mem.c static int __init __reserved_mem_alloc_size(unsigned long node,
node               85 drivers/of/of_reserved_mem.c 	prop = of_get_flat_dt_prop(node, "size", &len);
node               95 drivers/of/of_reserved_mem.c 	nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL;
node               97 drivers/of/of_reserved_mem.c 	prop = of_get_flat_dt_prop(node, "alignment", &len);
node              109 drivers/of/of_reserved_mem.c 	    && of_flat_dt_is_compatible(node, "shared-dma-pool")
node              110 drivers/of/of_reserved_mem.c 	    && of_get_flat_dt_prop(node, "reusable", NULL)
node              111 drivers/of/of_reserved_mem.c 	    && !of_get_flat_dt_prop(node, "no-map", NULL)) {
node              118 drivers/of/of_reserved_mem.c 	prop = of_get_flat_dt_prop(node, "alloc-ranges", &len);
node              246 drivers/of/of_reserved_mem.c 		unsigned long node = rmem->fdt_node;
node              252 drivers/of/of_reserved_mem.c 		nomap = of_get_flat_dt_prop(node, "no-map", NULL) != NULL;
node              253 drivers/of/of_reserved_mem.c 		prop = of_get_flat_dt_prop(node, "phandle", &len);
node              255 drivers/of/of_reserved_mem.c 			prop = of_get_flat_dt_prop(node, "linux,phandle", &len);
node              260 drivers/of/of_reserved_mem.c 			err = __reserved_mem_alloc_size(node, rmem->name,
node              275 drivers/of/of_reserved_mem.c static inline struct reserved_mem *__find_rmem(struct device_node *node)
node              279 drivers/of/of_reserved_mem.c 	if (!node->phandle)
node              283 drivers/of/of_reserved_mem.c 		if (reserved_mem[i].phandle == node->phandle)
node              410 drivers/of/overlay.c 		struct target *target, struct device_node *node)
node              418 drivers/of/overlay.c 	node_kbasename = kbasename(node->full_name);
node              430 drivers/of/overlay.c 		tchild->name = __of_get_property(node, "name", NULL);
node              436 drivers/of/overlay.c 		phandle = __of_get_property(node, "phandle", &size);
node              449 drivers/of/overlay.c 		ret = build_changeset_next_level(ovcs, &target_child, node);
node              454 drivers/of/overlay.c 	if (node->phandle && tchild->phandle) {
node              459 drivers/of/overlay.c 		ret = build_changeset_next_level(ovcs, &target_child, node);
node              544 drivers/of/overlay.c 	list_for_each_entry_continue(ce_2, &ovcs->cset.entries, node) {
node              578 drivers/of/overlay.c 	list_for_each_entry_continue(ce_2, &ovcs->cset.entries, node) {
node              616 drivers/of/overlay.c 	list_for_each_entry(ce_1, &ovcs->cset.entries, node) {
node              691 drivers/of/overlay.c 	struct device_node *node;
node              698 drivers/of/overlay.c 		node = of_find_node_by_phandle(val);
node              699 drivers/of/overlay.c 		if (!node)
node              702 drivers/of/overlay.c 		return node;
node              707 drivers/of/overlay.c 		node =  of_find_node_by_path(path);
node              708 drivers/of/overlay.c 		if (!node)
node              711 drivers/of/overlay.c 		return node;
node              735 drivers/of/overlay.c 	struct device_node *node, *overlay_node;
node              767 drivers/of/overlay.c 	for_each_child_of_node(tree, node) {
node              768 drivers/of/overlay.c 		overlay_node = of_get_child_by_name(node, "__overlay__");
node              775 drivers/of/overlay.c 	node = of_get_child_by_name(tree, "__symbols__");
node              776 drivers/of/overlay.c 	if (node) {
node              778 drivers/of/overlay.c 		of_node_put(node);
node              788 drivers/of/overlay.c 	for_each_child_of_node(tree, node) {
node              789 drivers/of/overlay.c 		overlay_node = of_get_child_by_name(node, "__overlay__");
node              795 drivers/of/overlay.c 		fragment->target = find_target(node);
node              809 drivers/of/overlay.c 	node = of_get_child_by_name(tree, "__symbols__");
node              810 drivers/of/overlay.c 	if (node) {
node              813 drivers/of/overlay.c 		fragment->overlay = node;
node             1111 drivers/of/overlay.c 		list_for_each_entry(ce, &ovcs->cset.entries, node) {
node             1144 drivers/of/overlay.c 	list_for_each_entry(remove_ce, &remove_ovcs->cset.entries, node) {
node               65 drivers/of/pdt.c static struct property * __init of_pdt_build_one_prop(phandle node, char *prev,
node               90 drivers/of/pdt.c 		err = of_pdt_prom_ops->nextprop(node, prev, p->name);
node               95 drivers/of/pdt.c 		p->length = of_pdt_prom_ops->getproplen(node, p->name);
node              102 drivers/of/pdt.c 			len = of_pdt_prom_ops->getproperty(node, p->name,
node              112 drivers/of/pdt.c static struct property * __init of_pdt_build_prop_list(phandle node)
node              116 drivers/of/pdt.c 	head = tail = of_pdt_build_one_prop(node, NULL,
node              117 drivers/of/pdt.c 				     ".node", &node, sizeof(node));
node              119 drivers/of/pdt.c 	tail->next = of_pdt_build_one_prop(node, NULL, NULL, NULL, 0);
node              122 drivers/of/pdt.c 		tail->next = of_pdt_build_one_prop(node, tail->name,
node              130 drivers/of/pdt.c static char * __init of_pdt_get_one_property(phandle node, const char *name)
node              135 drivers/of/pdt.c 	len = of_pdt_prom_ops->getproplen(node, name);
node              138 drivers/of/pdt.c 		len = of_pdt_prom_ops->getproperty(node, name, buf, len);
node              144 drivers/of/pdt.c static struct device_node * __init of_pdt_create_node(phandle node,
node              149 drivers/of/pdt.c 	if (!node)
node              157 drivers/of/pdt.c 	dp->name = of_pdt_get_one_property(node, "name");
node              158 drivers/of/pdt.c 	dp->phandle = node;
node              160 drivers/of/pdt.c 	dp->properties = of_pdt_build_prop_list(node);
node              170 drivers/of/pdt.c 						   phandle node)
node              176 drivers/of/pdt.c 		dp = of_pdt_create_node(node, parent);
node              187 drivers/of/pdt.c 		dp->child = of_pdt_build_tree(dp, of_pdt_prom_ops->getchild(node));
node              189 drivers/of/pdt.c 		node = of_pdt_prom_ops->getsibling(node);
node               77 drivers/of/platform.c 	struct device_node *node = dev->of_node;
node               82 drivers/of/platform.c 	while (node->parent) {
node               87 drivers/of/platform.c 		reg = of_get_property(node, "reg", NULL);
node               88 drivers/of/platform.c 		if (reg && (addr = of_translate_address(node, reg)) != OF_BAD_ADDR) {
node               90 drivers/of/platform.c 				     addr, node, dev_name(dev));
node               96 drivers/of/platform.c 			     kbasename(node->full_name), dev_name(dev));
node               97 drivers/of/platform.c 		node = node->parent;
node              219 drivers/of/platform.c static struct amba_device *of_amba_device_create(struct device_node *node,
node              228 drivers/of/platform.c 	pr_debug("Creating amba device %pOF\n", node);
node              230 drivers/of/platform.c 	if (!of_device_is_available(node) ||
node              231 drivers/of/platform.c 	    of_node_test_and_set_flag(node, OF_POPULATED))
node              243 drivers/of/platform.c 	dev->dev.of_node = of_node_get(node);
node              244 drivers/of/platform.c 	dev->dev.fwnode = &node->fwnode;
node              253 drivers/of/platform.c 	prop = of_get_property(node, "arm,primecell-periphid", NULL);
node              259 drivers/of/platform.c 		dev->irq[i] = irq_of_parse_and_map(node, i);
node              261 drivers/of/platform.c 	ret = of_address_to_resource(node, 0, &dev->res);
node              264 drivers/of/platform.c 		       ret, node);
node              271 drivers/of/platform.c 		       ret, node);
node              280 drivers/of/platform.c 	of_node_clear_flag(node, OF_POPULATED);
node              284 drivers/of/platform.c static struct amba_device *of_amba_device_create(struct device_node *node,
node              516 drivers/of/platform.c 	struct device_node *node;
node              526 drivers/of/platform.c 	for_each_matching_node(node, reserved_mem_matches)
node              527 drivers/of/platform.c 		of_platform_device_create(node, NULL, NULL);
node              529 drivers/of/platform.c 	node = of_find_node_by_path("/firmware");
node              530 drivers/of/platform.c 	if (node) {
node              531 drivers/of/platform.c 		of_platform_populate(node, NULL, NULL, NULL);
node              532 drivers/of/platform.c 		of_node_put(node);
node              531 drivers/of/property.c int of_graph_parse_endpoint(const struct device_node *node,
node              534 drivers/of/property.c 	struct device_node *port_node = of_get_parent(node);
node              537 drivers/of/property.c 		  __func__, node);
node              541 drivers/of/property.c 	endpoint->local_node = node;
node              547 drivers/of/property.c 	of_property_read_u32(node, "reg", &endpoint->id);
node              565 drivers/of/property.c 	struct device_node *node, *port;
node              567 drivers/of/property.c 	node = of_get_child_by_name(parent, "ports");
node              568 drivers/of/property.c 	if (node)
node              569 drivers/of/property.c 		parent = node;
node              581 drivers/of/property.c 	of_node_put(node);
node              610 drivers/of/property.c 		struct device_node *node;
node              612 drivers/of/property.c 		node = of_get_child_by_name(parent, "ports");
node              613 drivers/of/property.c 		if (node)
node              614 drivers/of/property.c 			parent = node;
node              617 drivers/of/property.c 		of_node_put(node);
node              668 drivers/of/property.c 	struct device_node *node = NULL;
node              670 drivers/of/property.c 	for_each_endpoint_of_node(parent, node) {
node              671 drivers/of/property.c 		of_graph_parse_endpoint(node, &endpoint);
node              674 drivers/of/property.c 			return node;
node              688 drivers/of/property.c struct device_node *of_graph_get_remote_endpoint(const struct device_node *node)
node              691 drivers/of/property.c 	return of_parse_phandle(node, "remote-endpoint", 0);
node              702 drivers/of/property.c struct device_node *of_graph_get_port_parent(struct device_node *node)
node              706 drivers/of/property.c 	if (!node)
node              713 drivers/of/property.c 	of_node_get(node);
node              716 drivers/of/property.c 	for (depth = 3; depth && node; depth--) {
node              717 drivers/of/property.c 		node = of_get_next_parent(node);
node              718 drivers/of/property.c 		if (depth == 2 && !of_node_name_eq(node, "ports"))
node              721 drivers/of/property.c 	return node;
node              733 drivers/of/property.c 			       const struct device_node *node)
node              738 drivers/of/property.c 	np = of_graph_get_remote_endpoint(node);
node              755 drivers/of/property.c struct device_node *of_graph_get_remote_port(const struct device_node *node)
node              760 drivers/of/property.c 	np = of_graph_get_remote_endpoint(node);
node              788 drivers/of/property.c struct device_node *of_graph_get_remote_node(const struct device_node *node,
node              793 drivers/of/property.c 	endpoint_node = of_graph_get_endpoint_by_regs(node, port, endpoint);
node              796 drivers/of/property.c 			 port, endpoint, node);
node              843 drivers/of/property.c 	const struct device_node *node = to_of_node(fwnode);
node              846 drivers/of/property.c 		return of_property_count_elems_of_size(node, propname,
node              851 drivers/of/property.c 		return of_property_read_u8_array(node, propname, val, nval);
node              853 drivers/of/property.c 		return of_property_read_u16_array(node, propname, val, nval);
node              855 drivers/of/property.c 		return of_property_read_u32_array(node, propname, val, nval);
node              857 drivers/of/property.c 		return of_property_read_u64_array(node, propname, val, nval);
node              868 drivers/of/property.c 	const struct device_node *node = to_of_node(fwnode);
node              871 drivers/of/property.c 		of_property_read_string_array(node, propname, val, nval) :
node              872 drivers/of/property.c 		of_property_count_strings(node, propname);
node              893 drivers/of/property.c 	const struct device_node *node = to_of_node(fwnode);
node              896 drivers/of/property.c 	for_each_available_child_of_node(node, child)
node              968 drivers/of/property.c 	const struct device_node *node = to_of_node(fwnode);
node              969 drivers/of/property.c 	struct device_node *port_node = of_get_parent(node);
node              974 drivers/of/property.c 	of_property_read_u32(node, "reg", &endpoint->id);
node               24 drivers/of/resolver.c 	struct device_node *node;
node               30 drivers/of/resolver.c 	for_each_of_allnodes(node) {
node               31 drivers/of/resolver.c 		if (node->phandle != OF_PHANDLE_ILLEGAL &&
node               32 drivers/of/resolver.c 				node->phandle > phandle)
node               33 drivers/of/resolver.c 			phandle = node->phandle;
node              321 drivers/of/unittest.c 	struct hlist_node node;
node              337 drivers/of/unittest.c 		hash_for_each_possible(phandle_ht, nh, node, np->phandle) {
node              351 drivers/of/unittest.c 		hash_add(phandle_ht, &nh->node, np->phandle);
node              358 drivers/of/unittest.c 	hash_for_each_safe(phandle_ht, i, tmp, nh, node) {
node              359 drivers/of/unittest.c 		hash_del(&nh->node);
node               38 drivers/opp/core.c 	list_for_each_entry(opp_dev, &opp_table->dev_list, node)
node               50 drivers/opp/core.c 	list_for_each_entry(opp_table, &opp_tables, node) {
node              230 drivers/opp/core.c 		list_for_each_entry(opp, &opp_table->opp_list, node) {
node              309 drivers/opp/core.c 	list_for_each_entry(opp, &opp_table->opp_list, node) {
node              386 drivers/opp/core.c 	list_for_each_entry(temp_opp, &opp_table->opp_list, node) {
node              435 drivers/opp/core.c 	list_for_each_entry(temp_opp, &opp_table->opp_list, node) {
node              459 drivers/opp/core.c 	list_for_each_entry(temp_opp, &opp_table->opp_list, node) {
node              551 drivers/opp/core.c 	list_for_each_entry(temp_opp, &opp_table->opp_list, node) {
node              609 drivers/opp/core.c 	list_for_each_entry(temp_opp, &opp_table->opp_list, node) {
node              914 drivers/opp/core.c 	list_del(&opp_dev->node);
node              930 drivers/opp/core.c 	list_add(&opp_dev->node, &opp_table->dev_list);
node              994 drivers/opp/core.c 	list_add(&opp_table->node, &opp_tables);
node             1056 drivers/opp/core.c 	list_for_each_entry_safe(opp_dev, temp, &opp_table->dev_list, node) {
node             1069 drivers/opp/core.c 	list_del(&opp_table->node);
node             1079 drivers/opp/core.c 	list_for_each_entry_safe(opp, tmp, &opp_table->opp_list, node) {
node             1124 drivers/opp/core.c 	list_del(&opp->node);
node             1181 drivers/opp/core.c 	list_for_each_entry(opp, &opp_table->opp_list, node) {
node             1222 drivers/opp/core.c 	list_for_each_entry_safe(opp, temp, &opp_table->opp_list, node) {
node             1255 drivers/opp/core.c 	INIT_LIST_HEAD(&opp->node);
node             1299 drivers/opp/core.c 	list_for_each_entry(opp, &opp_table->opp_list, node) {
node             1301 drivers/opp/core.c 			*head = &opp->node;
node             1349 drivers/opp/core.c 	list_add(&new_opp->node, head);
node             1982 drivers/opp/core.c 	list_for_each_entry(opp, &src_table->opp_list, node) {
node             2066 drivers/opp/core.c 	list_for_each_entry(tmp_opp, &opp_table->opp_list, node) {
node              224 drivers/opp/cpu.c 		list_for_each_entry(opp_dev, &opp_table->dev_list, node)
node              153 drivers/opp/debugfs.c 	list_for_each_entry(new_dev, &opp_table->dev_list, node)
node               51 drivers/opp/of.c 	list_for_each_entry(opp_table, &opp_tables, node) {
node               82 drivers/opp/of.c 	list_for_each_entry(opp, &opp_table->opp_list, node) {
node              124 drivers/opp/of.c 	list_for_each_entry(opp_table, &opp_tables, node) {
node              693 drivers/opp/of.c 	list_for_each_entry(opp, &opp_table->opp_list, node)
node               72 drivers/opp/opp.h 	struct list_head node;
node              107 drivers/opp/opp.h 	struct list_head node;
node              164 drivers/opp/opp.h 	struct list_head node;
node              698 drivers/parisc/lba_pci.c 	list_for_each_entry(child, &bus->children, node)
node              241 drivers/pci/controller/dwc/pci-dra7xx.c 	struct device_node *node = dev->of_node;
node              242 drivers/pci/controller/dwc/pci-dra7xx.c 	struct device_node *pcie_intc_node =  of_get_next_child(node, NULL);
node             1018 drivers/pci/controller/dwc/pci-imx6.c 	struct device_node *node = dev->of_node;
node             1037 drivers/pci/controller/dwc/pci-imx6.c 	np = of_parse_phandle(node, "fsl,imx7d-pcie-phy", 0);
node             1059 drivers/pci/controller/dwc/pci-imx6.c 	imx6_pcie->reset_gpio = of_get_named_gpio(node, "reset-gpio", 0);
node             1060 drivers/pci/controller/dwc/pci-imx6.c 	imx6_pcie->gpio_active_high = of_property_read_bool(node,
node             1149 drivers/pci/controller/dwc/pci-imx6.c 	if (of_property_read_u32(node, "fsl,tx-deemph-gen1",
node             1153 drivers/pci/controller/dwc/pci-imx6.c 	if (of_property_read_u32(node, "fsl,tx-deemph-gen2-3p5db",
node             1157 drivers/pci/controller/dwc/pci-imx6.c 	if (of_property_read_u32(node, "fsl,tx-deemph-gen2-6db",
node             1161 drivers/pci/controller/dwc/pci-imx6.c 	if (of_property_read_u32(node, "fsl,tx-swing-full",
node             1165 drivers/pci/controller/dwc/pci-imx6.c 	if (of_property_read_u32(node, "fsl,tx-swing-low",
node             1170 drivers/pci/controller/dwc/pci-imx6.c 	ret = of_property_read_u32(node, "fsl,max-link-speed",
node              115 drivers/pci/controller/dwc/pcie-armada8k.c 	struct device_node *node = dev->of_node;
node              120 drivers/pci/controller/dwc/pcie-armada8k.c 		pcie->phy[i] = devm_of_phy_get_by_index(dev, node, i);
node              505 drivers/pci/controller/dwc/pcie-designware-host.c 	list_for_each_entry(child, &pp->root_bus->children, node)
node             1037 drivers/pci/controller/dwc/pcie-tegra194.c 	list_for_each_entry(child, &pp->root_bus->children, node) {
node              772 drivers/pci/controller/pci-aardvark.c 	struct device_node *node = dev->of_node;
node              807 drivers/pci/controller/pci-aardvark.c 		pci_msi_create_irq_domain(of_node_to_fwnode(node),
node              826 drivers/pci/controller/pci-aardvark.c 	struct device_node *node = dev->of_node;
node              831 drivers/pci/controller/pci-aardvark.c 	pcie_intc_node =  of_get_next_child(node, NULL);
node             1043 drivers/pci/controller/pci-mvebu.c 		list_for_each_entry(child, &bus->children, node)
node             2825 drivers/pci/controller/pci-tegra.c 	list_for_each_entry(child, &host->bus->children, node)
node              357 drivers/pci/controller/pci-thunder-pem.c 	int node = acpi_get_node(root->device->handle);
node              360 drivers/pci/controller/pci-thunder-pem.c 	if (node == NUMA_NO_NODE)
node              361 drivers/pci/controller/pci-thunder-pem.c 		node = 0;
node              364 drivers/pci/controller/pci-thunder-pem.c 	index -= node * PEM_MAX_DOM_IN_NODE;
node              365 drivers/pci/controller/pci-thunder-pem.c 	res_pem->start = PEM_RES_BASE | FIELD_PREP(PEM_NODE_MASK, node) |
node              215 drivers/pci/controller/pci-versatile.c 	list_for_each_entry(child, &bus->children, node)
node               33 drivers/pci/controller/pci-xgene-msi.c 	struct device_node	*node;
node              250 drivers/pci/controller/pci-xgene-msi.c 	msi->msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(msi->node),
node              463 drivers/pci/controller/pci-xgene-msi.c 	xgene_msi->node = pdev->dev.of_node;
node               65 drivers/pci/controller/pci-xgene.c 	struct device_node	*node;
node              220 drivers/pci/controller/pci-xgene.c 	entry = list_first_entry(&list, struct resource_entry, node);
node              539 drivers/pci/controller/pci-xgene.c 	struct device_node *np = port->node;
node              622 drivers/pci/controller/pci-xgene.c 	port->node = of_node_get(dn);
node              626 drivers/pci/controller/pci-xgene.c 	if (of_device_is_compatible(port->node, "apm,xgene-pcie"))
node              665 drivers/pci/controller/pci-xgene.c 	list_for_each_entry(child, &bus->children, node)
node              709 drivers/pci/controller/pcie-altera.c 	struct device_node *node = dev->of_node;
node              712 drivers/pci/controller/pcie-altera.c 	pcie->irq_domain = irq_domain_add_linear(node, PCI_NUM_INTX,
node              873 drivers/pci/controller/pcie-altera.c 	list_for_each_entry(child, &bus->children, node)
node              438 drivers/pci/controller/pcie-iproc-msi.c static int iproc_msi_alloc_domains(struct device_node *node,
node              446 drivers/pci/controller/pcie-iproc-msi.c 	msi->msi_domain = pci_msi_create_irq_domain(of_node_to_fwnode(node),
node              511 drivers/pci/controller/pcie-iproc-msi.c int iproc_msi_init(struct iproc_pcie *pcie, struct device_node *node)
node              517 drivers/pci/controller/pcie-iproc-msi.c 	if (!of_device_is_compatible(node, "brcm,iproc-msi"))
node              520 drivers/pci/controller/pcie-iproc-msi.c 	if (!of_find_property(node, "msi-controller", NULL))
node              536 drivers/pci/controller/pcie-iproc-msi.c 	msi->nr_irqs = of_irq_count(node);
node              577 drivers/pci/controller/pcie-iproc-msi.c 	if (of_find_property(node, "brcm,pcie-msi-inten", NULL))
node              592 drivers/pci/controller/pcie-iproc-msi.c 		unsigned int irq = irq_of_parse_and_map(node, i);
node              613 drivers/pci/controller/pcie-iproc-msi.c 	ret = iproc_msi_alloc_domains(node, msi);
node             1199 drivers/pci/controller/pcie-iproc.c 			head = &tmp->node;
node             1564 drivers/pci/controller/pcie-iproc.c 	list_for_each_entry(child, &host->bus->children, node)
node              114 drivers/pci/controller/pcie-iproc.h int iproc_msi_init(struct iproc_pcie *pcie, struct device_node *node);
node              118 drivers/pci/controller/pcie-iproc.h 				 struct device_node *node)
node              157 drivers/pci/controller/pcie-mediatek.c 	int (*setup_irq)(struct mtk_pcie_port *port, struct device_node *node);
node              571 drivers/pci/controller/pcie-mediatek.c 				    struct device_node *node)
node              578 drivers/pci/controller/pcie-mediatek.c 	pcie_intc_node = of_get_next_child(node, NULL);
node              641 drivers/pci/controller/pcie-mediatek.c 			      struct device_node *node)
node              648 drivers/pci/controller/pcie-mediatek.c 	err = mtk_pcie_init_irq_domain(port, node);
node              894 drivers/pci/controller/pcie-mediatek.c 			       struct device_node *node,
node              964 drivers/pci/controller/pcie-mediatek.c 		err = pcie->soc->setup_irq(port, node);
node             1022 drivers/pci/controller/pcie-mediatek.c 	struct device_node *node = dev->of_node, *child;
node             1060 drivers/pci/controller/pcie-mediatek.c 	for_each_available_child_of_node(node, child) {
node              427 drivers/pci/controller/pcie-mobiveil.c 	struct device_node *node = dev->of_node;
node              453 drivers/pci/controller/pcie-mobiveil.c 	if (of_property_read_u32(node, "apio-wins", &pcie->apio_wins))
node              456 drivers/pci/controller/pcie-mobiveil.c 	if (of_property_read_u32(node, "ppio-wins", &pcie->ppio_wins))
node              836 drivers/pci/controller/pcie-mobiveil.c 	struct device_node *node = dev->of_node;
node              840 drivers/pci/controller/pcie-mobiveil.c 	pcie->intx_domain = irq_domain_add_linear(node, PCI_NUM_INTX,
node              939 drivers/pci/controller/pcie-mobiveil.c 	list_for_each_entry(child, &bus->children, node)
node              489 drivers/pci/controller/pcie-rcar.c 	list_for_each_entry(child, &bus->children, node)
node             1065 drivers/pci/controller/pcie-rockchip-host.c 	list_for_each_entry(child, &bus->children, node)
node               29 drivers/pci/controller/pcie-rockchip.c 	struct device_node *node = dev->of_node;
node               59 drivers/pci/controller/pcie-rockchip.c 	err = of_property_read_u32(node, "num-lanes", &rockchip->lanes);
node               67 drivers/pci/controller/pcie-rockchip.c 	rockchip->link_gen = of_pci_get_max_link_speed(node);
node              547 drivers/pci/controller/pcie-xilinx-nwl.c 	struct device_node *node = dev->of_node;
node              550 drivers/pci/controller/pcie-xilinx-nwl.c 	legacy_intc_node = of_get_next_child(node, NULL);
node              888 drivers/pci/controller/pcie-xilinx-nwl.c 	list_for_each_entry(child, &bus->children, node)
node              504 drivers/pci/controller/pcie-xilinx.c 	struct device_node *node = dev->of_node;
node              509 drivers/pci/controller/pcie-xilinx.c 	pcie_intc_node = of_get_next_child(node, NULL);
node              526 drivers/pci/controller/pcie-xilinx.c 		port->msi_domain = irq_domain_add_linear(node,
node              583 drivers/pci/controller/pcie-xilinx.c 	struct device_node *node = dev->of_node;
node              587 drivers/pci/controller/pcie-xilinx.c 	err = of_address_to_resource(node, 0, &regs);
node              597 drivers/pci/controller/pcie-xilinx.c 	port->irq = irq_of_parse_and_map(node, 0);
node              681 drivers/pci/controller/pcie-xilinx.c 	list_for_each_entry(child, &bus->children, node)
node               68 drivers/pci/controller/vmd.c 	struct list_head	node;
node              147 drivers/pci/controller/vmd.c 	list_add_tail_rcu(&vmdirq->node, &vmdirq->irq->irq_list);
node              163 drivers/pci/controller/vmd.c 		list_del_rcu(&vmdirq->node);
node              238 drivers/pci/controller/vmd.c 	INIT_LIST_HEAD(&vmdirq->node);
node              673 drivers/pci/controller/vmd.c 	sd->node = pcibus_to_node(vmd->dev->bus);
node              709 drivers/pci/controller/vmd.c 	list_for_each_entry(child, &vmd->bus->children, node)
node              726 drivers/pci/controller/vmd.c 	list_for_each_entry_rcu(vmdirq, &irqs->irq_list, node)
node               81 drivers/pci/hotplug/acpiphp.h 	struct list_head node;
node              152 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry_safe(slot, next, &bridge->slots, node) {
node              269 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry(slot, &bridge->slots, node)
node              285 drivers/pci/hotplug/acpiphp_glue.c 	list_add_tail(&slot->node, &bridge->slots);
node              335 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry(slot, &bridge->slots, node) {
node              377 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry(tmp, &bus->children, node) {
node              697 drivers/pci/hotplug/acpiphp_glue.c 	list_for_each_entry(slot, &bridge->slots, node) {
node              462 drivers/pci/hotplug/cpqphp.h 				   struct pci_resource *node)
node              464 drivers/pci/hotplug/cpqphp.h 	if (!node || !head)
node              466 drivers/pci/hotplug/cpqphp.h 	node->next = *head;
node              467 drivers/pci/hotplug/cpqphp.h 	*head = node;
node              400 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_resource *node;
node              425 drivers/pci/hotplug/cpqphp_ctrl.c 	node = *head;
node              427 drivers/pci/hotplug/cpqphp_ctrl.c 	if (node->length & (alignment - 1)) {
node              436 drivers/pci/hotplug/cpqphp_ctrl.c 		temp_dword = (node->length | (alignment-1)) + 1 - alignment;
node              438 drivers/pci/hotplug/cpqphp_ctrl.c 		split_node->base = node->base;
node              441 drivers/pci/hotplug/cpqphp_ctrl.c 		node->length -= temp_dword;
node              442 drivers/pci/hotplug/cpqphp_ctrl.c 		node->base += split_node->length;
node              446 drivers/pci/hotplug/cpqphp_ctrl.c 		split_node->next = node;
node              449 drivers/pci/hotplug/cpqphp_ctrl.c 	if (node->length < alignment)
node              453 drivers/pci/hotplug/cpqphp_ctrl.c 	if (*head == node) {
node              454 drivers/pci/hotplug/cpqphp_ctrl.c 		*head = node->next;
node              457 drivers/pci/hotplug/cpqphp_ctrl.c 		while (prevnode->next != node)
node              460 drivers/pci/hotplug/cpqphp_ctrl.c 		prevnode->next = node->next;
node              462 drivers/pci/hotplug/cpqphp_ctrl.c 	node->next = NULL;
node              464 drivers/pci/hotplug/cpqphp_ctrl.c 	return node;
node              476 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_resource *node;
node              485 drivers/pci/hotplug/cpqphp_ctrl.c 	node = *head;
node              487 drivers/pci/hotplug/cpqphp_ctrl.c 	while (node->next) {
node              488 drivers/pci/hotplug/cpqphp_ctrl.c 		prevnode = node;
node              489 drivers/pci/hotplug/cpqphp_ctrl.c 		node = node->next;
node              493 drivers/pci/hotplug/cpqphp_ctrl.c 	if (node->length < alignment)
node              496 drivers/pci/hotplug/cpqphp_ctrl.c 	if (node->base & (alignment - 1)) {
node              498 drivers/pci/hotplug/cpqphp_ctrl.c 		temp_dword = (node->base | (alignment-1)) + 1;
node              499 drivers/pci/hotplug/cpqphp_ctrl.c 		if ((node->length - (temp_dword - node->base)) < alignment)
node              502 drivers/pci/hotplug/cpqphp_ctrl.c 		node->length -= (temp_dword - node->base);
node              503 drivers/pci/hotplug/cpqphp_ctrl.c 		node->base = temp_dword;
node              506 drivers/pci/hotplug/cpqphp_ctrl.c 	if (node->length & (alignment - 1))
node              510 drivers/pci/hotplug/cpqphp_ctrl.c 	return node;
node              512 drivers/pci/hotplug/cpqphp_ctrl.c 	kfree(node);
node              529 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_resource *node;
node              542 drivers/pci/hotplug/cpqphp_ctrl.c 	for (node = *head; node; node = node->next) {
node              543 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->length < size)
node              546 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->base & (size - 1)) {
node              550 drivers/pci/hotplug/cpqphp_ctrl.c 			temp_dword = (node->base | (size-1)) + 1;
node              553 drivers/pci/hotplug/cpqphp_ctrl.c 			if ((node->length - (temp_dword - node->base)) < size)
node              561 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->base = node->base;
node              562 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->length = temp_dword - node->base;
node              563 drivers/pci/hotplug/cpqphp_ctrl.c 			node->base = temp_dword;
node              564 drivers/pci/hotplug/cpqphp_ctrl.c 			node->length -= split_node->length;
node              567 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
node              568 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
node              572 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->length > size) {
node              581 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->base = node->base + size;
node              582 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->length = node->length - size;
node              583 drivers/pci/hotplug/cpqphp_ctrl.c 			node->length = size;
node              586 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
node              587 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
node              591 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->base & 0x300L)
node              597 drivers/pci/hotplug/cpqphp_ctrl.c 		if (*head == node) {
node              598 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = node->next;
node              601 drivers/pci/hotplug/cpqphp_ctrl.c 			while (prevnode->next != node)
node              604 drivers/pci/hotplug/cpqphp_ctrl.c 			prevnode->next = node->next;
node              606 drivers/pci/hotplug/cpqphp_ctrl.c 		node->next = NULL;
node              610 drivers/pci/hotplug/cpqphp_ctrl.c 	return node;
node              723 drivers/pci/hotplug/cpqphp_ctrl.c 	struct pci_resource *node;
node              733 drivers/pci/hotplug/cpqphp_ctrl.c 	for (node = *head; node; node = node->next) {
node              735 drivers/pci/hotplug/cpqphp_ctrl.c 		    __func__, size, node, node->base, node->length);
node              736 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->length < size)
node              739 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->base & (size - 1)) {
node              744 drivers/pci/hotplug/cpqphp_ctrl.c 			temp_dword = (node->base | (size-1)) + 1;
node              747 drivers/pci/hotplug/cpqphp_ctrl.c 			if ((node->length - (temp_dword - node->base)) < size)
node              755 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->base = node->base;
node              756 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->length = temp_dword - node->base;
node              757 drivers/pci/hotplug/cpqphp_ctrl.c 			node->base = temp_dword;
node              758 drivers/pci/hotplug/cpqphp_ctrl.c 			node->length -= split_node->length;
node              760 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
node              761 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
node              765 drivers/pci/hotplug/cpqphp_ctrl.c 		if (node->length > size) {
node              775 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->base = node->base + size;
node              776 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->length = node->length - size;
node              777 drivers/pci/hotplug/cpqphp_ctrl.c 			node->length = size;
node              780 drivers/pci/hotplug/cpqphp_ctrl.c 			split_node->next = node->next;
node              781 drivers/pci/hotplug/cpqphp_ctrl.c 			node->next = split_node;
node              787 drivers/pci/hotplug/cpqphp_ctrl.c 		if (*head == node) {
node              788 drivers/pci/hotplug/cpqphp_ctrl.c 			*head = node->next;
node              791 drivers/pci/hotplug/cpqphp_ctrl.c 			while (prevnode->next != node)
node              794 drivers/pci/hotplug/cpqphp_ctrl.c 			prevnode->next = node->next;
node              796 drivers/pci/hotplug/cpqphp_ctrl.c 		node->next = NULL;
node              799 drivers/pci/hotplug/cpqphp_ctrl.c 	return node;
node             1421 drivers/pci/hotplug/cpqphp_pci.c 	struct pci_resource *node;
node             1428 drivers/pci/hotplug/cpqphp_pci.c 	node = func->io_head;
node             1430 drivers/pci/hotplug/cpqphp_pci.c 	while (node) {
node             1431 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
node             1432 drivers/pci/hotplug/cpqphp_pci.c 		return_resource(&(resources->io_head), node);
node             1433 drivers/pci/hotplug/cpqphp_pci.c 		node = t_node;
node             1436 drivers/pci/hotplug/cpqphp_pci.c 	node = func->mem_head;
node             1438 drivers/pci/hotplug/cpqphp_pci.c 	while (node) {
node             1439 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
node             1440 drivers/pci/hotplug/cpqphp_pci.c 		return_resource(&(resources->mem_head), node);
node             1441 drivers/pci/hotplug/cpqphp_pci.c 		node = t_node;
node             1444 drivers/pci/hotplug/cpqphp_pci.c 	node = func->p_mem_head;
node             1446 drivers/pci/hotplug/cpqphp_pci.c 	while (node) {
node             1447 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
node             1448 drivers/pci/hotplug/cpqphp_pci.c 		return_resource(&(resources->p_mem_head), node);
node             1449 drivers/pci/hotplug/cpqphp_pci.c 		node = t_node;
node             1452 drivers/pci/hotplug/cpqphp_pci.c 	node = func->bus_head;
node             1454 drivers/pci/hotplug/cpqphp_pci.c 	while (node) {
node             1455 drivers/pci/hotplug/cpqphp_pci.c 		t_node = node->next;
node             1456 drivers/pci/hotplug/cpqphp_pci.c 		return_resource(&(resources->bus_head), node);
node             1457 drivers/pci/hotplug/cpqphp_pci.c 		node = t_node;
node               38 drivers/pci/of.c 	struct device_node *node;
node               41 drivers/pci/of.c 		node = pcibios_get_phb_of_node(bus);
node               43 drivers/pci/of.c 		node = of_node_get(bus->self->dev.of_node);
node               44 drivers/pci/of.c 		if (node && of_property_read_bool(node, "external-facing"))
node               48 drivers/pci/of.c 	bus->dev.of_node = node;
node              106 drivers/pci/of.c static inline int __of_pci_pci_compare(struct device_node *node,
node              111 drivers/pci/of.c 	devfn = of_pci_get_devfn(node);
node              121 drivers/pci/of.c 	struct device_node *node, *node2;
node              123 drivers/pci/of.c 	for_each_child_of_node(parent, node) {
node              124 drivers/pci/of.c 		if (__of_pci_pci_compare(node, devfn))
node              125 drivers/pci/of.c 			return node;
node              131 drivers/pci/of.c 		if (of_node_name_eq(node, "multifunc-device")) {
node              132 drivers/pci/of.c 			for_each_child_of_node(node, node2) {
node              134 drivers/pci/of.c 					of_node_put(node);
node              173 drivers/pci/of.c int of_pci_parse_bus_range(struct device_node *node, struct resource *res)
node              178 drivers/pci/of.c 	error = of_property_read_u32_array(node, "bus-range", bus_range,
node              183 drivers/pci/of.c 	res->name = node->name;
node              201 drivers/pci/of.c int of_get_pci_domain_nr(struct device_node *node)
node              206 drivers/pci/of.c 	error = of_property_read_u32(node, "linux,pci-domain", &domain);
node              545 drivers/pci/of.c int of_pci_get_max_link_speed(struct device_node *node)
node              549 drivers/pci/of.c 	if (of_property_read_u32(node, "max-link-speed", &max_link_speed) ||
node               55 drivers/pci/pci-acpi.c 	entry = list_first_entry(&list, struct resource_entry, node);
node               25 drivers/pci/pci-driver.c 	struct list_head node;
node               71 drivers/pci/pci-driver.c 	list_add_tail(&dynid->node, &drv->dynids.list);
node               83 drivers/pci/pci-driver.c 	list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) {
node               84 drivers/pci/pci-driver.c 		list_del(&dynid->node);
node              183 drivers/pci/pci-driver.c 	list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) {
node              190 drivers/pci/pci-driver.c 			list_del(&dynid->node);
node              263 drivers/pci/pci-driver.c 	list_for_each_entry(dynid, &drv->dynids.list, node) {
node              334 drivers/pci/pci-driver.c 	int error, node, cpu;
node              342 drivers/pci/pci-driver.c 	node = dev_to_node(&dev->dev);
node              351 drivers/pci/pci-driver.c 	if (node < 0 || node >= MAX_NUMNODES || !node_online(node) ||
node              355 drivers/pci/pci-driver.c 		cpu = cpumask_any_and(cpumask_of_node(node), cpu_online_mask);
node              323 drivers/pci/pci-sysfs.c 	int node, ret;
node              328 drivers/pci/pci-sysfs.c 	ret = kstrtoint(buf, 0, &node);
node              332 drivers/pci/pci-sysfs.c 	if ((node < 0 && node != NUMA_NO_NODE) || node >= MAX_NUMNODES)
node              335 drivers/pci/pci-sysfs.c 	if (node != NUMA_NO_NODE && !node_online(node))
node              340 drivers/pci/pci-sysfs.c 		  node);
node              342 drivers/pci/pci-sysfs.c 	dev->numa_node = node;
node              158 drivers/pci/pci.c 	list_for_each_entry(tmp, &bus->children, node) {
node              604 drivers/pci/pci.h int of_pci_parse_bus_range(struct device_node *node, struct resource *res);
node              605 drivers/pci/pci.h int of_get_pci_domain_nr(struct device_node *node);
node              606 drivers/pci/pci.h int of_pci_get_max_link_speed(struct device_node *node);
node              614 drivers/pci/pci.h of_pci_parse_bus_range(struct device_node *node, struct resource *res)
node              620 drivers/pci/pci.h of_get_pci_domain_nr(struct device_node *node)
node              626 drivers/pci/pci.h of_pci_get_max_link_speed(struct device_node *node)
node              553 drivers/pci/probe.c 	INIT_LIST_HEAD(&b->node);
node              903 drivers/pci/probe.c 		list_move_tail(&window->node, &bridge->windows);
node              928 drivers/pci/probe.c 	list_add_tail(&bus->node, &pci_root_buses);
node             1066 drivers/pci/probe.c 		list_add_tail(&child->node, &parent->children);
node             2955 drivers/pci/probe.c 		list_for_each_entry(child, &bus->children, node)
node               52 drivers/pci/remove.c 	list_del(&bus->node);
node              118 drivers/pci/search.c 	list_for_each_entry(tmp, &bus->children, node) {
node              167 drivers/pci/search.c 	n = from ? from->node.next : pci_root_buses.next;
node              169 drivers/pci/search.c 		b = list_entry(n, struct pci_bus, node);
node             1439 drivers/pci/setup-bus.c 	list_for_each_entry(child, &b->children, node)
node             1620 drivers/pci/setup-bus.c 	list_for_each_entry(child_bus, &bus->children, node) {
node             1814 drivers/pci/setup-bus.c 	list_for_each_entry(root_bus, &pci_root_buses, node) {
node               72 drivers/pci/xen-pcifront.c 	sd->sd.node = first_online_node;
node              307 drivers/pcmcia/cistpl.c 	list_for_each_entry(cis, &s->cis_cache, node) {
node              325 drivers/pcmcia/cistpl.c 			list_add(&cis->node, &s->cis_cache);
node              339 drivers/pcmcia/cistpl.c 	list_for_each_entry(cis, &s->cis_cache, node)
node              341 drivers/pcmcia/cistpl.c 			list_del(&cis->node);
node              361 drivers/pcmcia/cistpl.c 		cis = list_entry(l, struct cis_cache_entry, node);
node              362 drivers/pcmcia/cistpl.c 		list_del(&cis->node);
node              385 drivers/pcmcia/cistpl.c 	list_for_each_entry(cis, &s->cis_cache, node) {
node               39 drivers/pcmcia/cs_internal.h 	struct list_head	node;
node               81 drivers/pcmcia/ds.c 	struct list_head		node;
node              124 drivers/pcmcia/ds.c 	list_add_tail(&dynid->node, &pdrv->dynids.list);
node              141 drivers/pcmcia/ds.c 	list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) {
node              142 drivers/pcmcia/ds.c 		list_del(&dynid->node);
node              913 drivers/pcmcia/ds.c 	list_for_each_entry(dynid, &p_drv->dynids.list, node) {
node               85 drivers/pcmcia/soc_common.h 	struct list_head	node;
node             1100 drivers/pcmcia/yenta_socket.c 			node) {
node              164 drivers/perf/arm-ccn.c 	struct hlist_node node;
node              178 drivers/perf/arm-ccn.c 	struct arm_ccn_component *node;
node              187 drivers/perf/arm-ccn.c static int arm_ccn_node_to_xp(int node)
node              189 drivers/perf/arm-ccn.c 	return node / CCN_NUM_XP_PORTS;
node              192 drivers/perf/arm-ccn.c static int arm_ccn_node_to_xp_port(int node)
node              194 drivers/perf/arm-ccn.c 	return node % CCN_NUM_XP_PORTS;
node              232 drivers/perf/arm-ccn.c static CCN_FORMAT_ATTR(node, "config:0-7");
node              671 drivers/perf/arm-ccn.c 		source = &ccn->node[node_xp];
node              782 drivers/perf/arm-ccn.c 		if (!arm_ccn_pmu_type_eq(type, ccn->node[node_xp].type)) {
node             1205 drivers/perf/arm-ccn.c static int arm_ccn_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node             1207 drivers/perf/arm-ccn.c 	struct arm_ccn_dt *dt = hlist_entry_safe(node, struct arm_ccn_dt, node);
node             1307 drivers/perf/arm-ccn.c 					 &ccn->dt.node);
node             1317 drivers/perf/arm-ccn.c 					    &ccn->dt.node);
node             1332 drivers/perf/arm-ccn.c 					    &ccn->dt.node);
node             1403 drivers/perf/arm-ccn.c 		component = &ccn->node[id];
node             1409 drivers/perf/arm-ccn.c 		component = &ccn->node[id];
node             1514 drivers/perf/arm-ccn.c 	ccn->node = devm_kcalloc(ccn->dev, ccn->num_nodes, sizeof(*ccn->node),
node             1516 drivers/perf/arm-ccn.c 	ccn->xp = devm_kcalloc(ccn->dev, ccn->num_xps, sizeof(*ccn->node),
node             1518 drivers/perf/arm-ccn.c 	if (!ccn->node || !ccn->xp)
node              767 drivers/perf/arm_dsu_pmu.c static int dsu_pmu_cpu_online(unsigned int cpu, struct hlist_node *node)
node              769 drivers/perf/arm_dsu_pmu.c 	struct dsu_pmu *dsu_pmu = hlist_entry_safe(node, struct dsu_pmu,
node              785 drivers/perf/arm_dsu_pmu.c static int dsu_pmu_cpu_teardown(unsigned int cpu, struct hlist_node *node)
node              788 drivers/perf/arm_dsu_pmu.c 	struct dsu_pmu *dsu_pmu = hlist_entry_safe(node, struct dsu_pmu,
node              615 drivers/perf/arm_pmu.c static int arm_perf_starting_cpu(unsigned int cpu, struct hlist_node *node)
node              617 drivers/perf/arm_pmu.c 	struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node);
node              638 drivers/perf/arm_pmu.c static int arm_perf_teardown_cpu(unsigned int cpu, struct hlist_node *node)
node              640 drivers/perf/arm_pmu.c 	struct arm_pmu *pmu = hlist_entry_safe(node, struct arm_pmu, node);
node              757 drivers/perf/arm_pmu.c 				       &cpu_pmu->node);
node              769 drivers/perf/arm_pmu.c 					    &cpu_pmu->node);
node              778 drivers/perf/arm_pmu.c 					    &cpu_pmu->node);
node               60 drivers/perf/arm_pmu_platform.c static bool pmu_has_irq_affinity(struct device_node *node)
node               62 drivers/perf/arm_pmu_platform.c 	return !!of_find_property(node, "interrupt-affinity", NULL);
node               65 drivers/perf/arm_pmu_platform.c static int pmu_parse_irq_affinity(struct device_node *node, int i)
node               75 drivers/perf/arm_pmu_platform.c 	if (!pmu_has_irq_affinity(node))
node               78 drivers/perf/arm_pmu_platform.c 	dn = of_parse_phandle(node, "interrupt-affinity", i);
node               81 drivers/perf/arm_pmu_platform.c 			i, node);
node              196 drivers/perf/arm_pmu_platform.c 	struct device_node *node = pdev->dev.of_node;
node              210 drivers/perf/arm_pmu_platform.c 	if (node && (of_id = of_match_node(of_table, pdev->dev.of_node))) {
node              229 drivers/perf/arm_pmu_platform.c 		pr_info("%pOF: failed to probe PMU!\n", node);
node              246 drivers/perf/arm_pmu_platform.c 	pr_info("%pOF: failed to register PMU devices!\n", node);
node              102 drivers/perf/arm_smmuv3_pmu.c 	struct hlist_node node;
node              585 drivers/perf/arm_smmuv3_pmu.c static int smmu_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              590 drivers/perf/arm_smmuv3_pmu.c 	smmu_pmu = hlist_entry_safe(node, struct smmu_pmu, node);
node              814 drivers/perf/arm_smmuv3_pmu.c 					       &smmu_pmu->node);
node              836 drivers/perf/arm_smmuv3_pmu.c 	cpuhp_state_remove_instance_nocalls(cpuhp_state_num, &smmu_pmu->node);
node              845 drivers/perf/arm_smmuv3_pmu.c 	cpuhp_state_remove_instance_nocalls(cpuhp_state_num, &smmu_pmu->node);
node             1071 drivers/perf/arm_spe_pmu.c static int arm_spe_pmu_cpu_startup(unsigned int cpu, struct hlist_node *node)
node             1075 drivers/perf/arm_spe_pmu.c 	spe_pmu = hlist_entry_safe(node, struct arm_spe_pmu, hotplug_node);
node             1083 drivers/perf/arm_spe_pmu.c static int arm_spe_pmu_cpu_teardown(unsigned int cpu, struct hlist_node *node)
node             1087 drivers/perf/arm_spe_pmu.c 	spe_pmu = hlist_entry_safe(node, struct arm_spe_pmu, hotplug_node);
node               72 drivers/perf/fsl_imx8_ddr_perf.c 	struct	hlist_node node;
node              518 drivers/perf/fsl_imx8_ddr_perf.c static int ddr_perf_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              520 drivers/perf/fsl_imx8_ddr_perf.c 	struct ddr_pmu *pmu = hlist_entry_safe(node, struct ddr_pmu, node);
node              583 drivers/perf/fsl_imx8_ddr_perf.c 	ret = cpuhp_state_add_instance_nocalls(pmu->cpuhp_state, &pmu->node);
node              621 drivers/perf/fsl_imx8_ddr_perf.c 	cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node);
node              634 drivers/perf/fsl_imx8_ddr_perf.c 	cpuhp_state_remove_instance_nocalls(pmu->cpuhp_state, &pmu->node);
node              374 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c 				       &ddrc_pmu->node);
node              401 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c 					    &ddrc_pmu->node);
node              413 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c 				    &ddrc_pmu->node);
node              385 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c 				       &hha_pmu->node);
node              412 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c 					    &hha_pmu->node);
node              424 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c 				    &hha_pmu->node);
node              375 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c 				       &l3c_pmu->node);
node              402 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c 					    &l3c_pmu->node);
node              414 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c 				    &l3c_pmu->node);
node              383 drivers/perf/hisilicon/hisi_uncore_pmu.c int hisi_uncore_pmu_online_cpu(unsigned int cpu, struct hlist_node *node)
node              385 drivers/perf/hisilicon/hisi_uncore_pmu.c 	struct hisi_pmu *hisi_pmu = hlist_entry_safe(node, struct hisi_pmu,
node              386 drivers/perf/hisilicon/hisi_uncore_pmu.c 						     node);
node              406 drivers/perf/hisilicon/hisi_uncore_pmu.c int hisi_uncore_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              408 drivers/perf/hisilicon/hisi_uncore_pmu.c 	struct hisi_pmu *hisi_pmu = hlist_entry_safe(node, struct hisi_pmu,
node              409 drivers/perf/hisilicon/hisi_uncore_pmu.c 						     node);
node               67 drivers/perf/hisilicon/hisi_uncore_pmu.h 	struct hlist_node node;
node               97 drivers/perf/hisilicon/hisi_uncore_pmu.h int hisi_uncore_pmu_online_cpu(unsigned int cpu, struct hlist_node *node);
node               98 drivers/perf/hisilicon/hisi_uncore_pmu.h int hisi_uncore_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node);
node              152 drivers/perf/qcom_l2_pmu.c 	struct hlist_node node;
node              812 drivers/perf/qcom_l2_pmu.c static int l2cache_pmu_online_cpu(unsigned int cpu, struct hlist_node *node)
node              817 drivers/perf/qcom_l2_pmu.c 	l2cache_pmu = hlist_entry_safe(node, struct l2cache_pmu, node);
node              847 drivers/perf/qcom_l2_pmu.c static int l2cache_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              854 drivers/perf/qcom_l2_pmu.c 	l2cache_pmu = hlist_entry_safe(node, struct l2cache_pmu, node);
node              993 drivers/perf/qcom_l2_pmu.c 				       &l2cache_pmu->node);
node             1012 drivers/perf/qcom_l2_pmu.c 				    &l2cache_pmu->node);
node             1023 drivers/perf/qcom_l2_pmu.c 				    &l2cache_pmu->node);
node              156 drivers/perf/qcom_l3_pmu.c 	struct hlist_node	node;
node              706 drivers/perf/qcom_l3_pmu.c static int qcom_l3_cache_pmu_online_cpu(unsigned int cpu, struct hlist_node *node)
node              708 drivers/perf/qcom_l3_pmu.c 	struct l3cache_pmu *l3pmu = hlist_entry_safe(node, struct l3cache_pmu, node);
node              717 drivers/perf/qcom_l3_pmu.c static int qcom_l3_cache_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node              719 drivers/perf/qcom_l3_pmu.c 	struct l3cache_pmu *l3pmu = hlist_entry_safe(node, struct l3cache_pmu, node);
node              790 drivers/perf/qcom_l3_pmu.c 	ret = cpuhp_state_add_instance(CPUHP_AP_PERF_ARM_QCOM_L3_ONLINE, &l3pmu->node);
node               69 drivers/perf/thunderx2_pmu.c 	int node;
node              579 drivers/perf/thunderx2_pmu.c 	cpu = cpumask_any_and(cpumask_of_node(tx2_pmu->node),
node              627 drivers/perf/thunderx2_pmu.c 	list_for_each_entry(rentry, &list, node) {
node              651 drivers/perf/thunderx2_pmu.c 	tx2_pmu->node = dev_to_node(dev);
node              662 drivers/perf/thunderx2_pmu.c 				"uncore_l3c_%d", tx2_pmu->node);
node              674 drivers/perf/thunderx2_pmu.c 				"uncore_dmc_%d", tx2_pmu->node);
node              728 drivers/perf/thunderx2_pmu.c 		(tx2_pmu->node == cpu_to_node(cpu)))
node              751 drivers/perf/thunderx2_pmu.c 			cpumask_of_node(tx2_pmu->node),
node              803 drivers/perf/thunderx2_pmu.c 			if (tx2_pmu->node == dev_to_node(dev)) {
node              122 drivers/perf/xgene_pmu.c 	struct hlist_node node;
node             1794 drivers/perf/xgene_pmu.c static int xgene_pmu_online_cpu(unsigned int cpu, struct hlist_node *node)
node             1796 drivers/perf/xgene_pmu.c 	struct xgene_pmu *xgene_pmu = hlist_entry_safe(node, struct xgene_pmu,
node             1797 drivers/perf/xgene_pmu.c 						       node);
node             1808 drivers/perf/xgene_pmu.c static int xgene_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
node             1810 drivers/perf/xgene_pmu.c 	struct xgene_pmu *xgene_pmu = hlist_entry_safe(node, struct xgene_pmu,
node             1811 drivers/perf/xgene_pmu.c 						       node);
node             1929 drivers/perf/xgene_pmu.c 				      &xgene_pmu->node);
node             1949 drivers/perf/xgene_pmu.c 				    &xgene_pmu->node);
node             1972 drivers/perf/xgene_pmu.c 				    &xgene_pmu->node);
node              126 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c 	struct device_node *node = dev->of_node, *child;
node              133 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c 	if (of_get_child_count(node) == 0) {
node              151 drivers/phy/broadcom/phy-bcm-cygnus-pcie.c 	for_each_available_child_of_node(node, child) {
node              218 drivers/phy/broadcom/phy-bcm-sr-pcie.c 	struct device_node *node = dev->of_node;
node              235 drivers/phy/broadcom/phy-bcm-sr-pcie.c 	core->cdru = syscon_regmap_lookup_by_phandle(node, "brcm,sr-cdru");
node              241 drivers/phy/broadcom/phy-bcm-sr-pcie.c 	core->mhb = syscon_regmap_lookup_by_phandle(node, "brcm,sr-mhb");
node              286 drivers/phy/broadcom/phy-bcm-sr-usb.c static int bcm_usb_phy_create(struct device *dev, struct device_node *node,
node              309 drivers/phy/broadcom/phy-bcm-sr-usb.c 			phy_cfg[idx].phy = devm_phy_create(dev, node,
node              326 drivers/phy/broadcom/phy-bcm-sr-usb.c 		phy_cfg->phy = devm_phy_create(dev, node, &sr_phy_ops);
node              169 drivers/phy/cadence/phy-cadence-sierra.c 	int i, ret, node = 0;
node              232 drivers/phy/cadence/phy-cadence-sierra.c 		sp->phys[node].lnk_rst =
node              235 drivers/phy/cadence/phy-cadence-sierra.c 		if (IS_ERR(sp->phys[node].lnk_rst)) {
node              238 drivers/phy/cadence/phy-cadence-sierra.c 			ret = PTR_ERR(sp->phys[node].lnk_rst);
node              243 drivers/phy/cadence/phy-cadence-sierra.c 			ret = cdns_sierra_get_optional(&sp->phys[node], child);
node              257 drivers/phy/cadence/phy-cadence-sierra.c 		sp->phys[node].phy = gphy;
node              258 drivers/phy/cadence/phy-cadence-sierra.c 		phy_set_drvdata(gphy, &sp->phys[node]);
node              264 drivers/phy/cadence/phy-cadence-sierra.c 		node++;
node              266 drivers/phy/cadence/phy-cadence-sierra.c 	sp->nsubnodes = node;
node              278 drivers/phy/cadence/phy-cadence-sierra.c 	node++;
node              280 drivers/phy/cadence/phy-cadence-sierra.c 	for (i = 0; i < node; i++)
node               80 drivers/phy/phy-core.c 	list_add_tail(&pl->node, &phys);
node              104 drivers/phy/phy-core.c 	list_for_each_entry(pl, &phys, node)
node              107 drivers/phy/phy-core.c 			list_del(&pl->node);
node              121 drivers/phy/phy-core.c 	list_for_each_entry(p, &phys, node)
node              131 drivers/phy/phy-core.c static struct phy_provider *of_phy_provider_lookup(struct device_node *node)
node              137 drivers/phy/phy-core.c 		if (phy_provider->dev->of_node == node)
node              141 drivers/phy/phy-core.c 			if (child == node)
node              834 drivers/phy/phy-core.c struct phy *phy_create(struct device *dev, struct device_node *node,
node              860 drivers/phy/phy-core.c 	phy->dev.of_node = node ?: dev->of_node;
node              910 drivers/phy/phy-core.c struct phy *devm_phy_create(struct device *dev, struct device_node *node,
node              919 drivers/phy/phy-core.c 	phy = phy_create(dev, node, ops);
node              321 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	struct device_node *node = rphy->dev->of_node;
node              331 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	of_property_read_string(node, "clock-output-names", &init.name);
node              351 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	ret = of_clk_add_provider(node, of_clk_src_simple_get, rphy->clk480m);
node              363 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	of_clk_del_provider(node);
node              373 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	struct device_node *node = rphy->dev->of_node;
node              376 drivers/phy/rockchip/phy-rockchip-inno-usb2.c 	if (of_property_read_bool(node, "extcon")) {
node              842 drivers/phy/samsung/phy-exynos5-usbdrd.c 	struct device_node *node = dev->of_node;
node              888 drivers/phy/samsung/phy-exynos5-usbdrd.c 	channel = of_alias_get_id(node, "usbdrdphy");
node              167 drivers/phy/samsung/phy-exynos5250-sata.c 	struct device_node *node;
node              187 drivers/phy/samsung/phy-exynos5250-sata.c 	node = of_parse_phandle(dev->of_node,
node              189 drivers/phy/samsung/phy-exynos5250-sata.c 	if (!node)
node              192 drivers/phy/samsung/phy-exynos5250-sata.c 	sata_phy->client = of_find_i2c_device_by_node(node);
node             1136 drivers/phy/st/phy-miphy28lp.c static int miphy28lp_probe_resets(struct device_node *node,
node             1143 drivers/phy/st/phy-miphy28lp.c 		of_reset_control_get_shared(node, "miphy-sw-rst");
node              397 drivers/phy/ti/phy-am654-serdes.c 	struct device_node *node = am654_phy->of_node;
node              416 drivers/phy/ti/phy-am654-serdes.c 	regmap_node = of_parse_phandle(node, "ti,serdes-clk", 0);
node              430 drivers/phy/ti/phy-am654-serdes.c 	num_parents = of_clk_get_parent_count(node);
node              444 drivers/phy/ti/phy-am654-serdes.c 	of_clk_parent_fill(node, parent_names, num_parents);
node              559 drivers/phy/ti/phy-am654-serdes.c 	struct device_node *node = dev->of_node;
node              589 drivers/phy/ti/phy-am654-serdes.c 	am654_phy->of_node = node;
node              603 drivers/phy/ti/phy-am654-serdes.c 		ret = of_property_read_string_index(node, "clock-output-names",
node              621 drivers/phy/ti/phy-am654-serdes.c 	ret = of_clk_add_provider(node, of_clk_src_onecell_get, clk_data);
node              641 drivers/phy/ti/phy-am654-serdes.c 	of_clk_del_provider(node);
node              649 drivers/phy/ti/phy-am654-serdes.c 	struct device_node *node = am654_phy->of_node;
node              652 drivers/phy/ti/phy-am654-serdes.c 	of_clk_del_provider(node);
node              143 drivers/phy/ti/phy-da8xx-usb.c 	struct device_node	*node = dev->of_node;
node              172 drivers/phy/ti/phy-da8xx-usb.c 	d_phy->usb11_phy = devm_phy_create(dev, node, &da8xx_usb11_phy_ops);
node              178 drivers/phy/ti/phy-da8xx-usb.c 	d_phy->usb20_phy = devm_phy_create(dev, node, &da8xx_usb20_phy_ops);
node              188 drivers/phy/ti/phy-da8xx-usb.c 	if (node) {
node              297 drivers/phy/ti/phy-gmii-sel.c 	struct device_node *node = dev->of_node;
node              313 drivers/phy/ti/phy-gmii-sel.c 	priv->regmap = syscon_node_to_regmap(node->parent);
node              287 drivers/phy/ti/phy-omap-usb2.c 	struct device_node *node = pdev->dev.of_node;
node              326 drivers/phy/ti/phy-omap-usb2.c 	phy->syscon_phy_power = syscon_regmap_lookup_by_phandle(node,
node              333 drivers/phy/ti/phy-omap-usb2.c 		control_node = of_parse_phandle(node, "ctrl-module", 0);
node              347 drivers/phy/ti/phy-omap-usb2.c 		if (of_property_read_u32_index(node,
node              673 drivers/phy/ti/phy-ti-pipe3.c 	struct device_node *node = dev->of_node;
node              677 drivers/phy/ti/phy-ti-pipe3.c 	phy->phy_power_syscon = syscon_regmap_lookup_by_phandle(node,
node              684 drivers/phy/ti/phy-ti-pipe3.c 		if (of_property_read_u32_index(node,
node              693 drivers/phy/ti/phy-ti-pipe3.c 		control_node = of_parse_phandle(node, "ctrl-module", 0);
node              709 drivers/phy/ti/phy-ti-pipe3.c 		phy->pcs_syscon = syscon_regmap_lookup_by_phandle(node,
node              716 drivers/phy/ti/phy-ti-pipe3.c 			if (of_property_read_u32_index(node,
node              727 drivers/phy/ti/phy-ti-pipe3.c 		phy->dpll_reset_syscon = syscon_regmap_lookup_by_phandle(node,
node              734 drivers/phy/ti/phy-ti-pipe3.c 			if (of_property_read_u32_index(node,
node             2632 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		struct device_node *node;
node             2635 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		node = of_parse_phandle(ctx->dev->of_node,
node             2637 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		if (node) {
node             2638 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 			map = syscon_node_to_regmap(node);
node             2639 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 			of_node_put(node);
node             2651 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		struct device_node *node;
node             2654 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		node = of_parse_phandle(ctx->dev->of_node,
node             2656 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 		if (node) {
node             2657 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 			map = syscon_node_to_regmap(node->parent);
node             2658 drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c 			of_node_put(node);
node               51 drivers/pinctrl/berlin/berlin.c 					 struct device_node *node,
node               64 drivers/pinctrl/berlin/berlin.c 	ret = of_property_read_string(node, "function", &function_name);
node               67 drivers/pinctrl/berlin/berlin.c 			"missing function property in node %pOFn\n", node);
node               71 drivers/pinctrl/berlin/berlin.c 	ngroups = of_property_count_strings(node, "groups");
node               74 drivers/pinctrl/berlin/berlin.c 			"missing groups property in node %pOFn\n", node);
node               85 drivers/pinctrl/berlin/berlin.c 	of_property_for_each_string(node, "groups", prop, group_name) {
node              108 drivers/pinctrl/core.c 	list_for_each_entry(pctldev, &pinctrldev_list, node) {
node              127 drivers/pinctrl/core.c 	list_for_each_entry(pctldev, &pinctrldev_list, node)
node              295 drivers/pinctrl/core.c 	list_for_each_entry(range, &pctldev->gpio_ranges, node) {
node              333 drivers/pinctrl/core.c 	list_for_each_entry(pctldev, &pinctrldev_list, node) {
node              336 drivers/pinctrl/core.c 		list_for_each_entry(range, &pctldev->gpio_ranges, node) {
node              376 drivers/pinctrl/core.c 	list_for_each_entry(pctldev, &pinctrldev_list, node) {
node              405 drivers/pinctrl/core.c 	list_add_tail(&range->node, &pctldev->gpio_ranges);
node              466 drivers/pinctrl/core.c 	list_for_each_entry(range, &pctldev->gpio_ranges, node) {
node              511 drivers/pinctrl/core.c 	list_del(&range->node);
node              917 drivers/pinctrl/core.c 	list_for_each_entry(state, &p->states, node)
node              936 drivers/pinctrl/core.c 	list_add_tail(&state->node, &p->states);
node             1001 drivers/pinctrl/core.c 	list_add_tail(&setting->node, &state->settings);
node             1011 drivers/pinctrl/core.c 	list_for_each_entry(p, &pinctrl_list, node)
node             1102 drivers/pinctrl/core.c 	list_add_tail(&p->node, &pinctrl_list);
node             1159 drivers/pinctrl/core.c 	list_for_each_entry_safe(state, n1, &p->states, node) {
node             1160 drivers/pinctrl/core.c 		list_for_each_entry_safe(setting, n2, &state->settings, node) {
node             1162 drivers/pinctrl/core.c 			list_del(&setting->node);
node             1165 drivers/pinctrl/core.c 		list_del(&state->node);
node             1172 drivers/pinctrl/core.c 		list_del(&p->node);
node             1250 drivers/pinctrl/core.c 		list_for_each_entry(setting, &p->state->settings, node) {
node             1260 drivers/pinctrl/core.c 	list_for_each_entry(setting, &state->settings, node) {
node             1290 drivers/pinctrl/core.c 	list_for_each_entry(setting2, &state->settings, node) {
node             1291 drivers/pinctrl/core.c 		if (&setting2->node == &setting->node)
node             1446 drivers/pinctrl/core.c 	list_add_tail(&maps_node->node, &pinctrl_maps);
node             1471 drivers/pinctrl/core.c 	list_for_each_entry(maps_node, &pinctrl_maps, node) {
node             1473 drivers/pinctrl/core.c 			list_del(&maps_node->node);
node             1693 drivers/pinctrl/core.c 	list_for_each_entry(range, &pctldev->gpio_ranges, node) {
node             1725 drivers/pinctrl/core.c 	list_for_each_entry(pctldev, &pinctrldev_list, node) {
node             1808 drivers/pinctrl/core.c 	list_for_each_entry(p, &pinctrl_list, node) {
node             1813 drivers/pinctrl/core.c 		list_for_each_entry(state, &p->states, node) {
node             1816 drivers/pinctrl/core.c 			list_for_each_entry(setting, &state->settings, node) {
node             1969 drivers/pinctrl/core.c 	INIT_LIST_HEAD(&pctldev->node);
node             2065 drivers/pinctrl/core.c 	list_add_tail(&pctldev->node, &pinctrldev_list);
node             2159 drivers/pinctrl/core.c 	list_del(&pctldev->node);
node             2166 drivers/pinctrl/core.c 	list_for_each_entry_safe(range, n, &pctldev->gpio_ranges, node)
node             2167 drivers/pinctrl/core.c 		list_del(&range->node);
node               43 drivers/pinctrl/core.h 	struct list_head node;
node               78 drivers/pinctrl/core.h 	struct list_head node;
node               93 drivers/pinctrl/core.h 	struct list_head node;
node              132 drivers/pinctrl/core.h 	struct list_head node;
node              180 drivers/pinctrl/core.h 	struct list_head node;
node              250 drivers/pinctrl/core.h 	list_for_each_entry(_maps_node_, &pinctrl_maps, node) \
node               23 drivers/pinctrl/devicetree.c 	struct list_head node;
node               53 drivers/pinctrl/devicetree.c 	list_for_each_entry_safe(dt_map, n1, &p->dt_maps, node) {
node               55 drivers/pinctrl/devicetree.c 		list_del(&dt_map->node);
node               93 drivers/pinctrl/devicetree.c 	list_add_tail(&dt_map->node, &p->dt_maps);
node              491 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 				      struct device_node *node,
node              507 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	pins = of_find_property(node, "pinmux", NULL);
node              510 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 				node);
node              514 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	err = pinconf_generic_parse_dt_config(node, pctldev, &configs,
node              543 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 		err = of_property_read_u32_index(node, "pinmux",
node             1022 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	struct device_node *np = pdev->dev.of_node, *node;
node             1038 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	node = of_parse_phandle(np, "mediatek,pctl-regmap", 0);
node             1039 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	if (node) {
node             1040 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 		pctl->regmap1 = syscon_node_to_regmap(node);
node             1051 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	node = of_parse_phandle(np, "mediatek,pctl-regmap", 1);
node             1052 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 	if (node) {
node             1053 drivers/pinctrl/mediatek/pinctrl-mtk-common.c 		pctl->regmap2 = syscon_node_to_regmap(node);
node              426 drivers/pinctrl/mediatek/pinctrl-paris.c 				       struct device_node *node,
node              441 drivers/pinctrl/mediatek/pinctrl-paris.c 	pins = of_find_property(node, "pinmux", NULL);
node              444 drivers/pinctrl/mediatek/pinctrl-paris.c 			node);
node              448 drivers/pinctrl/mediatek/pinctrl-paris.c 	err = pinconf_generic_parse_dt_config(node, pctldev, &configs,
node              477 drivers/pinctrl/mediatek/pinctrl-paris.c 		err = of_property_read_u32_index(node, "pinmux", i, &pinfunc);
node              621 drivers/pinctrl/meson/pinctrl-meson.c 					 struct device_node *node, char *name)
node              627 drivers/pinctrl/meson/pinctrl-meson.c 	i = of_property_match_string(node, "reg-names", name);
node              628 drivers/pinctrl/meson/pinctrl-meson.c 	if (of_address_to_resource(node, i, &res))
node              637 drivers/pinctrl/meson/pinctrl-meson.c 						  "%pOFn-%s", node,
node              646 drivers/pinctrl/meson/pinctrl-meson.c 				  struct device_node *node)
node              650 drivers/pinctrl/meson/pinctrl-meson.c 	for_each_child_of_node(node, np) {
node              131 drivers/pinctrl/pinctrl-at91-pio4.c 	struct device_node	*node;
node              470 drivers/pinctrl/pinctrl-at91-pio4.c 	if (np->parent == atmel_pioctrl->node)
node              946 drivers/pinctrl/pinctrl-at91-pio4.c 	atmel_pioctrl->node = dev->of_node;
node             1939 drivers/pinctrl/pinctrl-ingenic.c 				     struct device_node *node)
node             1946 drivers/pinctrl/pinctrl-ingenic.c 	err = of_property_read_u32(node, "reg", &bank);
node             1971 drivers/pinctrl/pinctrl-ingenic.c 	jzgc->gc.of_node = node;
node             1980 drivers/pinctrl/pinctrl-ingenic.c 	if (of_property_read_bool(node, "gpio-ranges")) {
node             1989 drivers/pinctrl/pinctrl-ingenic.c 	jzgc->irq = irq_of_parse_and_map(node, 0);
node             2023 drivers/pinctrl/pinctrl-ingenic.c 	struct device_node *node;
node             2126 drivers/pinctrl/pinctrl-ingenic.c 	for_each_child_of_node(dev->of_node, node) {
node             2127 drivers/pinctrl/pinctrl-ingenic.c 		if (of_match_node(ingenic_gpio_of_match, node)) {
node             2128 drivers/pinctrl/pinctrl-ingenic.c 			err = ingenic_gpio_probe(jzpc, node);
node             1347 drivers/pinctrl/pinctrl-pistachio.c 	struct device_node *node = pctl->dev->of_node;
node             1357 drivers/pinctrl/pinctrl-pistachio.c 		child = of_get_child_by_name(node, child_name);
node             3187 drivers/pinctrl/pinctrl-rockchip.c 		struct device_node *node;
node             3189 drivers/pinctrl/pinctrl-rockchip.c 		node = of_parse_phandle(bank->of_node->parent,
node             3191 drivers/pinctrl/pinctrl-rockchip.c 		if (!node) {
node             3208 drivers/pinctrl/pinctrl-rockchip.c 		of_node_put(node);
node             3228 drivers/pinctrl/pinctrl-rockchip.c 	struct device_node *node = pdev->dev.of_node;
node             3234 drivers/pinctrl/pinctrl-rockchip.c 	match = of_match_node(rockchip_pinctrl_dt_match, node);
node             3237 drivers/pinctrl/pinctrl-rockchip.c 	for_each_child_of_node(node, np) {
node             3404 drivers/pinctrl/pinctrl-rockchip.c 	struct device_node *np = pdev->dev.of_node, *node;
node             3427 drivers/pinctrl/pinctrl-rockchip.c 	node = of_parse_phandle(np, "rockchip,grf", 0);
node             3428 drivers/pinctrl/pinctrl-rockchip.c 	if (node) {
node             3429 drivers/pinctrl/pinctrl-rockchip.c 		info->regmap_base = syscon_node_to_regmap(node);
node             3463 drivers/pinctrl/pinctrl-rockchip.c 	node = of_parse_phandle(np, "rockchip,pmu", 0);
node             3464 drivers/pinctrl/pinctrl-rockchip.c 	if (node) {
node             3465 drivers/pinctrl/pinctrl-rockchip.c 		info->regmap_pmu = syscon_node_to_regmap(node);
node               96 drivers/pinctrl/pinctrl-single.c 	struct list_head node;
node              110 drivers/pinctrl/pinctrl-single.c 	struct list_head node;
node              392 drivers/pinctrl/pinctrl-single.c 		frange = list_entry(pos, struct pcs_gpiofunc_range, node);
node             1316 drivers/pinctrl/pinctrl-single.c static int pcs_add_gpio_func(struct device_node *node, struct pcs_device *pcs)
node             1325 drivers/pinctrl/pinctrl-single.c 		ret = of_parse_phandle_with_args(node, propname, cellname,
node             1341 drivers/pinctrl/pinctrl-single.c 		list_add_tail(&range->node, &pcs->gpiofuncs);
node             1356 drivers/pinctrl/pinctrl-single.c 	struct list_head node;
node             1377 drivers/pinctrl/pinctrl-single.c 		pcswi = list_entry(pos, struct pcs_interrupt, node);
node             1458 drivers/pinctrl/pinctrl-single.c 		pcswi = list_entry(pos, struct pcs_interrupt, node);
node             1524 drivers/pinctrl/pinctrl-single.c 	list_add_tail(&pcswi->node, &pcs->irqs);
node              986 drivers/pinctrl/samsung/pinctrl-samsung.c 	struct device_node *node = pdev->dev.of_node;
node              990 drivers/pinctrl/samsung/pinctrl-samsung.c 	id = of_alias_get_id(node, "pinctrl");
node             1010 drivers/pinctrl/samsung/pinctrl-samsung.c 	struct device_node *node = pdev->dev.of_node;
node             1075 drivers/pinctrl/samsung/pinctrl-samsung.c 	for_each_child_of_node(node, np) {
node              271 drivers/pinctrl/samsung/pinctrl-samsung.h 	struct list_head		node;
node              506 drivers/pinctrl/stm32/pinctrl-stm32.c 				      struct device_node *node,
node              523 drivers/pinctrl/stm32/pinctrl-stm32.c 	pins = of_find_property(node, "pinmux", NULL);
node              526 drivers/pinctrl/stm32/pinctrl-stm32.c 				node);
node              530 drivers/pinctrl/stm32/pinctrl-stm32.c 	err = pinconf_generic_parse_dt_config(node, pctldev, &configs,
node              559 drivers/pinctrl/stm32/pinctrl-stm32.c 		err = of_property_read_u32_index(node, "pinmux",
node              154 drivers/pinctrl/sunxi/pinctrl-sunxi.c static bool sunxi_pctrl_has_bias_prop(struct device_node *node)
node              156 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	return of_find_property(node, "bias-pull-up", NULL) ||
node              157 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		of_find_property(node, "bias-pull-down", NULL) ||
node              158 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		of_find_property(node, "bias-disable", NULL) ||
node              159 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		of_find_property(node, "allwinner,pull", NULL);
node              162 drivers/pinctrl/sunxi/pinctrl-sunxi.c static bool sunxi_pctrl_has_drive_prop(struct device_node *node)
node              164 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	return of_find_property(node, "drive-strength", NULL) ||
node              165 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		of_find_property(node, "allwinner,drive", NULL);
node              168 drivers/pinctrl/sunxi/pinctrl-sunxi.c static int sunxi_pctrl_parse_bias_prop(struct device_node *node)
node              173 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_find_property(node, "bias-pull-up", NULL))
node              176 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_find_property(node, "bias-pull-down", NULL))
node              179 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_find_property(node, "bias-disable", NULL))
node              183 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_property_read_u32(node, "allwinner,pull", &val))
node              198 drivers/pinctrl/sunxi/pinctrl-sunxi.c static int sunxi_pctrl_parse_drive_prop(struct device_node *node)
node              203 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (!of_property_read_u32(node, "drive-strength", &val)) {
node              217 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_property_read_u32(node, "allwinner,drive", &val))
node              223 drivers/pinctrl/sunxi/pinctrl-sunxi.c static const char *sunxi_pctrl_parse_function_prop(struct device_node *node)
node              229 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	ret = of_property_read_string(node, "function", &function);
node              234 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	ret = of_property_read_string(node, "allwinner,function", &function);
node              241 drivers/pinctrl/sunxi/pinctrl-sunxi.c static const char *sunxi_pctrl_find_pins_prop(struct device_node *node,
node              247 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	count = of_property_count_strings(node, "pins");
node              254 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	count = of_property_count_strings(node, "allwinner,pins");
node              263 drivers/pinctrl/sunxi/pinctrl-sunxi.c static unsigned long *sunxi_pctrl_build_pin_config(struct device_node *node,
node              270 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (sunxi_pctrl_has_drive_prop(node))
node              272 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (sunxi_pctrl_has_bias_prop(node))
node              285 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (sunxi_pctrl_has_drive_prop(node)) {
node              286 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		int drive = sunxi_pctrl_parse_drive_prop(node);
node              296 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (sunxi_pctrl_has_bias_prop(node)) {
node              297 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		int pull = sunxi_pctrl_parse_bias_prop(node);
node              320 drivers/pinctrl/sunxi/pinctrl-sunxi.c 				      struct device_node *node,
node              334 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	function = sunxi_pctrl_parse_function_prop(node);
node              337 drivers/pinctrl/sunxi/pinctrl-sunxi.c 			node);
node              341 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	pin_prop = sunxi_pctrl_find_pins_prop(node, &npins);
node              344 drivers/pinctrl/sunxi/pinctrl-sunxi.c 			node);
node              360 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	pinconfig = sunxi_pctrl_build_pin_config(node, &configlen);
node              366 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	of_property_for_each_string(node, pin_prop, prop, group) {
node             1089 drivers/pinctrl/sunxi/pinctrl-sunxi.c 				      struct device_node *node,
node             1318 drivers/pinctrl/sunxi/pinctrl-sunxi.c 					struct device_node *node)
node             1327 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (of_clk_get_parent_count(node) != 3)
node             1331 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	if (!of_find_property(node, "input-debounce", NULL))
node             1346 drivers/pinctrl/sunxi/pinctrl-sunxi.c 		ret = of_property_read_u32_index(node, "input-debounce",
node             1383 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	struct device_node *node = pdev->dev.of_node;
node             1502 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	ret = of_clk_get_parent_count(node);
node             1530 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	pctl->domain = irq_domain_add_linear(node,
node             1561 drivers/pinctrl/sunxi/pinctrl-sunxi.c 	sunxi_pinctrl_setup_debounce(pctl, node);
node               29 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c 	struct list_head node;
node              705 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c 	list_for_each_entry(r, &priv->reg_regions, node) {
node              721 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c 	list_for_each_entry(r, &priv->reg_regions, node) {
node              760 drivers/pinctrl/uniphier/pinctrl-uniphier-core.c 	list_add_tail(&region->node, &priv->reg_regions);
node               49 drivers/platform/chrome/cros_ec_chardev.c 	struct list_head node;
node              117 drivers/platform/chrome/cros_ec_chardev.c 	list_add_tail(&event->node, &priv->events);
node              149 drivers/platform/chrome/cros_ec_chardev.c 	event = list_first_entry(&priv->events, struct ec_event, node);
node              150 drivers/platform/chrome/cros_ec_chardev.c 	list_del(&event->node);
node              263 drivers/platform/chrome/cros_ec_chardev.c 	list_for_each_entry_safe(event, e, &priv->events, node) {
node              264 drivers/platform/chrome/cros_ec_chardev.c 		list_del(&event->node);
node               28 drivers/platform/olpc/olpc-ec.c 	struct list_head node;
node               87 drivers/platform/olpc/olpc-ec.c 		desc = list_first_entry(&ec->cmd_q, struct ec_cmd_desc, node);
node               88 drivers/platform/olpc/olpc-ec.c 		list_del(&desc->node);
node              118 drivers/platform/olpc/olpc-ec.c 	INIT_LIST_HEAD(&desc->node);
node              121 drivers/platform/olpc/olpc-ec.c 	list_add_tail(&desc->node, &ec->cmd_q);
node              853 drivers/platform/x86/asus-wmi.c static int asus_register_rfkill_notifier(struct asus_wmi *asus, char *node)
node              858 drivers/platform/x86/asus-wmi.c 	status = acpi_get_handle(NULL, node, &handle);
node              865 drivers/platform/x86/asus-wmi.c 		pr_warn("Failed to register notify on %s\n", node);
node              870 drivers/platform/x86/asus-wmi.c static void asus_unregister_rfkill_notifier(struct asus_wmi *asus, char *node)
node              875 drivers/platform/x86/asus-wmi.c 	status = acpi_get_handle(NULL, node, &handle);
node              882 drivers/platform/x86/asus-wmi.c 		pr_err("Error removing rfkill notify handler %s\n", node);
node             2362 drivers/platform/x86/asus-wmi.c 	struct asus_wmi_debugfs_node *node = inode->i_private;
node             2364 drivers/platform/x86/asus-wmi.c 	return single_open(file, node->show, node->asus);
node             2396 drivers/platform/x86/asus-wmi.c 		struct asus_wmi_debugfs_node *node = &asus_wmi_debug_files[i];
node             2398 drivers/platform/x86/asus-wmi.c 		node->asus = asus;
node             2399 drivers/platform/x86/asus-wmi.c 		debugfs_create_file(node->name, S_IFREG | S_IRUGO,
node             2400 drivers/platform/x86/asus-wmi.c 				    asus->debug.root, node,
node              634 drivers/platform/x86/eeepc-laptop.c static void eeepc_rfkill_hotplug_update(struct eeepc_laptop *eeepc, char *node)
node              639 drivers/platform/x86/eeepc-laptop.c 	status = acpi_get_handle(NULL, node, &handle);
node              656 drivers/platform/x86/eeepc-laptop.c 					  char *node)
node              661 drivers/platform/x86/eeepc-laptop.c 	status = acpi_get_handle(NULL, node, &handle);
node              671 drivers/platform/x86/eeepc-laptop.c 		pr_warn("Failed to register notify on %s\n", node);
node              682 drivers/platform/x86/eeepc-laptop.c 					     char *node)
node              687 drivers/platform/x86/eeepc-laptop.c 	status = acpi_get_handle(NULL, node, &handle);
node              697 drivers/platform/x86/eeepc-laptop.c 			node);
node              176 drivers/platform/x86/intel_cht_int33fe.c 	if (mux_ref.node) {
node              177 drivers/platform/x86/intel_cht_int33fe.c 		fwnode_handle_put(software_node_fwnode(mux_ref.node));
node              178 drivers/platform/x86/intel_cht_int33fe.c 		mux_ref.node = NULL;
node              204 drivers/platform/x86/intel_cht_int33fe.c 	mux_ref.node = software_node_find_by_name(NULL, "intel-xhci-usb-sw");
node              205 drivers/platform/x86/intel_cht_int33fe.c 	if (!mux_ref.node) {
node              228 drivers/platform/x86/intel_menlow.c 	struct list_head node;
node              395 drivers/platform/x86/intel_menlow.c 	list_add_tail(&attr->node, &intel_menlow_attr_list);
node              472 drivers/platform/x86/intel_menlow.c 	list_for_each_entry_safe(pos, next, &intel_menlow_attr_list, node) {
node              473 drivers/platform/x86/intel_menlow.c 		list_del(&pos->node);
node             6863 drivers/platform/x86/thinkpad_acpi.c 	list_for_each_entry(child, &device->children, node) {
node              192 drivers/pnp/pnpbios/core.c 	struct pnp_bios_node *node;
node              198 drivers/pnp/pnpbios/core.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              199 drivers/pnp/pnpbios/core.c 	if (!node)
node              201 drivers/pnp/pnpbios/core.c 	if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) {
node              202 drivers/pnp/pnpbios/core.c 		kfree(node);
node              205 drivers/pnp/pnpbios/core.c 	pnpbios_read_resources_from_node(dev, node);
node              207 drivers/pnp/pnpbios/core.c 	kfree(node);
node              214 drivers/pnp/pnpbios/core.c 	struct pnp_bios_node *node;
node              221 drivers/pnp/pnpbios/core.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              222 drivers/pnp/pnpbios/core.c 	if (!node)
node              224 drivers/pnp/pnpbios/core.c 	if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) {
node              225 drivers/pnp/pnpbios/core.c 		kfree(node);
node              228 drivers/pnp/pnpbios/core.c 	if (pnpbios_write_resources_to_node(dev, node) < 0) {
node              229 drivers/pnp/pnpbios/core.c 		kfree(node);
node              232 drivers/pnp/pnpbios/core.c 	ret = pnp_bios_set_dev_node(node->handle, (char)PNPMODE_DYNAMIC, node);
node              233 drivers/pnp/pnpbios/core.c 	kfree(node);
node              239 drivers/pnp/pnpbios/core.c static void pnpbios_zero_data_stream(struct pnp_bios_node *node)
node              241 drivers/pnp/pnpbios/core.c 	unsigned char *p = (char *)node->data;
node              242 drivers/pnp/pnpbios/core.c 	unsigned char *end = (char *)(node->data + node->size);
node              266 drivers/pnp/pnpbios/core.c 	struct pnp_bios_node *node;
node              273 drivers/pnp/pnpbios/core.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              274 drivers/pnp/pnpbios/core.c 	if (!node)
node              277 drivers/pnp/pnpbios/core.c 	if (pnp_bios_get_dev_node(&nodenum, (char)PNPMODE_DYNAMIC, node)) {
node              278 drivers/pnp/pnpbios/core.c 		kfree(node);
node              281 drivers/pnp/pnpbios/core.c 	pnpbios_zero_data_stream(node);
node              283 drivers/pnp/pnpbios/core.c 	ret = pnp_bios_set_dev_node(dev->number, (char)PNPMODE_DYNAMIC, node);
node              284 drivers/pnp/pnpbios/core.c 	kfree(node);
node              299 drivers/pnp/pnpbios/core.c static int __init insert_device(struct pnp_bios_node *node)
node              309 drivers/pnp/pnpbios/core.c 		if (dev->number == node->handle)
node              313 drivers/pnp/pnpbios/core.c 	pnp_eisa_id_to_string(node->eisa_id & PNP_EISA_ID_MASK, id);
node              314 drivers/pnp/pnpbios/core.c 	dev = pnp_alloc_dev(&pnpbios_protocol, node->handle, id);
node              318 drivers/pnp/pnpbios/core.c 	pnpbios_parse_data_stream(dev, node);
node              320 drivers/pnp/pnpbios/core.c 	dev->flags = node->flags;
node              341 drivers/pnp/pnpbios/core.c 	pnpbios_interface_attach_device(node);
node              351 drivers/pnp/pnpbios/core.c 	struct pnp_bios_node *node;
node              353 drivers/pnp/pnpbios/core.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              354 drivers/pnp/pnpbios/core.c 	if (!node)
node              364 drivers/pnp/pnpbios/core.c 			    (&nodenum, (char)PNPMODE_DYNAMIC, node))
node              368 drivers/pnp/pnpbios/core.c 			    (&nodenum, (char)PNPMODE_STATIC, node))
node              372 drivers/pnp/pnpbios/core.c 		if (insert_device(node) == 0)
node              382 drivers/pnp/pnpbios/core.c 	kfree(node);
node              153 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_parse_data_stream(struct pnp_dev *dev, struct pnp_bios_node * node);
node              154 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_read_resources_from_node(struct pnp_dev *dev, struct pnp_bios_node *node);
node              155 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_write_resources_to_node(struct pnp_dev *dev, struct pnp_bios_node *node);
node              162 drivers/pnp/pnpbios/pnpbios.h extern int pnpbios_interface_attach_device(struct pnp_bios_node * node);
node              166 drivers/pnp/pnpbios/pnpbios.h static inline int pnpbios_interface_attach_device(struct pnp_bios_node * node) { return 0; }
node              125 drivers/pnp/pnpbios/proc.c 	struct pnp_bios_node *node;
node              128 drivers/pnp/pnpbios/proc.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              129 drivers/pnp/pnpbios/proc.c 	if (!node)
node              135 drivers/pnp/pnpbios/proc.c 		if (pnp_bios_get_dev_node(&nodenum, PNPMODE_DYNAMIC, node))
node              138 drivers/pnp/pnpbios/proc.c 			     node->handle, node->eisa_id,
node              139 drivers/pnp/pnpbios/proc.c 			     node->type_code, node->flags);
node              149 drivers/pnp/pnpbios/proc.c 	kfree(node);
node              156 drivers/pnp/pnpbios/proc.c 	struct pnp_bios_node *node;
node              161 drivers/pnp/pnpbios/proc.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              162 drivers/pnp/pnpbios/proc.c 	if (!node)
node              164 drivers/pnp/pnpbios/proc.c 	if (pnp_bios_get_dev_node(&nodenum, boot, node)) {
node              165 drivers/pnp/pnpbios/proc.c 		kfree(node);
node              168 drivers/pnp/pnpbios/proc.c 	len = node->size - sizeof(struct pnp_bios_node);
node              169 drivers/pnp/pnpbios/proc.c 	seq_write(m, node->data, len);
node              170 drivers/pnp/pnpbios/proc.c 	kfree(node);
node              183 drivers/pnp/pnpbios/proc.c 	struct pnp_bios_node *node;
node              188 drivers/pnp/pnpbios/proc.c 	node = kzalloc(node_info.max_node_size, GFP_KERNEL);
node              189 drivers/pnp/pnpbios/proc.c 	if (!node)
node              191 drivers/pnp/pnpbios/proc.c 	if (pnp_bios_get_dev_node(&nodenum, boot, node)) {
node              195 drivers/pnp/pnpbios/proc.c 	if (count != node->size - sizeof(struct pnp_bios_node)) {
node              199 drivers/pnp/pnpbios/proc.c 	if (copy_from_user(node->data, buf, count)) {
node              203 drivers/pnp/pnpbios/proc.c 	if (pnp_bios_set_dev_node(node->handle, boot, node) != 0) {
node              209 drivers/pnp/pnpbios/proc.c 	kfree(node);
node              222 drivers/pnp/pnpbios/proc.c int pnpbios_interface_attach_device(struct pnp_bios_node *node)
node              226 drivers/pnp/pnpbios/proc.c 	sprintf(name, "%02x", node->handle);
node              232 drivers/pnp/pnpbios/proc.c 				 (void *)(long)(node->handle));
node              238 drivers/pnp/pnpbios/proc.c 			     (void *)(long)(node->handle + 0x100)))
node              771 drivers/pnp/pnpbios/rsparser.c 					struct pnp_bios_node *node)
node              773 drivers/pnp/pnpbios/rsparser.c 	unsigned char *p = (char *)node->data;
node              774 drivers/pnp/pnpbios/rsparser.c 	unsigned char *end = (char *)(node->data + node->size);
node              789 drivers/pnp/pnpbios/rsparser.c 				     struct pnp_bios_node *node)
node              791 drivers/pnp/pnpbios/rsparser.c 	unsigned char *p = (char *)node->data;
node              792 drivers/pnp/pnpbios/rsparser.c 	unsigned char *end = (char *)(node->data + node->size);
node              801 drivers/pnp/pnpbios/rsparser.c 				    struct pnp_bios_node *node)
node              803 drivers/pnp/pnpbios/rsparser.c 	unsigned char *p = (char *)node->data;
node              804 drivers/pnp/pnpbios/rsparser.c 	unsigned char *end = (char *)(node->data + node->size);
node               87 drivers/power/avs/smartreflex.c 	list_for_each_entry(sr_info, &sr_list, node) {
node              228 drivers/power/avs/smartreflex.c 	list_del(&sr_info->node);
node              678 drivers/power/avs/smartreflex.c 	list_for_each_entry(sr_info, &sr_list, node)
node              881 drivers/power/avs/smartreflex.c 	list_add(&sr_info->node, &sr_list);
node              945 drivers/power/avs/smartreflex.c 	list_del(&sr_info->node);
node              974 drivers/power/avs/smartreflex.c 	list_del(&sr_info->node);
node               94 drivers/power/supply/ab8500_btemp.c 	struct list_head node;
node              126 drivers/power/supply/ab8500_btemp.c 	return list_first_entry(&ab8500_btemp_list, struct ab8500_btemp, node);
node             1102 drivers/power/supply/ab8500_btemp.c 	list_add_tail(&di->node, &ab8500_btemp_list);
node              200 drivers/power/supply/ab8500_fg.c 	struct list_head node;
node              249 drivers/power/supply/ab8500_fg.c 					node);
node             2993 drivers/power/supply/ab8500_fg.c 	list_del(&di->node);
node             3210 drivers/power/supply/ab8500_fg.c 	list_add_tail(&di->node, &ab8500_fg_list);
node               73 drivers/power/supply/cpcap-battery.c 	struct list_head node;
node              552 drivers/power/supply/cpcap-battery.c 	list_for_each_entry(d, &ddata->irq_list, node) {
node              619 drivers/power/supply/cpcap-battery.c 	list_add(&d->node, &ddata->irq_list);
node              145 drivers/power/supply/cpcap-charger.c 	struct list_head node;
node              517 drivers/power/supply/cpcap-charger.c 	list_add(&d->node, &ddata->irq_list);
node              336 drivers/powercap/powercap_sys.c 	list_for_each_entry(pos, &powercap_cntrl_list, node) {
node              621 drivers/powercap/powercap_sys.c 	INIT_LIST_HEAD(&control_type->node);
node              633 drivers/powercap/powercap_sys.c 	list_add_tail(&control_type->node, &powercap_cntrl_list);
node              649 drivers/powercap/powercap_sys.c 	list_for_each_entry(pos, &powercap_cntrl_list, node) {
node              651 drivers/powercap/powercap_sys.c 			list_del(&control_type->node);
node              396 drivers/ptp/ptp_qoriq.c 				 struct device_node *node)
node              407 drivers/ptp/ptp_qoriq.c 	clk = of_clk_get(node, 0);
node              451 drivers/ptp/ptp_qoriq.c 	struct device_node *node = ptp_qoriq->dev->of_node;
node              457 drivers/ptp/ptp_qoriq.c 	if (!node)
node              463 drivers/ptp/ptp_qoriq.c 	if (of_property_read_u32(node, "fsl,cksel", &ptp_qoriq->cksel))
node              466 drivers/ptp/ptp_qoriq.c 	if (of_property_read_bool(node, "fsl,extts-fifo"))
node              471 drivers/ptp/ptp_qoriq.c 	if (of_property_read_u32(node,
node              473 drivers/ptp/ptp_qoriq.c 	    of_property_read_u32(node,
node              475 drivers/ptp/ptp_qoriq.c 	    of_property_read_u32(node,
node              477 drivers/ptp/ptp_qoriq.c 	    of_property_read_u32(node,
node              479 drivers/ptp/ptp_qoriq.c 	    of_property_read_u32(node,
node              481 drivers/ptp/ptp_qoriq.c 	    of_property_read_u32(node,
node              485 drivers/ptp/ptp_qoriq.c 		if (ptp_qoriq_auto_config(ptp_qoriq, node))
node              489 drivers/ptp/ptp_qoriq.c 	if (of_property_read_bool(node, "little-endian")) {
node              498 drivers/ptp/ptp_qoriq.c 	if (of_device_is_compatible(node, "fsl,etsec-ptp")) {
node              198 drivers/pwm/pwm-lp3943.c 	struct device_node *node = dev->of_node;
node              205 drivers/pwm/pwm-lp3943.c 	if (!node)
node              218 drivers/pwm/pwm-lp3943.c 		if (!of_get_property(node, name[i], &proplen))
node              230 drivers/pwm/pwm-lp3943.c 		err = of_property_read_u32_array(node, name[i], output,
node               24 drivers/pwm/pwm-tipwmss.c 	struct device_node *node = pdev->dev.of_node;
node               29 drivers/pwm/pwm-tipwmss.c 	ret = of_platform_populate(node, NULL, NULL, &pdev->dev);
node              122 drivers/rapidio/devices/rio_mport_cdev.c 	struct list_head node;
node              164 drivers/rapidio/devices/rio_mport_cdev.c 	struct list_head	node;
node              385 drivers/rapidio/devices/rio_mport_cdev.c 	list_add_tail(&map->node, &md->mappings);
node              401 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry(map, &md->mappings, node) {
node              475 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry_safe(map, _map, &md->mappings, node) {
node              536 drivers/rapidio/devices/rio_mport_cdev.c 	struct list_head node;
node              760 drivers/rapidio/devices/rio_mport_cdev.c 		list_add_tail(&req->node, &priv->async_list);
node              904 drivers/rapidio/devices/rio_mport_cdev.c 		list_for_each_entry(map, &md->mappings, node) {
node             1032 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry(req, &priv->async_list, node) {
node             1034 drivers/rapidio/devices/rio_mport_cdev.c 			list_del(&req->node);
node             1082 drivers/rapidio/devices/rio_mport_cdev.c 	list_add_tail(&req->node, &priv->async_list);
node             1109 drivers/rapidio/devices/rio_mport_cdev.c 	list_add_tail(&map->node, &md->mappings);
node             1156 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry_safe(map, _map, &md->mappings, node) {
node             1236 drivers/rapidio/devices/rio_mport_cdev.c 	list_add_tail(&map->node, &md->mappings);
node             1261 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry(map, &md->mappings, node) {
node             1342 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry_safe(map, _map, &md->mappings, node) {
node             1971 drivers/rapidio/devices/rio_mport_cdev.c 		list_for_each_entry_safe(req, req_next, &list, node) {
node             1975 drivers/rapidio/devices/rio_mport_cdev.c 			list_del(&req->node);
node             2040 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry_safe(map, _map, &chdev->mappings, node) {
node             2152 drivers/rapidio/devices/rio_mport_cdev.c 	list_del(&map->node);
node             2209 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry(map, &md->mappings, node) {
node             2429 drivers/rapidio/devices/rio_mport_cdev.c 	list_add_tail(&md->node, &mport_devs);
node             2514 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry_safe(map, _map, &md->mappings, node) {
node             2569 drivers/rapidio/devices/rio_mport_cdev.c 	list_for_each_entry(chdev, &mport_devs, node) {
node             2572 drivers/rapidio/devices/rio_mport_cdev.c 			list_del(&chdev->node);
node              413 drivers/rapidio/devices/tsi721.c 		list_for_each_entry(dbell, &mport->dbells, node) {
node             1173 drivers/rapidio/devices/tsi721.c 				list_add_tail(&map->node, &ib_win->mappings);
node             1208 drivers/rapidio/devices/tsi721.c 		list_add_tail(&map->node, &ib_win->mappings);
node             1266 drivers/rapidio/devices/tsi721.c 						    &ib_win->mappings, node) {
node             1268 drivers/rapidio/devices/tsi721.c 						list_del(&map->node);
node              839 drivers/rapidio/devices/tsi721.h 	struct list_head node;
node              897 drivers/rapidio/rio-scan.c 		list_for_each_entry(rswitch, &net->switches, node) {
node             1029 drivers/rapidio/rio-scan.c 	list_for_each_entry(rswitch, &net->switches, node) {
node               36 drivers/rapidio/rio.c 	struct list_head node;
node              113 drivers/rapidio/rio.c 		INIT_LIST_HEAD(&net->node);
node              131 drivers/rapidio/rio.c 	list_add_tail(&net->node, &rio_nets);
node              141 drivers/rapidio/rio.c 	if (!list_empty(&net->node))
node              142 drivers/rapidio/rio.c 		list_del(&net->node);
node              186 drivers/rapidio/rio.c 			list_add_tail(&rdev->rswitch->node,
node              212 drivers/rapidio/rio.c 			list_del(&rdev->rswitch->node);
node              416 drivers/rapidio/rio.c 	list_add_tail(&dbell->node, &mport->dbells);
node              480 drivers/rapidio/rio.c 	list_for_each_entry(dbell, &mport->dbells, node) {
node              482 drivers/rapidio/rio.c 			list_del(&dbell->node);
node              574 drivers/rapidio/rio.c 	list_add_tail(&pwrite->node, &mport->pwrites);
node              597 drivers/rapidio/rio.c 	list_for_each_entry(pwrite, &mport->pwrites, node) {
node              599 drivers/rapidio/rio.c 			list_del(&pwrite->node);
node             1211 drivers/rapidio/rio.c 	list_for_each_entry(pwrite, &mport->pwrites, node)
node             1941 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node) {
node             1984 drivers/rapidio/rio.c 	list_for_each_entry(scan, &rio_scans, node) {
node             2012 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node) {
node             2020 drivers/rapidio/rio.c 	list_add_tail(&scan->node, &rio_scans);
node             2052 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node)
node             2057 drivers/rapidio/rio.c 	list_for_each_entry(scan, &rio_scans, node) {
node             2059 drivers/rapidio/rio.c 			list_del(&scan->node);
node             2081 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node) {
node             2157 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node) {
node             2193 drivers/rapidio/rio.c 	list_for_each_entry(port, &rio_mports, node) {
node             2253 drivers/rapidio/rio.c 	list_for_each_entry(scan, &rio_scans, node) {
node             2262 drivers/rapidio/rio.c 	list_add_tail(&port->node, &rio_mports);
node             2321 drivers/rapidio/rio.c 	list_del(&port->node);
node              139 drivers/rapidio/rio_cm.c 	struct list_head node;
node              197 drivers/rapidio/rio_cm.c 	struct list_head node;
node              208 drivers/rapidio/rio_cm.c 	struct list_head node;
node              407 drivers/rapidio/rio_cm.c 	list_add_tail(&req->node, &ch->accept_queue);
node              675 drivers/rapidio/rio_cm.c 		list_for_each_entry_safe(req, _req, &cm->tx_reqs, node) {
node              676 drivers/rapidio/rio_cm.c 			list_del(&req->node);
node              718 drivers/rapidio/rio_cm.c 	list_add_tail(&treq->node, &cm->tx_reqs);
node             1140 drivers/rapidio/rio_cm.c 	req = list_first_entry(&ch->accept_queue, struct conn_req, node);
node             1141 drivers/rapidio/rio_cm.c 	list_del(&req->node);
node             1154 drivers/rapidio/rio_cm.c 	list_for_each_entry(peer, &new_ch->cmdev->peers, node) {
node             1596 drivers/rapidio/rio_cm.c 	list_for_each_entry(peer, &cm->peers, node) {
node             1811 drivers/rapidio/rio_cm.c 	list_for_each_entry(peer, &cm->peers, node) {
node             1973 drivers/rapidio/rio_cm.c 	list_add_tail(&peer->node, &cm->peers);
node             2020 drivers/rapidio/rio_cm.c 	list_for_each_entry(peer, &cm->peers, node) {
node             2024 drivers/rapidio/rio_cm.c 			list_del(&peer->node);
node             2230 drivers/rapidio/rio_cm.c 	list_for_each_entry_safe(peer, temp, &cm->peers, node) {
node             2232 drivers/rapidio/rio_cm.c 		list_del(&peer->node);
node             1459 drivers/regulator/core.c 	struct regulator_map *node;
node             1470 drivers/regulator/core.c 	list_for_each_entry(node, &regulator_map_list, list) {
node             1471 drivers/regulator/core.c 		if (node->dev_name && consumer_dev_name) {
node             1472 drivers/regulator/core.c 			if (strcmp(node->dev_name, consumer_dev_name) != 0)
node             1474 drivers/regulator/core.c 		} else if (node->dev_name || consumer_dev_name) {
node             1478 drivers/regulator/core.c 		if (strcmp(node->supply, supply) != 0)
node             1483 drivers/regulator/core.c 			 dev_name(&node->regulator->dev),
node             1484 drivers/regulator/core.c 			 node->regulator->desc->name,
node             1490 drivers/regulator/core.c 	node = kzalloc(sizeof(struct regulator_map), GFP_KERNEL);
node             1491 drivers/regulator/core.c 	if (node == NULL)
node             1494 drivers/regulator/core.c 	node->regulator = rdev;
node             1495 drivers/regulator/core.c 	node->supply = supply;
node             1498 drivers/regulator/core.c 		node->dev_name = kstrdup(consumer_dev_name, GFP_KERNEL);
node             1499 drivers/regulator/core.c 		if (node->dev_name == NULL) {
node             1500 drivers/regulator/core.c 			kfree(node);
node             1505 drivers/regulator/core.c 	list_add(&node->list, &regulator_map_list);
node             1511 drivers/regulator/core.c 	struct regulator_map *node, *n;
node             1513 drivers/regulator/core.c 	list_for_each_entry_safe(node, n, &regulator_map_list, list) {
node             1514 drivers/regulator/core.c 		if (rdev == node->regulator) {
node             1515 drivers/regulator/core.c 			list_del(&node->list);
node             1516 drivers/regulator/core.c 			kfree(node->dev_name);
node             1517 drivers/regulator/core.c 			kfree(node);
node             1716 drivers/regulator/core.c 	struct device_node *node;
node             1724 drivers/regulator/core.c 		node = of_get_regulator(dev, supply);
node             1725 drivers/regulator/core.c 		if (node) {
node             1726 drivers/regulator/core.c 			r = of_find_regulator_by_node(node);
node              667 drivers/regulator/da9063-regulator.c 	struct device_node *node;
node              674 drivers/regulator/da9063-regulator.c 	node = of_get_child_by_name(pdev->dev.parent->of_node, "regulators");
node              675 drivers/regulator/da9063-regulator.c 	if (!node) {
node              680 drivers/regulator/da9063-regulator.c 	num = of_regulator_match(&pdev->dev, node, da9063_matches,
node              682 drivers/regulator/da9063-regulator.c 	of_node_put(node);
node              256 drivers/regulator/da9211-regulator.c 	struct device_node *node;
node              259 drivers/regulator/da9211-regulator.c 	node = of_get_child_by_name(dev->of_node, "regulators");
node              260 drivers/regulator/da9211-regulator.c 	if (!node) {
node              265 drivers/regulator/da9211-regulator.c 	num = of_regulator_match(dev, node, da9211_matches,
node              267 drivers/regulator/da9211-regulator.c 	of_node_put(node);
node               61 drivers/regulator/internal.h 				 struct device_node **node);
node               81 drivers/regulator/internal.h 			   struct device_node **node)
node              231 drivers/regulator/mc13783-regulator.c #define MC13783_DEFINE(prefix, name, node, reg, vsel_reg, voltages)	\
node              232 drivers/regulator/mc13783-regulator.c 	MC13xxx_DEFINE(MC13783_REG_, name, node, reg, vsel_reg, voltages, \
node              235 drivers/regulator/mc13783-regulator.c #define MC13783_FIXED_DEFINE(prefix, name, node, reg, voltages)		\
node              236 drivers/regulator/mc13783-regulator.c 	MC13xxx_FIXED_DEFINE(MC13783_REG_, name, node, reg, voltages,	\
node              239 drivers/regulator/mc13783-regulator.c #define MC13783_GPO_DEFINE(prefix, name, node, reg, voltages)		\
node              240 drivers/regulator/mc13783-regulator.c 	MC13xxx_GPO_DEFINE(MC13783_REG_, name, node, reg, voltages,	\
node              425 drivers/regulator/mc13783-regulator.c 		struct device_node *node = NULL;
node              431 drivers/regulator/mc13783-regulator.c 			node = mc13xxx_data[i].node;
node              441 drivers/regulator/mc13783-regulator.c 		config.of_node = node;
node              249 drivers/regulator/mc13892-regulator.c #define MC13892_FIXED_DEFINE(name, node, reg, voltages)			\
node              250 drivers/regulator/mc13892-regulator.c 	MC13xxx_FIXED_DEFINE(MC13892_, name, node, reg, voltages,	\
node              253 drivers/regulator/mc13892-regulator.c #define MC13892_GPO_DEFINE(name, node, reg, voltages)			\
node              254 drivers/regulator/mc13892-regulator.c 	MC13xxx_GPO_DEFINE(MC13892_, name, node, reg, voltages,		\
node              257 drivers/regulator/mc13892-regulator.c #define MC13892_SW_DEFINE(name, node, reg, vsel_reg, voltages)		\
node              258 drivers/regulator/mc13892-regulator.c 	MC13xxx_DEFINE(MC13892_, name, node, reg, vsel_reg, voltages,	\
node              261 drivers/regulator/mc13892-regulator.c #define MC13892_DEFINE_REGU(name, node, reg, vsel_reg, voltages)	\
node              262 drivers/regulator/mc13892-regulator.c 	MC13xxx_DEFINE(MC13892_, name, node, reg, vsel_reg, voltages, \
node              595 drivers/regulator/mc13892-regulator.c 		struct device_node *node = NULL;
node              601 drivers/regulator/mc13892-regulator.c 			node = mc13xxx_data[i].node;
node              611 drivers/regulator/mc13892-regulator.c 		config.of_node = node;
node              195 drivers/regulator/mc13xxx-regulator-core.c 				p->node = child;
node              274 drivers/regulator/of_regulator.c 					  struct device_node *node,
node              279 drivers/regulator/of_regulator.c 	if (!node)
node              286 drivers/regulator/of_regulator.c 	if (of_get_regulation_constraints(dev, node, &init_data, desc))
node              325 drivers/regulator/of_regulator.c int of_regulator_match(struct device *dev, struct device_node *node,
node              335 drivers/regulator/of_regulator.c 	if (!dev || !node)
node              355 drivers/regulator/of_regulator.c 	for_each_child_of_node(node, child) {
node              433 drivers/regulator/of_regulator.c 					    struct device_node **node)
node              464 drivers/regulator/of_regulator.c 	*node = child;
node              488 drivers/regulator/of_regulator.c 	struct device_node *node = rdev->dev.of_node;
node              491 drivers/regulator/of_regulator.c 	n_phandles = of_count_phandle_with_args(node,
node              547 drivers/regulator/of_regulator.c 	struct device_node *node = rdev->dev.of_node;
node              564 drivers/regulator/of_regulator.c 		c_node = of_parse_phandle(node,
node              580 drivers/regulator/of_regulator.c 		if (!of_coupling_find_node(c_node, node, &index)) {
node              620 drivers/regulator/of_regulator.c 	struct device_node *node = rdev->dev.of_node;
node              624 drivers/regulator/of_regulator.c 	c_node = of_parse_phandle(node, "regulator-coupled-with", index);
node             1467 drivers/regulator/palmas-regulator.c 			      struct device_node *node,
node             1475 drivers/regulator/palmas-regulator.c 	regulators = of_get_child_by_name(node, "regulators");
node             1550 drivers/regulator/palmas-regulator.c 	pdata->ldo6_vibrator = of_property_read_bool(node, "ti,ldo6-vibrator");
node             1599 drivers/regulator/palmas-regulator.c 	struct device_node *node = pdev->dev.of_node;
node             1622 drivers/regulator/palmas-regulator.c 	if (of_device_is_compatible(node, "ti,tps659038-pmic")) {
node             1634 drivers/regulator/palmas-regulator.c 	ret = palmas_dt_to_pdata(&pdev->dev, node, pdata, driver_data);
node              414 drivers/regulator/qcom-rpmh-regulator.c 			struct device_node *node, const char *pmic_id,
node              427 drivers/regulator/qcom-rpmh-regulator.c 		if (of_node_name_eq(node, rpmh_data->name))
node              431 drivers/regulator/qcom-rpmh-regulator.c 		dev_err(dev, "Unknown regulator %pOFn\n", node);
node              441 drivers/regulator/qcom-rpmh-regulator.c 			node, rpmh_resource_name);
node              459 drivers/regulator/qcom-rpmh-regulator.c 	vreg->always_wait_for_ack = of_property_read_bool(node,
node              467 drivers/regulator/qcom-rpmh-regulator.c 	init_data = of_get_regulator_init_data(dev, node, &vreg->rdesc);
node              480 drivers/regulator/qcom-rpmh-regulator.c 	reg_config.of_node	= node;
node              487 drivers/regulator/qcom-rpmh-regulator.c 			node, ret);
node              492 drivers/regulator/qcom-rpmh-regulator.c 		node, rpmh_resource_name, vreg->addr);
node              885 drivers/regulator/qcom-rpmh-regulator.c 	struct device_node *node;
node              900 drivers/regulator/qcom-rpmh-regulator.c 	for_each_available_child_of_node(dev->of_node, node) {
node              903 drivers/regulator/qcom-rpmh-regulator.c 			of_node_put(node);
node              907 drivers/regulator/qcom-rpmh-regulator.c 		ret = rpmh_regulator_init_vreg(vreg, dev, node, pmic_id,
node              910 drivers/regulator/qcom-rpmh-regulator.c 			of_node_put(node);
node              654 drivers/regulator/qcom_rpm-regulator.c 				 struct device_node *node,
node              669 drivers/regulator/qcom_rpm-regulator.c 	ret = of_property_read_u32(node, key, &freq);
node              686 drivers/regulator/qcom_rpm-regulator.c static int rpm_reg_of_parse(struct device_node *node,
node              699 drivers/regulator/qcom_rpm-regulator.c 	if (of_property_read_bool(node, key)) {
node              708 drivers/regulator/qcom_rpm-regulator.c 		ret = rpm_reg_of_parse_freq(dev, node, vreg);
node              715 drivers/regulator/qcom_rpm-regulator.c 		pwm = !of_property_read_bool(node, key);
node              728 drivers/regulator/qcom_rpm-regulator.c 		ret = of_property_read_u32(node, key, &val);
node              374 drivers/regulator/qcom_spmi-regulator.c 	struct list_head			node;
node             1699 drivers/regulator/qcom_spmi-regulator.c 		struct device_node *node, struct spmi_regulator_init_data *data)
node             1710 drivers/regulator/qcom_spmi-regulator.c 	of_property_read_u32(node, "qcom,ocp-max-retries",
node             1712 drivers/regulator/qcom_spmi-regulator.c 	of_property_read_u32(node, "qcom,ocp-retry-delay",
node             1714 drivers/regulator/qcom_spmi-regulator.c 	of_property_read_u32(node, "qcom,pin-ctrl-enable",
node             1716 drivers/regulator/qcom_spmi-regulator.c 	of_property_read_u32(node, "qcom,pin-ctrl-hpm", &data->pin_ctrl_hpm);
node             1717 drivers/regulator/qcom_spmi-regulator.c 	of_property_read_u32(node, "qcom,vs-soft-start-strength",
node             1731 drivers/regulator/qcom_spmi-regulator.c static int spmi_regulator_of_parse(struct device_node *node,
node             1740 drivers/regulator/qcom_spmi-regulator.c 	spmi_regulator_get_dt_config(vreg, node, &data);
node             1966 drivers/regulator/qcom_spmi-regulator.c 	struct device_node *node = pdev->dev.of_node;
node             1986 drivers/regulator/qcom_spmi-regulator.c 	if (of_find_property(node, "qcom,saw-reg", &lenp)) {
node             1987 drivers/regulator/qcom_spmi-regulator.c 		syscon = of_parse_phandle(node, "qcom,saw-reg", 0);
node             1997 drivers/regulator/qcom_spmi-regulator.c 			reg_node = of_get_child_by_name(node, reg->name);
node             2036 drivers/regulator/qcom_spmi-regulator.c 			reg_node = of_get_child_by_name(node, reg->name);
node             2064 drivers/regulator/qcom_spmi-regulator.c 		INIT_LIST_HEAD(&vreg->node);
node             2065 drivers/regulator/qcom_spmi-regulator.c 		list_add(&vreg->node, vreg_list);
node             2071 drivers/regulator/qcom_spmi-regulator.c 	list_for_each_entry(vreg, vreg_list, node)
node             2082 drivers/regulator/qcom_spmi-regulator.c 	list_for_each_entry(vreg, vreg_list, node)
node              165 drivers/regulator/tps65086-regulator.c static int tps65086_of_parse_cb(struct device_node *node,
node              172 drivers/regulator/tps65086-regulator.c 	if (of_property_read_bool(node, "ti,regulator-step-size-25mv")) {
node              196 drivers/regulator/tps65086-regulator.c 	if (desc->id <= BUCK6 && of_property_read_bool(node, "ti,regulator-decay")) {
node              285 drivers/remoteproc/imx_rproc.c 		struct device_node *node;
node              288 drivers/remoteproc/imx_rproc.c 		node = of_parse_phandle(np, "memory-region", a);
node              289 drivers/remoteproc/imx_rproc.c 		err = of_address_to_resource(node, 0, &res);
node               32 drivers/remoteproc/qcom_common.c 	glink->edge = qcom_glink_smem_register(glink->dev, glink->node);
node               54 drivers/remoteproc/qcom_common.c 	glink->node = of_get_child_by_name(dev->parent->of_node, "glink-edge");
node               55 drivers/remoteproc/qcom_common.c 	if (!glink->node)
node               73 drivers/remoteproc/qcom_common.c 	if (!glink->node)
node               77 drivers/remoteproc/qcom_common.c 	of_node_put(glink->node);
node              128 drivers/remoteproc/qcom_common.c 	smd->edge = qcom_smd_register_edge(smd->dev, smd->node);
node              150 drivers/remoteproc/qcom_common.c 	smd->node = of_get_child_by_name(dev->parent->of_node, "smd-edge");
node              151 drivers/remoteproc/qcom_common.c 	if (!smd->node)
node              169 drivers/remoteproc/qcom_common.c 	if (!smd->node)
node              173 drivers/remoteproc/qcom_common.c 	of_node_put(smd->node);
node               15 drivers/remoteproc/qcom_common.h 	struct device_node *node;
node               23 drivers/remoteproc/qcom_common.h 	struct device_node *node;
node              382 drivers/remoteproc/qcom_q6v5_adsp.c 	struct device_node *node;
node              386 drivers/remoteproc/qcom_q6v5_adsp.c 	node = of_parse_phandle(adsp->dev->of_node, "memory-region", 0);
node              387 drivers/remoteproc/qcom_q6v5_adsp.c 	if (!node) {
node              392 drivers/remoteproc/qcom_q6v5_adsp.c 	ret = of_address_to_resource(node, 0, &r);
node             1363 drivers/remoteproc/qcom_q6v5_mss.c 	struct device_node *node;
node             1368 drivers/remoteproc/qcom_q6v5_mss.c 	node = of_parse_phandle(child, "memory-region", 0);
node             1369 drivers/remoteproc/qcom_q6v5_mss.c 	ret = of_address_to_resource(node, 0, &r);
node             1374 drivers/remoteproc/qcom_q6v5_mss.c 	of_node_put(node);
node             1386 drivers/remoteproc/qcom_q6v5_mss.c 	node = of_parse_phandle(child, "memory-region", 0);
node             1387 drivers/remoteproc/qcom_q6v5_mss.c 	ret = of_address_to_resource(node, 0, &r);
node             1392 drivers/remoteproc/qcom_q6v5_mss.c 	of_node_put(node);
node              222 drivers/remoteproc/qcom_q6v5_pas.c 	struct device_node *node;
node              226 drivers/remoteproc/qcom_q6v5_pas.c 	node = of_parse_phandle(adsp->dev->of_node, "memory-region", 0);
node              227 drivers/remoteproc/qcom_q6v5_pas.c 	if (!node) {
node              232 drivers/remoteproc/qcom_q6v5_pas.c 	ret = of_address_to_resource(node, 0, &r);
node              502 drivers/remoteproc/qcom_q6v5_wcss.c 	struct device_node *node;
node              505 drivers/remoteproc/qcom_q6v5_wcss.c 	node = of_parse_phandle(dev->of_node, "memory-region", 0);
node              506 drivers/remoteproc/qcom_q6v5_wcss.c 	if (node)
node              507 drivers/remoteproc/qcom_q6v5_wcss.c 		rmem = of_reserved_mem_lookup(node);
node              508 drivers/remoteproc/qcom_q6v5_wcss.c 	of_node_put(node);
node               25 drivers/remoteproc/qcom_sysmon.c 	struct list_head node;
node              402 drivers/remoteproc/qcom_sysmon.c 	sysmon->ssctl.sq_node = svc->node;
node              555 drivers/remoteproc/qcom_sysmon.c 	list_add(&sysmon->node, &sysmon_list);
node              572 drivers/remoteproc/qcom_sysmon.c 	list_del(&sysmon->node);
node              606 drivers/remoteproc/qcom_sysmon.c 	list_for_each_entry(sysmon, &sysmon_list, node) {
node              432 drivers/remoteproc/qcom_wcnss.c 	struct device_node *node;
node              436 drivers/remoteproc/qcom_wcnss.c 	node = of_parse_phandle(wcnss->dev->of_node, "memory-region", 0);
node              437 drivers/remoteproc/qcom_wcnss.c 	if (!node) {
node              442 drivers/remoteproc/qcom_wcnss.c 	ret = of_address_to_resource(node, 0, &r);
node              201 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(carveout, &rproc->carveouts, node) {
node              257 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(carveout, &rproc->carveouts, node) {
node              553 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&rvdev->node, &rproc->rvdevs);
node              583 drivers/remoteproc/remoteproc_core.c 	list_del(&rvdev->node);
node              642 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&trace->node, &rproc->traces);
node              718 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&mapping->node, &rproc->mappings);
node              809 drivers/remoteproc/remoteproc_core.c 		list_add_tail(&mapping->node, &rproc->mappings);
node              941 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&mem->node, &rproc->carveouts);
node             1098 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(subdev, &rproc->subdevs, node) {
node             1109 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_continue_reverse(subdev, &rproc->subdevs, node) {
node             1122 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(subdev, &rproc->subdevs, node) {
node             1133 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_continue_reverse(subdev, &rproc->subdevs, node) {
node             1145 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_reverse(subdev, &rproc->subdevs, node) {
node             1155 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_reverse(subdev, &rproc->subdevs, node) {
node             1180 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) {
node             1239 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->dump_segments, node) {
node             1240 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
node             1260 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(trace, ttmp, &rproc->traces, node) {
node             1263 drivers/remoteproc/remoteproc_core.c 		list_del(&trace->node);
node             1268 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->mappings, node) {
node             1278 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
node             1283 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) {
node             1286 drivers/remoteproc/remoteproc_core.c 		list_del(&entry->node);
node             1291 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry_safe(rvdev, rvtmp, &rproc->rvdevs, node)
node             1514 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&segment->node, &rproc->dump_segments);
node             1552 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&segment->node, &rproc->dump_segments);
node             1580 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(segment, &rproc->dump_segments, node) {
node             1609 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(segment, &rproc->dump_segments, node) {
node             1859 drivers/remoteproc/remoteproc_core.c 	list_for_each_entry(r, &rproc_list, node) {
node             1929 drivers/remoteproc/remoteproc_core.c 	list_add(&rproc->node, &rproc_list);
node             2144 drivers/remoteproc/remoteproc_core.c 	list_del(&rproc->node);
node             2162 drivers/remoteproc/remoteproc_core.c 	list_add_tail(&subdev->node, &rproc->subdevs);
node             2173 drivers/remoteproc/remoteproc_core.c 	list_del(&subdev->node);
node              290 drivers/remoteproc/remoteproc_debugfs.c 	list_for_each_entry(carveout, &rproc->carveouts, node) {
node               23 drivers/remoteproc/remoteproc_internal.h 	struct list_head node;
node              130 drivers/remoteproc/st_remoteproc.c 		rmem = of_reserved_mem_lookup(it.node);
node              137 drivers/remoteproc/st_remoteproc.c 		if (strcmp(it.node->name, "vdev0buffer")) {
node              144 drivers/remoteproc/st_remoteproc.c 						   it.node->name);
node              150 drivers/remoteproc/st_remoteproc.c 							   it.node->name);
node              212 drivers/remoteproc/stm32_rproc.c 		rmem = of_reserved_mem_lookup(it.node);
node              225 drivers/remoteproc/stm32_rproc.c 		if (strcmp(it.node->name, "vdev0buffer")) {
node              232 drivers/remoteproc/stm32_rproc.c 						   it.node->name);
node              242 drivers/remoteproc/stm32_rproc.c 							   it.node->name);
node              368 drivers/remoteproc/stm32_rproc.c 	list_for_each_entry(trace, &rproc->traces, node) {
node              371 drivers/remoteproc/stm32_rproc.c 		list_for_each_entry(segment, &rproc->dump_segments, node) {
node              600 drivers/reset/core.c struct reset_control *__of_reset_control_get(struct device_node *node,
node              610 drivers/reset/core.c 	if (!node)
node              614 drivers/reset/core.c 		index = of_property_match_string(node,
node              622 drivers/reset/core.c 	ret = of_parse_phandle_with_args(node, "resets", "#reset-cells",
node              840 drivers/reset/core.c static int of_reset_control_get_count(struct device_node *node)
node              844 drivers/reset/core.c 	if (!node)
node              847 drivers/reset/core.c 	count = of_count_phandle_with_args(node, "resets", "#reset-cells");
node               49 drivers/rpmsg/qcom_glink_native.c 	struct list_head node;
node               75 drivers/rpmsg/qcom_glink_native.c 	struct list_head node;
node              254 drivers/rpmsg/qcom_glink_native.c 	list_for_each_entry_safe(intent, tmp, &channel->done_intents, node) {
node              491 drivers/rpmsg/qcom_glink_native.c 	list_for_each_entry_safe(intent, tmp, &channel->done_intents, node) {
node              492 drivers/rpmsg/qcom_glink_native.c 		list_del(&intent->node);
node              531 drivers/rpmsg/qcom_glink_native.c 	list_add_tail(&intent->node, &channel->done_intents);
node              772 drivers/rpmsg/qcom_glink_native.c 	INIT_LIST_HEAD(&dcmd->node);
node              777 drivers/rpmsg/qcom_glink_native.c 	list_add_tail(&dcmd->node, &glink->rx_queue);
node             1338 drivers/rpmsg/qcom_glink_native.c static struct device_node *qcom_glink_match_channel(struct device_node *node,
node             1346 drivers/rpmsg/qcom_glink_native.c 	for_each_available_child_of_node(node, child) {
node             1385 drivers/rpmsg/qcom_glink_native.c 	struct device_node *node;
node             1431 drivers/rpmsg/qcom_glink_native.c 		node = qcom_glink_match_channel(glink->dev->of_node, name);
node             1432 drivers/rpmsg/qcom_glink_native.c 		rpdev->dev.of_node = node;
node             1528 drivers/rpmsg/qcom_glink_native.c 					struct glink_defer_cmd, node);
node             1529 drivers/rpmsg/qcom_glink_native.c 		list_del(&dcmd->node);
node             1573 drivers/rpmsg/qcom_glink_native.c 	list_for_each_entry_safe(dcmd, tmp, &glink->rx_queue, node)
node              186 drivers/rpmsg/qcom_glink_smem.c 					    struct device_node *node)
node              202 drivers/rpmsg/qcom_glink_smem.c 	dev->of_node = node;
node              204 drivers/rpmsg/qcom_glink_smem.c 	dev_set_name(dev, "%s:%pOFn", dev_name(parent->parent), node);
node             1321 drivers/rpmsg/qcom_smd.c 			       struct device_node *node,
node             1335 drivers/rpmsg/qcom_smd.c 	edge->of_node = of_node_get(node);
node             1338 drivers/rpmsg/qcom_smd.c 	ret = of_property_read_u32(node, key, &edge->edge_id);
node             1346 drivers/rpmsg/qcom_smd.c 	of_property_read_u32(node, key, &edge->remote_pid);
node             1357 drivers/rpmsg/qcom_smd.c 		syscon_np = of_parse_phandle(node, "qcom,ipc", 0);
node             1368 drivers/rpmsg/qcom_smd.c 		ret = of_property_read_u32_index(node, key, 1, &edge->ipc_offset);
node             1374 drivers/rpmsg/qcom_smd.c 		ret = of_property_read_u32_index(node, key, 2, &edge->ipc_bit);
node             1381 drivers/rpmsg/qcom_smd.c 	ret = of_property_read_string(node, "label", &edge->name);
node             1383 drivers/rpmsg/qcom_smd.c 		edge->name = node->name;
node             1385 drivers/rpmsg/qcom_smd.c 	irq = irq_of_parse_and_map(node, 0);
node             1393 drivers/rpmsg/qcom_smd.c 			       node->name, edge);
node             1445 drivers/rpmsg/qcom_smd.c 					     struct device_node *node)
node             1458 drivers/rpmsg/qcom_smd.c 	edge->dev.of_node = node;
node             1460 drivers/rpmsg/qcom_smd.c 	dev_set_name(&edge->dev, "%s:%pOFn", dev_name(parent), node);
node             1468 drivers/rpmsg/qcom_smd.c 	ret = qcom_smd_parse_edge(&edge->dev, node, edge);
node             1525 drivers/rpmsg/qcom_smd.c 	struct device_node *node;
node             1533 drivers/rpmsg/qcom_smd.c 	for_each_available_child_of_node(pdev->dev.of_node, node)
node             1534 drivers/rpmsg/qcom_smd.c 		qcom_smd_register_edge(&pdev->dev, node);
node              400 drivers/rtc/interface.c 		alarm->time = rtc_ktime_to_tm(rtc->aie_timer.node.expires);
node              471 drivers/rtc/interface.c 	rtc->aie_timer.node.expires = rtc_tm_to_ktime(alarm->time);
node              500 drivers/rtc/interface.c 	rtc->aie_timer.node.expires = rtc_tm_to_ktime(alarm->time);
node              505 drivers/rtc/interface.c 			 rtc->aie_timer.node.expires)) {
node              507 drivers/rtc/interface.c 		timerqueue_add(&rtc->timerqueue, &rtc->aie_timer.node);
node              576 drivers/rtc/interface.c 		rtc->uie_rtctimer.node.expires = ktime_add(now, onesec);
node              808 drivers/rtc/interface.c 	timerqueue_add(&rtc->timerqueue, &timer->node);
node              810 drivers/rtc/interface.c 	if (!next || ktime_before(timer->node.expires, next->expires)) {
node              814 drivers/rtc/interface.c 		alarm.time = rtc_ktime_to_tm(timer->node.expires);
node              821 drivers/rtc/interface.c 			timerqueue_del(&rtc->timerqueue, &timer->node);
node              855 drivers/rtc/interface.c 	timerqueue_del(&rtc->timerqueue, &timer->node);
node              858 drivers/rtc/interface.c 	if (next == &timer->node) {
node              906 drivers/rtc/interface.c 		timer = container_of(next, struct rtc_timer, node);
node              907 drivers/rtc/interface.c 		timerqueue_del(&rtc->timerqueue, &timer->node);
node              916 drivers/rtc/interface.c 			timer->node.expires = ktime_add(timer->node.expires,
node              919 drivers/rtc/interface.c 			timerqueue_add(&rtc->timerqueue, &timer->node);
node              940 drivers/rtc/interface.c 			timer = container_of(next, struct rtc_timer, node);
node              941 drivers/rtc/interface.c 			timerqueue_del(&rtc->timerqueue, &timer->node);
node              965 drivers/rtc/interface.c 	timerqueue_init(&timer->node);
node              988 drivers/rtc/interface.c 	timer->node.expires = expires;
node              244 drivers/rtc/rtc-88pm80x.c 	struct device_node *node = pdev->dev.of_node;
node              247 drivers/rtc/rtc-88pm80x.c 	if (!pdata && !node) {
node              189 drivers/rtc/rtc-ab-eoz9.c static int abeoz9_trickle_parse_dt(struct device_node *node)
node              193 drivers/rtc/rtc-ab-eoz9.c 	if (of_property_read_u32(node, "trickle-resistor-ohms", &ohms))
node              210 drivers/rtc/rtc-ab-eoz9.c static int abeoz9_rtc_setup(struct device *dev, struct device_node *node)
node              243 drivers/rtc/rtc-ab-eoz9.c 	ret = abeoz9_trickle_parse_dt(node);
node              142 drivers/rtc/rtc-bq32k.c static int trickle_charger_of_init(struct device *dev, struct device_node *node)
node              148 drivers/rtc/rtc-bq32k.c 	if (of_property_read_u32(node, "trickle-resistor-ohms" , &ohms))
node              158 drivers/rtc/rtc-bq32k.c 		if (of_property_read_bool(node, "trickle-diode-disable")) {
node              168 drivers/rtc/rtc-bq32k.c 		if (!of_property_read_bool(node, "trickle-diode-disable")) {
node             1374 drivers/rtc/rtc-cmos.c 	struct device_node *node = pdev->dev.of_node;
node             1377 drivers/rtc/rtc-cmos.c 	if (!node)
node             1380 drivers/rtc/rtc-cmos.c 	val = of_get_property(node, "ctrl-reg", NULL);
node             1384 drivers/rtc/rtc-cmos.c 	val = of_get_property(node, "freq-reg", NULL);
node             1505 drivers/rtc/rtc-ds1307.c 	struct device_node *node = ds1307->dev->of_node;
node             1530 drivers/rtc/rtc-ds1307.c 		of_property_read_string_index(node, "clock-output-names", i,
node             1540 drivers/rtc/rtc-ds1307.c 	if (!node)
node             1543 drivers/rtc/rtc-ds1307.c 	of_clk_add_provider(node, of_clk_src_onecell_get, onecell);
node              395 drivers/rtc/rtc-hym8563.c 	struct device_node *node = client->dev.of_node;
node              413 drivers/rtc/rtc-hym8563.c 	of_property_read_string(node, "clock-output-names", &init.name);
node              419 drivers/rtc/rtc-hym8563.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              547 drivers/rtc/rtc-m41t80.c 	struct device_node *node = client->dev.of_node;
node              570 drivers/rtc/rtc-m41t80.c 	of_property_read_string(node, "clock-output-names", &init.name);
node              575 drivers/rtc/rtc-m41t80.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              506 drivers/rtc/rtc-pcf8563.c 	struct device_node *node = client->dev.of_node;
node              526 drivers/rtc/rtc-pcf8563.c 	of_property_read_string(node, "clock-output-names", &init.name);
node              532 drivers/rtc/rtc-pcf8563.c 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
node              491 drivers/rtc/rtc-rv8803.c 	struct device_node *node = client->dev.of_node;
node              495 drivers/rtc/rtc-rv8803.c 	if (!node)
node              507 drivers/rtc/rtc-rv8803.c 	if (of_property_read_bool(node, "epson,vdet-disable"))
node              510 drivers/rtc/rtc-rv8803.c 	if (of_property_read_bool(node, "trickle-diode-disable"))
node              453 drivers/rtc/rtc-sc27xx.c 		rtc_ktime_to_tm(rtc->rtc->aie_timer.node.expires);
node              598 drivers/rtc/rtc-sc27xx.c 	struct device_node *node = pdev->dev.of_node;
node              610 drivers/rtc/rtc-sc27xx.c 	ret = of_property_read_u32(node, "reg", &rtc->base);
node              214 drivers/rtc/rtc-sun6i.c static void __init sun6i_rtc_clk_init(struct device_node *node,
node              241 drivers/rtc/rtc-sun6i.c 	rtc->base = of_io_request_and_map(node, 0, of_node_full_name(node));
node              264 drivers/rtc/rtc-sun6i.c 	if (!of_get_property(node, "clocks", NULL))
node              269 drivers/rtc/rtc-sun6i.c 		of_property_read_string_index(node, "clock-output-names", 2,
node              283 drivers/rtc/rtc-sun6i.c 	parents[1] = of_clk_get_parent_name(node, 0);
node              288 drivers/rtc/rtc-sun6i.c 	init.num_parents = of_clk_get_parent_count(node) + 1;
node              289 drivers/rtc/rtc-sun6i.c 	of_property_read_string_index(node, "clock-output-names", 0,
node              298 drivers/rtc/rtc-sun6i.c 	of_property_read_string_index(node, "clock-output-names", 1,
node              316 drivers/rtc/rtc-sun6i.c 	of_clk_add_hw_provider(node, of_clk_hw_onecell_get, clk_data);
node              328 drivers/rtc/rtc-sun6i.c static void __init sun6i_a31_rtc_clk_init(struct device_node *node)
node              330 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun6i_a31_rtc_data);
node              341 drivers/rtc/rtc-sun6i.c static void __init sun8i_a23_rtc_clk_init(struct device_node *node)
node              343 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun8i_a23_rtc_data);
node              356 drivers/rtc/rtc-sun6i.c static void __init sun8i_h3_rtc_clk_init(struct device_node *node)
node              358 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun8i_h3_rtc_data);
node              376 drivers/rtc/rtc-sun6i.c static void __init sun50i_h6_rtc_clk_init(struct device_node *node)
node              378 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun50i_h6_rtc_data);
node              392 drivers/rtc/rtc-sun6i.c static void __init sun8i_r40_rtc_clk_init(struct device_node *node)
node              394 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun8i_r40_rtc_data);
node              404 drivers/rtc/rtc-sun6i.c static void __init sun8i_v3_rtc_clk_init(struct device_node *node)
node              406 drivers/rtc/rtc-sun6i.c 	sun6i_rtc_clk_init(node, &sun8i_v3_rtc_data);
node              182 drivers/s390/char/tape.h 	struct list_head		node;
node              339 drivers/s390/char/tape_core.c 	list_for_each_entry(tmp, &tape_device_list, node) {
node              349 drivers/s390/char/tape_core.c 	list_add_tail(&device->node, &tmp->node);
node              359 drivers/s390/char/tape_core.c 	list_del_init(&device->node);
node              543 drivers/s390/char/tape_core.c 	INIT_LIST_HEAD(&device->node);
node              600 drivers/s390/char/tape_core.c 	list_for_each_entry(tmp, &tape_device_list, node) {
node              440 drivers/s390/cio/cmf.c 	struct ccw_device_private *node;
node              455 drivers/s390/cio/cmf.c 	list_for_each_entry(node, &cmb_area.list, cmb_list) {
node              457 drivers/s390/cio/cmf.c 		data = node->cmb;
node              468 drivers/s390/cio/cmf.c 	list_add_tail(&cdev->private->cmb_list, &node->cmb_list);
node              343 drivers/s390/crypto/vfio_ap_ops.c 	list_add(&matrix_mdev->node, &matrix_dev->mdev_list);
node              358 drivers/s390/crypto/vfio_ap_ops.c 	list_del(&matrix_mdev->node);
node              542 drivers/s390/crypto/vfio_ap_ops.c 	list_for_each_entry(lstdev, &matrix_dev->mdev_list, node) {
node             1042 drivers/s390/crypto/vfio_ap_ops.c 	list_for_each_entry(m, &matrix_dev->mdev_list, node) {
node               82 drivers/s390/crypto/vfio_ap_private.h 	struct list_head node;
node              130 drivers/s390/virtio/virtio_ccw.c 	struct list_head node;
node              302 drivers/s390/virtio/virtio_ccw.c 	list_for_each_entry(info, &vcdev->virtqueues, node)
node              456 drivers/s390/virtio/virtio_ccw.c 	list_del(&info->node);
node              587 drivers/s390/virtio/virtio_ccw.c 	list_add(&info->node, &vcdev->virtqueues);
node             1057 drivers/s390/virtio/virtio_ccw.c 	list_for_each_entry(info, &vcdev->virtqueues, node) {
node             1857 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 	symbol_node_t *node;
node             1892 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 		for(node = expression->referenced_syms.slh_first;
node             1893 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 		    node != NULL;
node             1894 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 		    node = node->links.sle_next) {
node             1895 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 			if ((node->symbol->type == MASK
node             1896 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 			  || node->symbol->type == FIELD
node             1897 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 			  || node->symbol->type == ENUM
node             1898 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 			  || node->symbol->type == ENUM_ENTRY)
node             1899 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 			 && symlist_search(&node->symbol->info.finfo->symrefs,
node             1904 drivers/scsi/aic7xxx/aicasm/aicasm_gram.y 					 node->symbol->name, symbol->name);
node              314 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c 	symbol_node_t *node;
node              317 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c 	while((node = SLIST_FIRST(symlist_src2)) != NULL) {
node              319 drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c 		SLIST_INSERT_HEAD(symlist_dest, node, links);
node             3948 drivers/scsi/bfa/bfa_svc.c bfa_fcport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node)
node             3951 drivers/scsi/bfa/bfa_svc.c 	if (node)
node              541 drivers/scsi/bfa/bfa_svc.h wwn_t bfa_fcport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node);
node               64 drivers/scsi/device_handler/scsi_dh_alua.c 	struct list_head	node;
node               84 drivers/scsi/device_handler/scsi_dh_alua.c 	struct list_head	node;
node              116 drivers/scsi/device_handler/scsi_dh_alua.c 	list_del(&pg->node);
node              186 drivers/scsi/device_handler/scsi_dh_alua.c 	list_for_each_entry(pg, &port_group_list, node) {
node              241 drivers/scsi/device_handler/scsi_dh_alua.c 	INIT_LIST_HEAD(&pg->node);
node              254 drivers/scsi/device_handler/scsi_dh_alua.c 	list_add(&pg->node, &port_group_list);
node              363 drivers/scsi/device_handler/scsi_dh_alua.c 			list_del_rcu(&h->node);
node              372 drivers/scsi/device_handler/scsi_dh_alua.c 		list_add_rcu(&h->node, &pg->dh_list);
node              660 drivers/scsi/device_handler/scsi_dh_alua.c 						&tmp_pg->dh_list, node) {
node              707 drivers/scsi/device_handler/scsi_dh_alua.c 			list_for_each_entry_rcu(h, &pg->dh_list, node) {
node             1124 drivers/scsi/device_handler/scsi_dh_alua.c 	INIT_LIST_HEAD(&h->node);
node             1154 drivers/scsi/device_handler/scsi_dh_alua.c 		list_del_rcu(&h->node);
node              155 drivers/scsi/device_handler/scsi_dh_rdac.c 	struct list_head	node; /* list of all controllers */
node              185 drivers/scsi/device_handler/scsi_dh_rdac.c 	struct list_head	node;
node              322 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_del(&ctlr->node);
node              331 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_for_each_entry(tmp, &ctlr_list, node) {
node              356 drivers/scsi/device_handler/scsi_dh_rdac.c 	list_add(&ctlr->node, &ctlr_list);
node              425 drivers/scsi/device_handler/scsi_dh_rdac.c 		list_for_each_entry_rcu(tmp, &h->ctlr->dh_list, node) {
node              456 drivers/scsi/device_handler/scsi_dh_rdac.c 			list_add_rcu(&h->node, &h->ctlr->dh_list);
node              780 drivers/scsi/device_handler/scsi_dh_rdac.c 		list_del_rcu(&h->node);
node              552 drivers/scsi/isci/init.c 		INIT_LIST_HEAD(&idev->node);
node              260 drivers/scsi/isci/port.c 					    node) {
node             1053 drivers/scsi/isci/remote_device.c 	list_del_init(&idev->node);
node             1527 drivers/scsi/isci/remote_device.c 	if (WARN_ONCE(!list_empty(&idev->node), "found non-idle remote device\n"))
node             1631 drivers/scsi/isci/remote_device.c 	INIT_LIST_HEAD(&isci_device->node);
node             1636 drivers/scsi/isci/remote_device.c 	list_add_tail(&isci_device->node, &isci_port->remote_dev_list);
node               95 drivers/scsi/isci/remote_device.h 	struct list_head node;
node               54 drivers/scsi/lpfc/lpfc_bsg.c 	struct list_head node;
node              114 drivers/scsi/lpfc/lpfc_bsg.c 	struct list_head node;
node              807 drivers/scsi/lpfc/lpfc_bsg.c 	list_del(&evt->node);
node              810 drivers/scsi/lpfc/lpfc_bsg.c 		ed = list_entry(evt->events_to_get.next, typeof(*ed), node);
node              811 drivers/scsi/lpfc/lpfc_bsg.c 		list_del(&ed->node);
node              817 drivers/scsi/lpfc/lpfc_bsg.c 		ed = list_entry(evt->events_to_see.next, typeof(*ed), node);
node              818 drivers/scsi/lpfc/lpfc_bsg.c 		list_del(&ed->node);
node              966 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_entry(evt, &phba->ct_ev_waiters, node) {
node             1123 drivers/scsi/lpfc/lpfc_bsg.c 		list_add(&evt_dat->node, &evt->events_to_see);
node             1232 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_entry(evt, &phba->ct_ev_waiters, node) {
node             1242 drivers/scsi/lpfc/lpfc_bsg.c 	if (&evt->node == &phba->ct_ev_waiters) {
node             1265 drivers/scsi/lpfc/lpfc_bsg.c 		list_add(&evt->node, &phba->ct_ev_waiters);
node             1317 drivers/scsi/lpfc/lpfc_bsg.c 	list_for_each_entry_safe(evt, evt_next, &phba->ct_ev_waiters, node) {
node             1324 drivers/scsi/lpfc/lpfc_bsg.c 					     struct event_data, node);
node             1325 drivers/scsi/lpfc/lpfc_bsg.c 			list_del(&evt_dat->node);
node             2670 drivers/scsi/lpfc/lpfc_bsg.c 	list_add(&evt->node, &phba->ct_ev_waiters);
node             2761 drivers/scsi/lpfc/lpfc_bsg.c 				     node))->immed_dat;
node             3220 drivers/scsi/lpfc/lpfc_bsg.c 	list_add(&evt->node, &phba->ct_ev_waiters);
node             3339 drivers/scsi/lpfc/lpfc_bsg.c 				   typeof(*evdat), node);
node             11089 drivers/scsi/lpfc/lpfc_init.c static int lpfc_cpu_offline(unsigned int cpu, struct hlist_node *node)
node             11091 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp);
node             11115 drivers/scsi/lpfc/lpfc_init.c static int lpfc_cpu_online(unsigned int cpu, struct hlist_node *node)
node             11117 drivers/scsi/lpfc/lpfc_init.c 	struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp);
node              412 drivers/scsi/mac53c94.c 	struct device_node *node = macio_get_of_node(mdev);
node              452 drivers/scsi/mac53c94.c 		printk(KERN_ERR "mac53c94: ioremap failed for %pOF\n", node);
node              456 drivers/scsi/mac53c94.c 	clkprop = of_get_property(node, "clock-frequency", &proplen);
node              459 drivers/scsi/mac53c94.c        		       "assuming 25MHz\n", node);
node              473 drivers/scsi/mac53c94.c        		       "command space for %pOF\n", node);
node              486 drivers/scsi/mac53c94.c 		       state->intr, node);
node             1574 drivers/scsi/qla2xxx/qla_target.c 	void *node;
node             1602 drivers/scsi/qla2xxx/qla_target.c 	btree_for_each_safe64(&tgt->lun_qpair_map, key, node)
node             4167 drivers/scsi/qla2xxx/qla_target.c 	void *node;
node             4176 drivers/scsi/qla2xxx/qla_target.c 	btree_for_each_safe64(&tgt->lun_qpair_map, key, node)
node             6446 drivers/scsi/qla2xxx/qla_target.c 	struct scsi_qla_host *node;
node             6449 drivers/scsi/qla2xxx/qla_target.c 	btree_for_each_safe32(&ha->tgt.host_map, key, node)
node              768 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	void *node;
node              772 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	node = btree_remove32(&lport->lport_fcport_map, nacl->nport_id);
node              773 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	if (WARN_ON(node && (node != se_nacl))) {
node              781 drivers/scsi/qla2xxx/tcm_qla2xxx.c 			       node, GFP_ATOMIC);
node             1652 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	struct se_node_acl *node;
node             1666 drivers/scsi/qla2xxx/tcm_qla2xxx.c 	btree_for_each_safe32(&lport->lport_fcport_map, key, node)
node               33 drivers/scsi/raid_class.c 	struct list_head node;
node              101 drivers/scsi/raid_class.c 	list_for_each_entry_safe(rc, next, &rd->component_list, node) {
node              102 drivers/scsi/raid_class.c 		list_del(&rc->node);
node              235 drivers/scsi/raid_class.c 	INIT_LIST_HEAD(&rc->node);
node              242 drivers/scsi/raid_class.c 	list_add_tail(&rc->node, &rd->component_list);
node              251 drivers/scsi/raid_class.c 	list_del(&rc->node);
node               30 drivers/scsi/scsi_devinfo.c 	struct list_head node;	/* our node for being on the master list */
node              270 drivers/scsi/scsi_devinfo.c 	list_for_each_entry(devinfo_table, &scsi_dev_info_list, node)
node              627 drivers/scsi/scsi_devinfo.c 		list_entry(dl->top, struct scsi_dev_info_list_table, node);
node              652 drivers/scsi/scsi_devinfo.c 				   node);
node              666 drivers/scsi/scsi_devinfo.c 		list_entry(dl->top, struct scsi_dev_info_list_table, node);
node              678 drivers/scsi/scsi_devinfo.c 					   node);
node              793 drivers/scsi/scsi_devinfo.c 	INIT_LIST_HEAD(&devinfo_table->node);
node              797 drivers/scsi/scsi_devinfo.c 	list_add_tail(&devinfo_table->node, &scsi_dev_info_list);
node              822 drivers/scsi/scsi_devinfo.c 	list_del(&devinfo_table->node);
node             2436 drivers/scsi/scsi_lib.c 			evt = list_entry(this, struct scsi_event, node);
node             2437 drivers/scsi/scsi_lib.c 			list_del(&evt->node);
node             2466 drivers/scsi/scsi_lib.c 	list_add_tail(&evt->node, &sdev->event_list);
node             2487 drivers/scsi/scsi_lib.c 	INIT_LIST_HEAD(&evt->node);
node              459 drivers/scsi/scsi_sysfs.c 		evt = list_entry(this, struct scsi_event, node);
node              460 drivers/scsi/scsi_sysfs.c 		list_del(&evt->node);
node             7863 drivers/scsi/smartpqi/smartpqi_init.c 	int node, cp_node;
node             7880 drivers/scsi/smartpqi/smartpqi_init.c 	node = dev_to_node(&pci_dev->dev);
node             7881 drivers/scsi/smartpqi/smartpqi_init.c 	if (node == NUMA_NO_NODE) {
node             7888 drivers/scsi/smartpqi/smartpqi_init.c 	ctrl_info = pqi_alloc_ctrl_info(node);
node               79 drivers/scsi/virtio_scsi.c 	struct hlist_node node;
node              421 drivers/sh/clk/core.c 	if (clk->node.next || clk->node.prev)
node              438 drivers/sh/clk/core.c 	list_add(&clk->node, &clock_list);
node              456 drivers/sh/clk/core.c 	list_del(&clk->node);
node              466 drivers/sh/clk/core.c 	list_for_each_entry(clkp, &clock_list, node)
node              576 drivers/sh/clk/core.c 	list_for_each_entry(clkp, &clock_list, node) {
node              613 drivers/sh/clk/core.c 	list_for_each_entry(clk, &clock_list, node)
node              153 drivers/siox/siox-core.c 	list_for_each_entry(sdevice, &smaster->devices, node) {
node              194 drivers/siox/siox-core.c 	list_for_each_entry(sdevice, &smaster->devices, node) {
node              362 drivers/siox/siox-core.c 		list_for_each_entry(sdevice, &smaster->devices, node) {
node              750 drivers/siox/siox-core.c 				       struct siox_device, node);
node              751 drivers/siox/siox-core.c 		list_del(&sdevice->node);
node              818 drivers/siox/siox-core.c 	list_add_tail(&sdevice->node, &smaster->devices);
node              858 drivers/siox/siox-core.c 	sdevice = container_of(smaster->devices.prev, struct siox_device, node);
node              859 drivers/siox/siox-core.c 	list_del(&sdevice->node);
node              156 drivers/slimbus/core.c 			   struct device_node *node)
node              166 drivers/slimbus/core.c 	if (node)
node              167 drivers/slimbus/core.c 		sbdev->dev.of_node = of_node_get(node);
node              180 drivers/slimbus/core.c 					     struct device_node *node)
node              190 drivers/slimbus/core.c 	ret = slim_add_device(ctrl, sbdev, node);
node              202 drivers/slimbus/core.c 	struct device_node *node;
node              207 drivers/slimbus/core.c 	for_each_child_of_node(ctrl->dev->of_node, node) {
node              214 drivers/slimbus/core.c 		compat = of_get_property(node, "compatible", NULL);
node              225 drivers/slimbus/core.c 		ret = of_property_read_u32_array(node, "reg", reg, 2);
node              237 drivers/slimbus/core.c 		sbdev = slim_alloc_device(ctrl, &e_addr, node);
node             1137 drivers/slimbus/qcom-ngd-ctrl.c 	struct device_node *node;
node             1139 drivers/slimbus/qcom-ngd-ctrl.c 	for_each_child_of_node(ctrl->ngd->pdev->dev.of_node, node) {
node             1140 drivers/slimbus/qcom-ngd-ctrl.c 		sbdev = of_slim_get_device(&ctrl->ctrl, node);
node             1263 drivers/slimbus/qcom-ngd-ctrl.c 	qmi->svc_info.sq_node = service->node;
node             1331 drivers/slimbus/qcom-ngd-ctrl.c 	struct device_node *node;
node             1336 drivers/slimbus/qcom-ngd-ctrl.c 	for_each_available_child_of_node(parent->of_node, node) {
node             1337 drivers/slimbus/qcom-ngd-ctrl.c 		if (of_property_read_u32(node, "reg", &id))
node             1342 drivers/slimbus/qcom-ngd-ctrl.c 			of_node_put(node);
node             1349 drivers/slimbus/qcom-ngd-ctrl.c 			of_node_put(node);
node             1355 drivers/slimbus/qcom-ngd-ctrl.c 		ngd->pdev->dev.of_node = node;
node              343 drivers/slimbus/slimbus.h 	struct list_head node;
node              116 drivers/slimbus/stream.c 	list_add_tail(&rt->node, &dev->stream_list);
node              469 drivers/slimbus/stream.c 	list_del(&stream->node);
node              200 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	struct device_node *node;
node              212 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	node = of_parse_phandle(dev->of_node, "flash", 0);
node              213 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	if (!node) {
node              216 drivers/soc/aspeed/aspeed-lpc-ctrl.c 		rc = of_address_to_resource(node, 1, &resm);
node              217 drivers/soc/aspeed/aspeed-lpc-ctrl.c 		of_node_put(node);
node              231 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	node = of_parse_phandle(dev->of_node, "memory-region", 0);
node              232 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	if (!node) {
node              235 drivers/soc/aspeed/aspeed-lpc-ctrl.c 		rc = of_address_to_resource(node, 0, &resm);
node              236 drivers/soc/aspeed/aspeed-lpc-ctrl.c 		of_node_put(node);
node              337 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	struct device_node *node;
node              349 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	node = of_parse_phandle(dev->of_node, "memory-region", 0);
node              350 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	if (node) {
node              351 drivers/soc/aspeed/aspeed-p2a-ctrl.c 		rc = of_address_to_resource(node, 0, &resm);
node              352 drivers/soc/aspeed/aspeed-p2a-ctrl.c 		of_node_put(node);
node              103 drivers/soc/fsl/dpaa2-console.c static int dpaa2_generic_console_open(struct inode *node, struct file *fp,
node              168 drivers/soc/fsl/dpaa2-console.c static int dpaa2_mc_console_open(struct inode *node, struct file *fp)
node              170 drivers/soc/fsl/dpaa2-console.c 	return dpaa2_generic_console_open(node, fp,
node              175 drivers/soc/fsl/dpaa2-console.c static int dpaa2_aiop_console_open(struct inode *node, struct file *fp)
node              177 drivers/soc/fsl/dpaa2-console.c 	return dpaa2_generic_console_open(node, fp,
node              182 drivers/soc/fsl/dpaa2-console.c static int dpaa2_console_close(struct inode *node, struct file *fp)
node               24 drivers/soc/fsl/dpio/dpio-service.c 	struct list_head node;
node               74 drivers/soc/fsl/dpio/dpio-service.c 	d = list_entry(dpio_list.next, struct dpaa2_io, node);
node               75 drivers/soc/fsl/dpio/dpio-service.c 	list_del(&d->node);
node               76 drivers/soc/fsl/dpio/dpio-service.c 	list_add_tail(&d->node, &dpio_list);
node              134 drivers/soc/fsl/dpio/dpio-service.c 	INIT_LIST_HEAD(&obj->node);
node              147 drivers/soc/fsl/dpio/dpio-service.c 	list_add_tail(&obj->node, &dpio_list);
node              170 drivers/soc/fsl/dpio/dpio-service.c 	list_del(&d->node);
node              273 drivers/soc/fsl/dpio/dpio-service.c 	list_add(&ctx->node, &d->notifications);
node              305 drivers/soc/fsl/dpio/dpio-service.c 	list_del(&ctx->node);
node              212 drivers/soc/fsl/qbman/bman_ccsr.c 	struct device_node *node = dev->of_node;
node              222 drivers/soc/fsl/qbman/bman_ccsr.c 			node);
node              264 drivers/soc/fsl/qbman/bman_ccsr.c 		dev_info(dev, "Can't get %pOF IRQ\n", node);
node              271 drivers/soc/fsl/qbman/bman_ccsr.c 			ret, node);
node              100 drivers/soc/fsl/qbman/bman_portal.c 	struct device_node *node = dev->of_node;
node              124 drivers/soc/fsl/qbman/bman_portal.c 		dev_err(dev, "Can't get %pOF property 'reg::CE'\n", node);
node              131 drivers/soc/fsl/qbman/bman_portal.c 		dev_err(dev, "Can't get %pOF property 'reg::CI'\n", node);
node             1481 drivers/soc/fsl/qbman/qman.c 	list_for_each_entry(cgr, &p->cgr_cbs, node)
node             2464 drivers/soc/fsl/qbman/qman.c 	list_add(&cgr->node, &p->cgr_cbs);
node             2504 drivers/soc/fsl/qbman/qman.c 	list_del(&cgr->node);
node             2509 drivers/soc/fsl/qbman/qman.c 	list_for_each_entry(i, &p->cgr_cbs, node)
node             2515 drivers/soc/fsl/qbman/qman.c 		list_add(&cgr->node, &p->cgr_cbs);
node             2526 drivers/soc/fsl/qbman/qman.c 		list_add(&cgr->node, &p->cgr_cbs);
node              755 drivers/soc/fsl/qbman/qman_ccsr.c 	struct device_node *node = dev->of_node;
node              766 drivers/soc/fsl/qbman/qman_ccsr.c 			node);
node              844 drivers/soc/fsl/qbman/qman_ccsr.c 			 node);
node              851 drivers/soc/fsl/qbman/qman_ccsr.c 			ret, node);
node              233 drivers/soc/fsl/qbman/qman_portal.c 	struct device_node *node = dev->of_node;
node              258 drivers/soc/fsl/qbman/qman_portal.c 		dev_err(dev, "Can't get %pOF property 'reg::CE'\n", node);
node              265 drivers/soc/fsl/qbman/qman_portal.c 		dev_err(dev, "Can't get %pOF property 'reg::CI'\n", node);
node              269 drivers/soc/fsl/qbman/qman_portal.c 	err = of_property_read_u32(node, "cell-index", &val);
node              271 drivers/soc/fsl/qbman/qman_portal.c 		dev_err(dev, "Can't get %pOF property 'cell-index'\n", node);
node              154 drivers/soc/fsl/qbman/qman_test_stash.c 	struct list_head node;
node              165 drivers/soc/fsl/qbman/qman_test_stash.c 	struct list_head node;
node              323 drivers/soc/fsl/qbman/qman_test_stash.c 	list_add_tail(&hp_cpu->node, &hp_cpu_list);
node              337 drivers/soc/fsl/qbman/qman_test_stash.c 		list_add_tail(&handler->node, &hp_cpu->handlers);
node              348 drivers/soc/fsl/qbman/qman_test_stash.c 	list_del(&hp_cpu->node);
node              353 drivers/soc/fsl/qbman/qman_test_stash.c 							node);
node              368 drivers/soc/fsl/qbman/qman_test_stash.c 		list_del(&handler->node);
node              448 drivers/soc/fsl/qbman/qman_test_stash.c 		list_for_each_entry(hp_cpu, &hp_cpu_list, node) {
node              454 drivers/soc/fsl/qbman/qman_test_stash.c 						struct hp_handler, node);
node              457 drivers/soc/fsl/qbman/qman_test_stash.c 						hp_cpu->iterator->node.next,
node              458 drivers/soc/fsl/qbman/qman_test_stash.c 						struct hp_handler, node);
node              476 drivers/soc/fsl/qbman/qman_test_stash.c 	hp_cpu = list_first_entry(&hp_cpu_list, struct hp_cpu, node);
node              477 drivers/soc/fsl/qbman/qman_test_stash.c 	handler = list_first_entry(&hp_cpu->handlers, struct hp_handler, node);
node              493 drivers/soc/fsl/qbman/qman_test_stash.c 		list_for_each_entry(hp_cpu, &hp_cpu_list, node) {
node              497 drivers/soc/fsl/qbman/qman_test_stash.c 						struct hp_handler, node);
node              500 drivers/soc/fsl/qbman/qman_test_stash.c 						hp_cpu->iterator->node.next,
node              501 drivers/soc/fsl/qbman/qman_test_stash.c 						struct hp_handler, node);
node              244 drivers/soc/fsl/qe/qe_ic.c static int qe_ic_host_match(struct irq_domain *h, struct device_node *node,
node              249 drivers/soc/fsl/qe/qe_ic.c 	return of_node == NULL || of_node == node;
node              316 drivers/soc/fsl/qe/qe_ic.c void __init qe_ic_init(struct device_node *node, unsigned int flags,
node              324 drivers/soc/fsl/qe/qe_ic.c 	ret = of_address_to_resource(node, 0, &res);
node              332 drivers/soc/fsl/qe/qe_ic.c 	qe_ic->irqhost = irq_domain_add_linear(node, NR_QE_IC_INTS,
node              343 drivers/soc/fsl/qe/qe_ic.c 	qe_ic->virq_high = irq_of_parse_and_map(node, 0);
node              344 drivers/soc/fsl/qe/qe_ic.c 	qe_ic->virq_low = irq_of_parse_and_map(node, 1);
node               30 drivers/soc/qcom/apr.c 	struct list_head node;
node               93 drivers/soc/qcom/apr.c 	list_add_tail(&abuf->node, &apr->rx_list);
node              177 drivers/soc/qcom/apr.c 		list_for_each_entry_safe(abuf, b, &apr->rx_list, node) {
node              180 drivers/soc/qcom/apr.c 			list_del(&abuf->node);
node              306 drivers/soc/qcom/apr.c 	struct device_node *node;
node              308 drivers/soc/qcom/apr.c 	for_each_child_of_node(dev->of_node, node) {
node              311 drivers/soc/qcom/apr.c 		if (of_property_read_u32(node, "reg", &id.svc_id))
node              316 drivers/soc/qcom/apr.c 		if (apr_add_device(dev, node, &id))
node              453 drivers/soc/qcom/qcom_aoss.c 				  struct device_node *node)
node              455 drivers/soc/qcom/qcom_aoss.c 	char *cdev_name = (char *)node->name;
node              461 drivers/soc/qcom/qcom_aoss.c 				(qmp->dev, node,
node              127 drivers/soc/qcom/qcom_gsbi.c 	struct device_node *node = pdev->dev.of_node;
node              148 drivers/soc/qcom/qcom_gsbi.c 	gsbi->tcsr = syscon_regmap_lookup_by_phandle(node, "syscon-tcsr");
node              151 drivers/soc/qcom/qcom_gsbi.c 		tcsr_node = of_parse_phandle(node, "syscon-tcsr", 0);
node              163 drivers/soc/qcom/qcom_gsbi.c 	if (of_property_read_u32(node, "cell-index", &gsbi_num)) {
node              173 drivers/soc/qcom/qcom_gsbi.c 	if (of_property_read_u32(node, "qcom,mode", &gsbi->mode)) {
node              179 drivers/soc/qcom/qcom_gsbi.c 	of_property_read_u32(node, "qcom,crci", &gsbi->crci);
node              216 drivers/soc/qcom/qcom_gsbi.c 	ret = of_platform_populate(node, NULL, NULL, &pdev->dev);
node               33 drivers/soc/qcom/qmi_interface.c 				unsigned int node, unsigned int port)
node               43 drivers/soc/qcom/qmi_interface.c 	if (!node && !port)
node               53 drivers/soc/qcom/qmi_interface.c 	svc->node = node;
node               73 drivers/soc/qcom/qmi_interface.c 				unsigned int node, unsigned int port)
node               80 drivers/soc/qcom/qmi_interface.c 		if (node != -1 && svc->node != node)
node              103 drivers/soc/qcom/qmi_interface.c 			 unsigned int node)
node              107 drivers/soc/qcom/qmi_interface.c 	qmi_recv_del_server(qmi, node, -1);
node              110 drivers/soc/qcom/qmi_interface.c 		ops->bye(qmi, node);
node              122 drivers/soc/qcom/qmi_interface.c 				unsigned int node, unsigned int port)
node              127 drivers/soc/qcom/qmi_interface.c 		ops->del_client(qmi, node, port);
node              142 drivers/soc/qcom/qmi_interface.c 		qmi_recv_bye(qmi, le32_to_cpu(pkt->client.node));
node              148 drivers/soc/qcom/qmi_interface.c 				    le32_to_cpu(pkt->server.node),
node              153 drivers/soc/qcom/qmi_interface.c 				    le32_to_cpu(pkt->server.node),
node              158 drivers/soc/qcom/qmi_interface.c 				    le32_to_cpu(pkt->client.node),
node              239 drivers/soc/qcom/qmi_interface.c 	pkt.server.node = cpu_to_le32(qmi->sq.sq_node);
node              173 drivers/soc/qcom/rmtfs_mem.c 	struct device_node *node = pdev->dev.of_node;
node              181 drivers/soc/qcom/rmtfs_mem.c 	rmem = of_reserved_mem_lookup(node);
node              187 drivers/soc/qcom/rmtfs_mem.c 	ret = of_property_read_u32(node, "qcom,client-id", &client_id);
node              229 drivers/soc/qcom/rmtfs_mem.c 	ret = of_property_read_u32(node, "qcom,vmid", &vmid);
node               91 drivers/soc/qcom/smp2p.c 	struct list_head node;
node              202 drivers/soc/qcom/smp2p.c 		list_for_each_entry(entry, &smp2p->inbound, node) {
node              213 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node) {
node              307 drivers/soc/qcom/smp2p.c 				    struct device_node *node)
node              309 drivers/soc/qcom/smp2p.c 	entry->domain = irq_domain_add_linear(node, 32, &smp2p_irq_ops, entry);
node              343 drivers/soc/qcom/smp2p.c 				     struct device_node *node)
node              357 drivers/soc/qcom/smp2p.c 	entry->state = qcom_smem_state_register(node, &smp2p_state_ops, entry);
node              444 drivers/soc/qcom/smp2p.c 	struct device_node *node;
node              500 drivers/soc/qcom/smp2p.c 	for_each_available_child_of_node(pdev->dev.of_node, node) {
node              510 drivers/soc/qcom/smp2p.c 		ret = of_property_read_string(node, "qcom,entry-name", &entry->name);
node              514 drivers/soc/qcom/smp2p.c 		if (of_property_read_bool(node, "interrupt-controller")) {
node              515 drivers/soc/qcom/smp2p.c 			ret = qcom_smp2p_inbound_entry(smp2p, entry, node);
node              519 drivers/soc/qcom/smp2p.c 			list_add(&entry->node, &smp2p->inbound);
node              521 drivers/soc/qcom/smp2p.c 			ret = qcom_smp2p_outbound_entry(smp2p, entry, node);
node              525 drivers/soc/qcom/smp2p.c 			list_add(&entry->node, &smp2p->outbound);
node              545 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node)
node              548 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->outbound, node)
node              568 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->inbound, node)
node              571 drivers/soc/qcom/smp2p.c 	list_for_each_entry(entry, &smp2p->outbound, node)
node              346 drivers/soc/qcom/smsm.c 	struct device_node *node = smsm->dev->of_node;
node              352 drivers/soc/qcom/smsm.c 	syscon = of_parse_phandle(node, key, 0);
node              360 drivers/soc/qcom/smsm.c 	ret = of_property_read_u32_index(node, key, 1, &host->ipc_offset);
node              366 drivers/soc/qcom/smsm.c 	ret = of_property_read_u32_index(node, key, 2, &host->ipc_bit);
node              383 drivers/soc/qcom/smsm.c 			      struct device_node *node)
node              388 drivers/soc/qcom/smsm.c 	irq = irq_of_parse_and_map(node, 0);
node              403 drivers/soc/qcom/smsm.c 	entry->domain = irq_domain_add_linear(node, 32, &smsm_irq_ops, entry);
node              458 drivers/soc/qcom/smsm.c 	struct device_node *node;
node              551 drivers/soc/qcom/smsm.c 	for_each_available_child_of_node(pdev->dev.of_node, node) {
node              552 drivers/soc/qcom/smsm.c 		if (!of_property_read_bool(node, "interrupt-controller"))
node              555 drivers/soc/qcom/smsm.c 		ret = of_property_read_u32(node, "reg", &id);
node              571 drivers/soc/qcom/smsm.c 		ret = smsm_inbound_entry(smsm, entry, node);
node              382 drivers/soc/rockchip/pm_domains.c 				      struct device_node *node)
node              391 drivers/soc/rockchip/pm_domains.c 	error = of_property_read_u32(node, "reg", &id);
node              395 drivers/soc/rockchip/pm_domains.c 			node, error);
node              401 drivers/soc/rockchip/pm_domains.c 			node, id);
node              408 drivers/soc/rockchip/pm_domains.c 			node, id);
node              419 drivers/soc/rockchip/pm_domains.c 	pd->num_clks = of_clk_get_parent_count(node);
node              427 drivers/soc/rockchip/pm_domains.c 			node, pd->num_clks);
node              432 drivers/soc/rockchip/pm_domains.c 		pd->clks[i].clk = of_clk_get(node, i);
node              437 drivers/soc/rockchip/pm_domains.c 				node, i, error);
node              446 drivers/soc/rockchip/pm_domains.c 	pd->num_qos = of_count_phandle_with_args(node, "pm_qos",
node              470 drivers/soc/rockchip/pm_domains.c 			qos_node = of_parse_phandle(node, "pm_qos", j);
node              489 drivers/soc/rockchip/pm_domains.c 			node, error);
node              493 drivers/soc/rockchip/pm_domains.c 	pd->genpd.name = node->name;
node              623 drivers/soc/rockchip/pm_domains.c 	struct device_node *node;
node              682 drivers/soc/rockchip/pm_domains.c 	for_each_available_child_of_node(np, node) {
node              683 drivers/soc/rockchip/pm_domains.c 		error = rockchip_pm_add_one_domain(pmu, node);
node              686 drivers/soc/rockchip/pm_domains.c 				node, error);
node              687 drivers/soc/rockchip/pm_domains.c 			of_node_put(node);
node              691 drivers/soc/rockchip/pm_domains.c 		error = rockchip_pm_add_subdomain(pmu, node);
node              694 drivers/soc/rockchip/pm_domains.c 				node, error);
node              695 drivers/soc/rockchip/pm_domains.c 			of_node_put(node);
node               95 drivers/soc/samsung/pm_domains.c static __init const char *exynos_get_domain_name(struct device_node *node)
node               99 drivers/soc/samsung/pm_domains.c 	if (of_property_read_string(node, "label", &name) < 0)
node              100 drivers/soc/samsung/pm_domains.c 		name = kbasename(node->full_name);
node              165 drivers/soc/sunxi/sunxi_sram.c static const struct sunxi_sram_data *sunxi_sram_of_parse(struct device_node *node,
node              175 drivers/soc/sunxi/sunxi_sram.c 	ret = of_parse_phandle_with_fixed_args(node, "allwinner,sram", 1, 0,
node              537 drivers/soc/ti/knav_dma.c 				struct device_node *node,
node              545 drivers/soc/ti/knav_dma.c 	ret = of_address_to_resource(node, index, &res);
node              548 drivers/soc/ti/knav_dma.c 			node, index);
node              555 drivers/soc/ti/knav_dma.c 			index, node);
node              623 drivers/soc/ti/knav_dma.c 	struct device_node *node = dma_node;
node              658 drivers/soc/ti/knav_dma.c 	dma->reg_global	 = pktdma_get_regs(dma, node, 0, &size);
node              666 drivers/soc/ti/knav_dma.c 	dma->reg_tx_chan = pktdma_get_regs(dma, node, 1, &size);
node              671 drivers/soc/ti/knav_dma.c 	dma->reg_rx_chan = pktdma_get_regs(dma, node, 2, &size);
node              676 drivers/soc/ti/knav_dma.c 	dma->reg_tx_sched = pktdma_get_regs(dma, node, 3, &size);
node              681 drivers/soc/ti/knav_dma.c 	dma->reg_rx_flow = pktdma_get_regs(dma, node, 4, &size);
node              689 drivers/soc/ti/knav_dma.c 	dma->enable_all	= (of_get_property(node, "ti,enable-all", NULL) != NULL);
node              690 drivers/soc/ti/knav_dma.c 	dma->loopback	= (of_get_property(node, "ti,loop-back",  NULL) != NULL);
node              692 drivers/soc/ti/knav_dma.c 	ret = of_property_read_u32(node, "ti,rx-retry-timeout", &timeout);
node              704 drivers/soc/ti/knav_dma.c 	strcpy(dma->name, node->name);
node              740 drivers/soc/ti/knav_dma.c 	struct device_node *node = pdev->dev.of_node;
node              744 drivers/soc/ti/knav_dma.c 	if (!node) {
node              767 drivers/soc/ti/knav_dma.c 	for_each_child_of_node(node, child) {
node              768 drivers/soc/ti/knav_dma.c 		ret = dma_init(node, child);
node              383 drivers/soc/ti/knav_qmss.h 					struct device_node *node,
node              472 drivers/soc/ti/knav_qmss_acc.c 			struct device_node *node,
node              487 drivers/soc/ti/knav_qmss_acc.c 	ret = of_property_read_u32_array(node, "accumulator", config, 5);
node              524 drivers/soc/ti/knav_qmss_acc.c 	if (of_get_property(node, "multi-queue", NULL)) {
node              769 drivers/soc/ti/knav_qmss_queue.c 	struct list_head *node;
node              827 drivers/soc/ti/knav_qmss_queue.c 	node = &region->pools;
node              835 drivers/soc/ti/knav_qmss_queue.c 	node = &pi->region_inst;
node              843 drivers/soc/ti/knav_qmss_queue.c 		list_add_tail(&pool->region_inst, node);
node             1073 drivers/soc/ti/knav_qmss_queue.c static const char *knav_queue_find_name(struct device_node *node)
node             1077 drivers/soc/ti/knav_qmss_queue.c 	if (of_property_read_string(node, "label", &name) < 0)
node             1078 drivers/soc/ti/knav_qmss_queue.c 		name = node->name;
node             1146 drivers/soc/ti/knav_qmss_queue.c 	struct device_node *node = pdev->dev.of_node;
node             1162 drivers/soc/ti/knav_qmss_queue.c 	if (!of_property_read_u32_array(node, name , temp, 2)) {
node             1217 drivers/soc/ti/knav_qmss_queue.c 					struct device_node *node)
node             1232 drivers/soc/ti/knav_qmss_queue.c 	range->name = knav_queue_find_name(node);
node             1233 drivers/soc/ti/knav_qmss_queue.c 	ret = of_property_read_u32_array(node, "qrange", temp, 2);
node             1246 drivers/soc/ti/knav_qmss_queue.c 		if (of_irq_parse_one(node, i, &oirq))
node             1274 drivers/soc/ti/knav_qmss_queue.c 	if (of_get_property(node, "qalloc-by-id", NULL))
node             1277 drivers/soc/ti/knav_qmss_queue.c 	if (of_get_property(node, "accumulator", NULL)) {
node             1278 drivers/soc/ti/knav_qmss_queue.c 		ret = knav_init_acc_range(kdev, node, range);
node             1377 drivers/soc/ti/knav_qmss_queue.c 					struct device_node *node, int index)
node             1383 drivers/soc/ti/knav_qmss_queue.c 	ret = of_address_to_resource(node, index, &res);
node             1386 drivers/soc/ti/knav_qmss_queue.c 			node, index);
node             1393 drivers/soc/ti/knav_qmss_queue.c 			index, node);
node             1761 drivers/soc/ti/knav_qmss_queue.c 	struct device_node *node = pdev->dev.of_node;
node             1768 drivers/soc/ti/knav_qmss_queue.c 	if (!node) {
node             1798 drivers/soc/ti/knav_qmss_queue.c 	if (of_property_read_u32_array(node, "queue-range", temp, 2)) {
node             1807 drivers/soc/ti/knav_qmss_queue.c 	qmgrs =  of_get_child_by_name(node, "qmgrs");
node             1819 drivers/soc/ti/knav_qmss_queue.c 	pdsps =  of_get_child_by_name(node, "pdsps");
node             1832 drivers/soc/ti/knav_qmss_queue.c 	queue_pools = of_get_child_by_name(node, "queue-pools");
node             1861 drivers/soc/ti/knav_qmss_queue.c 	regions =  of_get_child_by_name(node, "descriptor-regions");
node              123 drivers/soundwire/bus.c 	list_del_init(&slave->node);
node              415 drivers/soundwire/bus.c 	list_for_each_entry(slave, &bus->slaves, node) {
node              559 drivers/soundwire/bus.c 		list_for_each_entry_safe(slave, _s, &bus->slaves, node) {
node              166 drivers/soundwire/mipi_disco.c 	struct fwnode_handle *node;
node              182 drivers/soundwire/mipi_disco.c 		node = device_get_named_child_node(&slave->dev, name);
node              183 drivers/soundwire/mipi_disco.c 		if (!node) {
node              188 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-port-max-wordlength",
node              190 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-port-min-wordlength",
node              193 drivers/soundwire/mipi_disco.c 		nval = fwnode_property_count_u32(node, "mipi-sdw-port-wordlength-configs");
node              203 drivers/soundwire/mipi_disco.c 			fwnode_property_read_u32_array(node,
node              208 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-data-port-type",
node              211 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node,
node              215 drivers/soundwire/mipi_disco.c 		dpn[i].simple_ch_prep_sm = fwnode_property_read_bool(node,
node              218 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node,
node              222 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node,
node              226 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-min-channel-number",
node              229 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-max-channel-number",
node              232 drivers/soundwire/mipi_disco.c 		nval = fwnode_property_count_u32(node, "mipi-sdw-channel-number-list");
node              241 drivers/soundwire/mipi_disco.c 			fwnode_property_read_u32_array(node,
node              246 drivers/soundwire/mipi_disco.c 		nval = fwnode_property_count_u32(node, "mipi-sdw-channel-combination-list");
node              256 drivers/soundwire/mipi_disco.c 			fwnode_property_read_u32_array(node,
node              262 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node,
node              265 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-max-async-buffer",
node              268 drivers/soundwire/mipi_disco.c 		dpn[i].block_pack_mode = fwnode_property_read_bool(node,
node              271 drivers/soundwire/mipi_disco.c 		fwnode_property_read_u32(node, "mipi-sdw-port-encoding-type",
node               45 drivers/soundwire/slave.c 	list_add_tail(&slave->node, &bus->slaves);
node               57 drivers/soundwire/slave.c 		list_del(&slave->node);
node               83 drivers/soundwire/slave.c 	list_for_each_entry(adev, &parent->children, node) {
node              128 drivers/soundwire/slave.c 	struct device_node *node;
node              130 drivers/soundwire/slave.c 	for_each_child_of_node(bus->dev->of_node, node) {
node              137 drivers/soundwire/slave.c 		compat = of_get_property(node, "compatible", NULL);
node              150 drivers/soundwire/slave.c 		addr = of_get_property(node, "reg", &len);
node              164 drivers/soundwire/slave.c 		sdw_slave_add(bus, &id, of_fwnode_handle(node));
node              819 drivers/spi/spi-davinci.c 	struct device_node *node = pdev->dev.of_node;
node              843 drivers/spi/spi-davinci.c 	of_property_read_u32(node, "num-cs", &num_cs);
node              845 drivers/spi/spi-davinci.c 	of_property_read_u32(node, "ti,davinci-spi-intr-line", &intr_line);
node              142 drivers/spi/spi-jcore.c 	struct device_node *node = pdev->dev.of_node;
node              159 drivers/spi/spi-jcore.c 	master->dev.of_node = node;
node              140 drivers/spi/spi-omap2-mcspi.c 	struct list_head	node;
node             1055 drivers/spi/spi-omap2-mcspi.c 		list_add_tail(&cs->node, &ctx->cs);
node             1089 drivers/spi/spi-omap2-mcspi.c 		list_del(&cs->node);
node             1275 drivers/spi/spi-omap2-mcspi.c 	list_for_each_entry(cs, &ctx->cs, node) {
node             1349 drivers/spi/spi-omap2-mcspi.c 	list_for_each_entry(cs, &ctx->cs, node) {
node             1399 drivers/spi/spi-omap2-mcspi.c 	struct device_node	*node = pdev->dev.of_node;
node             1402 drivers/spi/spi-omap2-mcspi.c 	if (of_property_read_bool(node, "spi-slave"))
node             1420 drivers/spi/spi-omap2-mcspi.c 	master->dev.of_node = node;
node             1434 drivers/spi/spi-omap2-mcspi.c 		of_property_read_u32(node, "ti,spi-num-cs", &num_cs);
node             1436 drivers/spi/spi-omap2-mcspi.c 		if (of_get_property(node, "ti,pindir-d0-out-d1-in", NULL))
node             3659 drivers/spi/spi.c struct spi_device *of_find_spi_device_by_node(struct device_node *node)
node             3661 drivers/spi/spi.c 	struct device *dev = bus_find_device_by_of_node(&spi_bus_type, node);
node             3670 drivers/spi/spi.c static struct spi_controller *of_find_spi_controller_by_node(struct device_node *node)
node             3674 drivers/spi/spi.c 	dev = class_find_device_by_of_node(&spi_master_class, node);
node             3676 drivers/spi/spi.c 		dev = class_find_device_by_of_node(&spi_slave_class, node);
node              451 drivers/spmi/spmi.c 	struct device_node *node;
node              457 drivers/spmi/spmi.c 	for_each_available_child_of_node(ctrl->dev.of_node, node) {
node              461 drivers/spmi/spmi.c 		dev_dbg(&ctrl->dev, "adding child %pOF\n", node);
node              463 drivers/spmi/spmi.c 		err = of_property_read_u32_array(node, "reg", reg, 2);
node              467 drivers/spmi/spmi.c 				node, err);
node              474 drivers/spmi/spmi.c 				node);
node              479 drivers/spmi/spmi.c 			dev_err(&ctrl->dev, "invalid usid on node %pOF\n", node);
node              489 drivers/spmi/spmi.c 		sdev->dev.of_node = node;
node              379 drivers/staging/android/ion/ion.c 	plist_for_each_entry(heap, &dev->heaps, node) {
node              435 drivers/staging/android/ion/ion.c 	plist_for_each_entry(heap, &dev->heaps, node) {
node              634 drivers/staging/android/ion/ion.c 	plist_node_init(&heap->node, -heap->id);
node              635 drivers/staging/android/ion/ion.c 	plist_add(&heap->node, &dev->heaps);
node              140 drivers/staging/android/ion/ion.h 	struct plist_node node;
node              349 drivers/staging/android/vsoc.c 				     struct fd_scoped_permission_node *node)
node              351 drivers/staging/android/vsoc.c 	if (node) {
node              353 drivers/staging/android/vsoc.c 						&node->permission);
node              355 drivers/staging/android/vsoc.c 		list_del(&node->list);
node              357 drivers/staging/android/vsoc.c 		kfree(node);
node              545 drivers/staging/android/vsoc.c 			struct fd_scoped_permission_node *node = NULL;
node              547 drivers/staging/android/vsoc.c 			node = kzalloc(sizeof(*node), GFP_KERNEL);
node              549 drivers/staging/android/vsoc.c 			if (!node)
node              551 drivers/staging/android/vsoc.c 			INIT_LIST_HEAD(&node->list);
node              554 drivers/staging/android/vsoc.c 				 node,
node              558 drivers/staging/android/vsoc.c 				list_add(&node->list, &vsoc_dev.permissions);
node              561 drivers/staging/android/vsoc.c 				kfree(node);
node              569 drivers/staging/android/vsoc.c 			struct fd_scoped_permission_node *node =
node              572 drivers/staging/android/vsoc.c 			if (!node)
node              576 drivers/staging/android/vsoc.c 			     &node->permission, sizeof(node->permission)))
node             1059 drivers/staging/android/vsoc.c 	struct fd_scoped_permission_node *node = NULL;
node             1069 drivers/staging/android/vsoc.c 	node = private_data->fd_scoped_permission_node;
node             1070 drivers/staging/android/vsoc.c 	if (node) {
node             1076 drivers/staging/android/vsoc.c 		do_destroy_fd_scoped_permission_node(owner_region_p, node);
node              909 drivers/staging/fbtft/fbtft-core.c 	struct device_node *node = par->info->device->of_node;
node              915 drivers/staging/fbtft/fbtft-core.c 	if (!node)
node              918 drivers/staging/fbtft/fbtft-core.c 	prop = of_find_property(node, "init", NULL);
node             1142 drivers/staging/fbtft/fbtft-core.c static u32 fbtft_of_value(struct device_node *node, const char *propname)
node             1147 drivers/staging/fbtft/fbtft-core.c 	ret = of_property_read_u32(node, propname, &val);
node             1156 drivers/staging/fbtft/fbtft-core.c 	struct device_node *node = dev->of_node;
node             1159 drivers/staging/fbtft/fbtft-core.c 	if (!node) {
node             1168 drivers/staging/fbtft/fbtft-core.c 	pdata->display.width = fbtft_of_value(node, "width");
node             1169 drivers/staging/fbtft/fbtft-core.c 	pdata->display.height = fbtft_of_value(node, "height");
node             1170 drivers/staging/fbtft/fbtft-core.c 	pdata->display.regwidth = fbtft_of_value(node, "regwidth");
node             1171 drivers/staging/fbtft/fbtft-core.c 	pdata->display.buswidth = fbtft_of_value(node, "buswidth");
node             1172 drivers/staging/fbtft/fbtft-core.c 	pdata->display.backlight = fbtft_of_value(node, "backlight");
node             1173 drivers/staging/fbtft/fbtft-core.c 	pdata->display.bpp = fbtft_of_value(node, "bpp");
node             1174 drivers/staging/fbtft/fbtft-core.c 	pdata->display.debug = fbtft_of_value(node, "debug");
node             1175 drivers/staging/fbtft/fbtft-core.c 	pdata->rotate = fbtft_of_value(node, "rotate");
node             1176 drivers/staging/fbtft/fbtft-core.c 	pdata->bgr = of_property_read_bool(node, "bgr");
node             1177 drivers/staging/fbtft/fbtft-core.c 	pdata->fps = fbtft_of_value(node, "fps");
node             1178 drivers/staging/fbtft/fbtft-core.c 	pdata->txbuflen = fbtft_of_value(node, "txbuflen");
node             1179 drivers/staging/fbtft/fbtft-core.c 	pdata->startbyte = fbtft_of_value(node, "startbyte");
node             1180 drivers/staging/fbtft/fbtft-core.c 	of_property_read_string(node, "gamma", (const char **)&pdata->gamma);
node             1182 drivers/staging/fbtft/fbtft-core.c 	if (of_find_property(node, "led-gpios", NULL))
node             1184 drivers/staging/fbtft/fbtft-core.c 	if (of_find_property(node, "init", NULL))
node             1250 drivers/staging/fieldbus/anybuss/host.c 	struct device_node *node;
node             1254 drivers/staging/fieldbus/anybuss/host.c 	for_each_child_of_node(dev->of_node, node) {
node             1255 drivers/staging/fieldbus/anybuss/host.c 		if (anybus_of_get_host_idx(node) == host_idx)
node             1256 drivers/staging/fieldbus/anybuss/host.c 			return node;
node              187 drivers/staging/fieldbus/dev_core.c static int fieldbus_release(struct inode *node, struct file *filp)
node               30 drivers/staging/greybus/authentication.c 	struct list_head	node;
node               69 drivers/staging/greybus/authentication.c 	list_for_each_entry(cap, &cap_list, node) {
node              317 drivers/staging/greybus/authentication.c 	list_add(&cap->node, &cap_list);
node              356 drivers/staging/greybus/authentication.c 	list_del(&cap->node);
node              390 drivers/staging/greybus/authentication.c 	list_del(&cap->node);
node               27 drivers/staging/greybus/fw-download.c 	struct list_head	node;
node               99 drivers/staging/greybus/fw-download.c 	list_for_each_entry(fw_req, &fw_download->fw_requests, node) {
node              122 drivers/staging/greybus/fw-download.c 	list_del(&fw_req->node);
node              202 drivers/staging/greybus/fw-download.c 	list_add(&fw_req->node, &fw_download->fw_requests);
node              453 drivers/staging/greybus/fw-download.c 	list_for_each_entry(fw_req, &fw_download->fw_requests, node)
node              458 drivers/staging/greybus/fw-download.c 	list_for_each_entry_safe(fw_req, tmp, &fw_download->fw_requests, node) {
node               27 drivers/staging/greybus/fw-management.c 	struct list_head	node;
node               89 drivers/staging/greybus/fw-management.c 	list_for_each_entry(fw_mgmt, &fw_mgmt_list, node) {
node              610 drivers/staging/greybus/fw-management.c 	list_add(&fw_mgmt->node, &fw_mgmt_list);
node              650 drivers/staging/greybus/fw-management.c 	list_del(&fw_mgmt->node);
node              684 drivers/staging/greybus/fw-management.c 	list_del(&fw_mgmt->node);
node              245 drivers/staging/greybus/tools/loopback_test.c int open_sysfs(const char *sys_pfx, const char *node, int flags)
node              250 drivers/staging/greybus/tools/loopback_test.c 	snprintf(path, sizeof(path), "%s%s", sys_pfx, node);
node              259 drivers/staging/greybus/tools/loopback_test.c int read_sysfs_int_fd(int fd, const char *sys_pfx, const char *node)
node              264 drivers/staging/greybus/tools/loopback_test.c 		fprintf(stderr, "unable to read from %s%s %s\n", sys_pfx, node,
node              272 drivers/staging/greybus/tools/loopback_test.c float read_sysfs_float_fd(int fd, const char *sys_pfx, const char *node)
node              278 drivers/staging/greybus/tools/loopback_test.c 		fprintf(stderr, "unable to read from %s%s %s\n", sys_pfx, node,
node              286 drivers/staging/greybus/tools/loopback_test.c int read_sysfs_int(const char *sys_pfx, const char *node)
node              290 drivers/staging/greybus/tools/loopback_test.c 	fd = open_sysfs(sys_pfx, node, O_RDONLY);
node              291 drivers/staging/greybus/tools/loopback_test.c 	val = read_sysfs_int_fd(fd, sys_pfx, node);
node              296 drivers/staging/greybus/tools/loopback_test.c float read_sysfs_float(const char *sys_pfx, const char *node)
node              301 drivers/staging/greybus/tools/loopback_test.c 	fd = open_sysfs(sys_pfx, node, O_RDONLY);
node              302 drivers/staging/greybus/tools/loopback_test.c 	val = read_sysfs_float_fd(fd, sys_pfx, node);
node              307 drivers/staging/greybus/tools/loopback_test.c void write_sysfs_val(const char *sys_pfx, const char *node, int val)
node              312 drivers/staging/greybus/tools/loopback_test.c 	fd = open_sysfs(sys_pfx, node, O_RDWR);
node              315 drivers/staging/greybus/tools/loopback_test.c 		fprintf(stderr, "unable to write to %s%s %s\n", sys_pfx, node,
node              573 drivers/staging/greybus/tools/loopback_test.c int is_loopback_device(const char *path, const char *node)
node              577 drivers/staging/greybus/tools/loopback_test.c 	snprintf(file, MAX_SYSFS_PATH, "%s%s/iteration_count", path, node);
node               74 drivers/staging/media/imx/imx-media-dev.c 	struct device_node *node = dev->of_node;
node               82 drivers/staging/media/imx/imx-media-dev.c 	ret = imx_media_add_of_subdevs(imxmd, node);
node             1193 drivers/staging/media/imx/imx7-media-csi.c 	struct device_node *node = dev->of_node;
node             1238 drivers/staging/media/imx/imx7-media-csi.c 	ret = imx_media_of_add_csi(imxmd, node);
node              828 drivers/staging/media/imx/imx7-mipi-csis.c 	struct device_node *node = pdev->dev.of_node;
node              830 drivers/staging/media/imx/imx7-mipi-csis.c 	if (of_property_read_u32(node, "clock-frequency",
node              840 drivers/staging/media/imx/imx7-mipi-csis.c 	of_property_read_u32(node, "fsl,csis-hs-settle", &state->hs_settle);
node               56 drivers/staging/media/ipu3/ipu3-v4l2.c 	unsigned int node;
node               86 drivers/staging/media/ipu3/ipu3-v4l2.c 		node = imgu_map_node(imgu, i);
node               88 drivers/staging/media/ipu3/ipu3-v4l2.c 		if (node == IMGU_NODE_STAT_3A || node == IMGU_NODE_PARAMS)
node               90 drivers/staging/media/ipu3/ipu3-v4l2.c 		fmts[i] = imgu_pipe->queue_enabled[node] ?
node               91 drivers/staging/media/ipu3/ipu3-v4l2.c 			&imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp : NULL;
node              311 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              313 drivers/staging/media/ipu3/ipu3-v4l2.c 	unsigned int queue = imgu_node_to_queue(node->id);
node              327 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              329 drivers/staging/media/ipu3/ipu3-v4l2.c 	unsigned int queue = imgu_node_to_queue(node->id);
node              341 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              343 drivers/staging/media/ipu3/ipu3-v4l2.c 	unsigned int queue = imgu_node_to_queue(node->id);
node              351 drivers/staging/media/ipu3/ipu3-v4l2.c 		need_bytes = node->vdev_fmt.fmt.meta.buffersize;
node              353 drivers/staging/media/ipu3/ipu3-v4l2.c 		need_bytes = node->vdev_fmt.fmt.pix_mp.plane_fmt[0].sizeimage;
node              365 drivers/staging/media/ipu3/ipu3-v4l2.c 	list_add_tail(&buf->vid_buf.list, &node->buffers);
node              371 drivers/staging/media/ipu3/ipu3-v4l2.c 		imgu_queue_buffers(imgu, false, node->pipe);
node              374 drivers/staging/media/ipu3/ipu3-v4l2.c 		node->pipe, node->id);
node              384 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              386 drivers/staging/media/ipu3/ipu3-v4l2.c 	const struct v4l2_format *fmt = &node->vdev_fmt;
node              408 drivers/staging/media/ipu3/ipu3-v4l2.c 	INIT_LIST_HEAD(&node->buffers);
node              418 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node;
node              430 drivers/staging/media/ipu3/ipu3-v4l2.c 			node = &imgu->imgu_pipe[p].nodes[i];
node              432 drivers/staging/media/ipu3/ipu3-v4l2.c 				__func__, p, i, node->name, node->enabled);
node              433 drivers/staging/media/ipu3/ipu3-v4l2.c 			if (node == except)
node              435 drivers/staging/media/ipu3/ipu3-v4l2.c 			if (node->enabled && !vb2_start_streaming_called(&node->vbq))
node              444 drivers/staging/media/ipu3/ipu3-v4l2.c 				    struct imgu_video_device *node,
node              451 drivers/staging/media/ipu3/ipu3-v4l2.c 	list_for_each_entry_safe(b, b0, &node->buffers, list) {
node              463 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              469 drivers/staging/media/ipu3/ipu3-v4l2.c 		node->name, node->pipe, node->id);
node              476 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (!node->enabled) {
node              482 drivers/staging/media/ipu3/ipu3-v4l2.c 	pipe = node->pipe;
node              484 drivers/staging/media/ipu3/ipu3-v4l2.c 	r = media_pipeline_start(&node->vdev.entity, &imgu_pipe->pipeline);
node              489 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (!imgu_all_nodes_streaming(imgu, node))
node              508 drivers/staging/media/ipu3/ipu3-v4l2.c 	media_pipeline_stop(&node->vdev.entity);
node              510 drivers/staging/media/ipu3/ipu3-v4l2.c 	imgu_return_all_buffers(imgu, node, VB2_BUF_STATE_QUEUED);
node              520 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node =
node              525 drivers/staging/media/ipu3/ipu3-v4l2.c 	WARN_ON(!node->enabled);
node              527 drivers/staging/media/ipu3/ipu3-v4l2.c 	pipe = node->pipe;
node              528 drivers/staging/media/ipu3/ipu3-v4l2.c 	dev_dbg(dev, "Try to stream off node [%u][%u]", pipe, node->id);
node              536 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (imgu->streaming && imgu_all_nodes_streaming(imgu, node)) {
node              544 drivers/staging/media/ipu3/ipu3-v4l2.c 	imgu_return_all_buffers(imgu, node, VB2_BUF_STATE_ERROR);
node              545 drivers/staging/media/ipu3/ipu3-v4l2.c 	media_pipeline_stop(&node->vdev.entity);
node              587 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              591 drivers/staging/media/ipu3/ipu3-v4l2.c 	snprintf(cap->bus_info, sizeof(cap->bus_info), "PCI:%s", node->name);
node              638 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              640 drivers/staging/media/ipu3/ipu3-v4l2.c 	f->fmt = node->vdev_fmt.fmt;
node              649 drivers/staging/media/ipu3/ipu3-v4l2.c static int imgu_fmt(struct imgu_device *imgu, unsigned int pipe, int node,
node              662 drivers/staging/media/ipu3/ipu3-v4l2.c 	dev_dbg(dev, "set fmt node [%u][%u](try = %u)", pipe, node, try);
node              721 drivers/staging/media/ipu3/ipu3-v4l2.c 	css_q = imgu_node_to_queue(node);
node              740 drivers/staging/media/ipu3/ipu3-v4l2.c 		f->fmt = imgu_pipe->nodes[node].vdev_fmt.fmt;
node              776 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              781 drivers/staging/media/ipu3/ipu3-v4l2.c 		pix_mp->width, pix_mp->height, node->id);
node              787 drivers/staging/media/ipu3/ipu3-v4l2.c 	return imgu_fmt(imgu, node->pipe, node->id, f, true);
node              794 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              799 drivers/staging/media/ipu3/ipu3-v4l2.c 		pix_mp->width, pix_mp->height, node->id);
node              805 drivers/staging/media/ipu3/ipu3-v4l2.c 	return imgu_fmt(imgu, node->pipe, node->id, f, false);
node              822 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              826 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (fmt->index > 0 || fmt->type != node->vbq.type)
node              838 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = file_to_intel_imgu_node(file);
node              840 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (f->type != node->vbq.type)
node              843 drivers/staging/media/ipu3/ipu3-v4l2.c 	f->fmt = node->vdev_fmt.fmt;
node             1055 drivers/staging/media/ipu3/ipu3-v4l2.c static void imgu_node_to_v4l2(u32 node, struct video_device *vdev,
node             1061 drivers/staging/media/ipu3/ipu3-v4l2.c 	WARN_ON(node >= IMGU_NODE_NUM);
node             1063 drivers/staging/media/ipu3/ipu3-v4l2.c 	switch (node) {
node             1160 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct imgu_video_device *node = &imgu_pipe->nodes[node_num];
node             1161 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct video_device *vdev = &node->vdev;
node             1162 drivers/staging/media/ipu3/ipu3-v4l2.c 	struct vb2_queue *vbq = &node->vbq;
node             1188 drivers/staging/media/ipu3/ipu3-v4l2.c 	mutex_init(&node->lock);
node             1189 drivers/staging/media/ipu3/ipu3-v4l2.c 	INIT_LIST_HEAD(&node->buffers);
node             1192 drivers/staging/media/ipu3/ipu3-v4l2.c 	node->pad_fmt = def_bus_fmt;
node             1193 drivers/staging/media/ipu3/ipu3-v4l2.c 	node->id = node_num;
node             1194 drivers/staging/media/ipu3/ipu3-v4l2.c 	node->pipe = pipe;
node             1195 drivers/staging/media/ipu3/ipu3-v4l2.c 	imgu_node_to_v4l2(node_num, vdev, &node->vdev_fmt);
node             1196 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (node->vdev_fmt.type ==
node             1198 drivers/staging/media/ipu3/ipu3-v4l2.c 	    node->vdev_fmt.type ==
node             1200 drivers/staging/media/ipu3/ipu3-v4l2.c 		def_pix_fmt.pixelformat = node->output ?
node             1203 drivers/staging/media/ipu3/ipu3-v4l2.c 		node->vdev_fmt.fmt.pix_mp = def_pix_fmt;
node             1207 drivers/staging/media/ipu3/ipu3-v4l2.c 	r = media_entity_pads_init(&vdev->entity, 1, &node->vdev_pad);
node             1210 drivers/staging/media/ipu3/ipu3-v4l2.c 		mutex_destroy(&node->lock);
node             1213 drivers/staging/media/ipu3/ipu3-v4l2.c 	node->vdev_pad.flags = node->output ?
node             1218 drivers/staging/media/ipu3/ipu3-v4l2.c 	vbq->type = node->vdev_fmt.type;
node             1230 drivers/staging/media/ipu3/ipu3-v4l2.c 	vbq->lock = &node->lock;
node             1240 drivers/staging/media/ipu3/ipu3-v4l2.c 		 IMGU_NAME, pipe, node->name);
node             1243 drivers/staging/media/ipu3/ipu3-v4l2.c 	vdev->lock = &node->lock;
node             1245 drivers/staging/media/ipu3/ipu3-v4l2.c 	vdev->queue = &node->vbq;
node             1246 drivers/staging/media/ipu3/ipu3-v4l2.c 	vdev->vfl_dir = node->output ? VFL_DIR_TX : VFL_DIR_RX;
node             1257 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (node->enabled)
node             1259 drivers/staging/media/ipu3/ipu3-v4l2.c 	if (node->output) {
node             1276 drivers/staging/media/ipu3/ipu3-v4l2.c 					 unsigned int pipe, int node)
node             1281 drivers/staging/media/ipu3/ipu3-v4l2.c 	for (i = 0; i < node; i++) {
node               51 drivers/staging/media/ipu3/ipu3.c unsigned int imgu_node_to_queue(unsigned int node)
node               53 drivers/staging/media/ipu3/ipu3.c 	return imgu_node_map[node].css_queue;
node              110 drivers/staging/media/ipu3/ipu3.c 	unsigned int i, k, node;
node              116 drivers/staging/media/ipu3/ipu3.c 		node = imgu_map_node(imgu, i);
node              117 drivers/staging/media/ipu3/ipu3.c 		if (!imgu_pipe->queue_enabled[node] || i == IMGU_QUEUE_MASTER)
node              128 drivers/staging/media/ipu3/ipu3.c 		meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta;
node              129 drivers/staging/media/ipu3/ipu3.c 		mpix = &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp;
node              131 drivers/staging/media/ipu3/ipu3.c 		if (node == IMGU_NODE_STAT_3A || node == IMGU_NODE_PARAMS)
node              204 drivers/staging/media/ipu3/ipu3.c 						 unsigned int node,
node              210 drivers/staging/media/ipu3/ipu3.c 	if (WARN_ON(node >= IMGU_NODE_NUM))
node              214 drivers/staging/media/ipu3/ipu3.c 	list_for_each_entry(buf, &imgu_pipe->nodes[node].buffers, vid_buf.list) {
node              220 drivers/staging/media/ipu3/ipu3.c 	return imgu_dummybufs_get(imgu, imgu_node_map[node].css_queue, pipe);
node              229 drivers/staging/media/ipu3/ipu3.c 	unsigned int node;
node              245 drivers/staging/media/ipu3/ipu3.c 	for (node = IMGU_NODE_NUM - 1;
node              247 drivers/staging/media/ipu3/ipu3.c 	     node = node ? node - 1 : IMGU_NODE_NUM - 1) {
node              248 drivers/staging/media/ipu3/ipu3.c 		if (node == IMGU_NODE_VF &&
node              253 drivers/staging/media/ipu3/ipu3.c 		} else if (node == IMGU_NODE_PARAMS &&
node              254 drivers/staging/media/ipu3/ipu3.c 			   imgu_pipe->nodes[node].enabled) {
node              259 drivers/staging/media/ipu3/ipu3.c 			if (list_empty(&imgu_pipe->nodes[node].buffers))
node              262 drivers/staging/media/ipu3/ipu3.c 			ivb = list_first_entry(&imgu_pipe->nodes[node].buffers,
node              278 drivers/staging/media/ipu3/ipu3.c 		} else if (imgu_pipe->queue_enabled[node]) {
node              280 drivers/staging/media/ipu3/ipu3.c 				imgu_queue_getbuf(imgu, node, pipe);
node              297 drivers/staging/media/ipu3/ipu3.c 				imgu_node_map[node].name,
node              316 drivers/staging/media/ipu3/ipu3.c 		node, r);
node              322 drivers/staging/media/ipu3/ipu3.c 	for (node = 0; node < IMGU_NODE_NUM; node++) {
node              325 drivers/staging/media/ipu3/ipu3.c 		if (!imgu_pipe->queue_enabled[node])
node              330 drivers/staging/media/ipu3/ipu3.c 					 &imgu_pipe->nodes[node].buffers,
node              512 drivers/staging/media/ipu3/ipu3.c 		unsigned int node, pipe;
node              529 drivers/staging/media/ipu3/ipu3.c 		node = imgu_map_node(imgu, b->queue);
node              537 drivers/staging/media/ipu3/ipu3.c 			imgu_node_map[node].name,
node              547 drivers/staging/media/ipu3/ipu3.c 		if (!imgu_pipe->nodes[node].output) {
node              552 drivers/staging/media/ipu3/ipu3.c 				&imgu_pipe->nodes[node].sequence);
node              157 drivers/staging/media/ipu3/ipu3.h unsigned int imgu_node_to_queue(unsigned int node);
node              412 drivers/staging/mt7621-dma/mtk-hsdma.c 			list_del(&desc->vdesc.node);
node              284 drivers/staging/mt7621-pci/pci-mt7621.c 	struct device_node *node = dev->of_node;
node              289 drivers/staging/mt7621-pci/pci-mt7621.c 	if (of_pci_range_parser_init(&parser, node)) {
node              310 drivers/staging/mt7621-pci/pci-mt7621.c 			of_pci_range_to_resource(&range, node, res);
node              313 drivers/staging/mt7621-pci/pci-mt7621.c 	err = of_pci_parse_bus_range(node, &pcie->busn);
node              316 drivers/staging/mt7621-pci/pci-mt7621.c 		pcie->busn.name = node->name;
node              326 drivers/staging/mt7621-pci/pci-mt7621.c 				  struct device_node *node,
node              374 drivers/staging/mt7621-pci/pci-mt7621.c 	struct device_node *node = dev->of_node, *child;
node              384 drivers/staging/mt7621-pci/pci-mt7621.c 	err = of_address_to_resource(node, 0, &regs);
node              400 drivers/staging/mt7621-pci/pci-mt7621.c 	for_each_available_child_of_node(node, child) {
node              270 drivers/staging/nvec/nvec.c 	list_add_tail(&msg->node, &nvec->tx_data);
node              382 drivers/staging/nvec/nvec.c 		msg = list_first_entry(&nvec->tx_data, struct nvec_msg, node);
node              397 drivers/staging/nvec/nvec.c 			list_del_init(&msg->node);
node              445 drivers/staging/nvec/nvec.c 		msg = list_first_entry(&nvec->rx_data, struct nvec_msg, node);
node              446 drivers/staging/nvec/nvec.c 		list_del_init(&msg->node);
node              511 drivers/staging/nvec/nvec.c 	list_add_tail(&nvec->rx->node, &nvec->rx_data);
node              555 drivers/staging/nvec/nvec.c 		list_add_tail(&nvec->tx->node, &nvec->tx_data);
node              558 drivers/staging/nvec/nvec.c 					    node);
node               95 drivers/staging/nvec/nvec.h 	struct list_head node;
node              273 drivers/staging/octeon-usb/octeon-hcd.c 	struct list_head node;
node              317 drivers/staging/octeon-usb/octeon-hcd.c 	struct list_head node;
node             1134 drivers/staging/octeon-usb/octeon-hcd.c 	list_add_tail(&pipe->node, &usb->idle_pipes);
node             1355 drivers/staging/octeon-usb/octeon-hcd.c 				 node);
node             1500 drivers/staging/octeon-usb/octeon-hcd.c 				 node);
node             1847 drivers/staging/octeon-usb/octeon-hcd.c 	list_for_each_entry(pipe, list, node) {
node             1850 drivers/staging/octeon-usb/octeon-hcd.c 					 node);
node             1945 drivers/staging/octeon-usb/octeon-hcd.c 		list_for_each_entry(pipe, &usb->active_pipes[ttype], node) {
node             2092 drivers/staging/octeon-usb/octeon-hcd.c 	list_del(&transaction->node);
node             2094 drivers/staging/octeon-usb/octeon-hcd.c 		list_move_tail(&pipe->node, &usb->idle_pipes);
node             2160 drivers/staging/octeon-usb/octeon-hcd.c 		list_add_tail(&transaction->node, &pipe->transactions);
node             2162 drivers/staging/octeon-usb/octeon-hcd.c 		list_add_tail(&transaction->node, &pipe->transactions);
node             2163 drivers/staging/octeon-usb/octeon-hcd.c 		list_move_tail(&pipe->node,
node             2304 drivers/staging/octeon-usb/octeon-hcd.c 	if (list_first_entry(&pipe->transactions, typeof(*transaction), node) ==
node             2346 drivers/staging/octeon-usb/octeon-hcd.c 	list_for_each_entry_safe(transaction, next, &pipe->transactions, node) {
node             2371 drivers/staging/octeon-usb/octeon-hcd.c 	list_del(&pipe->node);
node             2678 drivers/staging/octeon-usb/octeon-hcd.c 				       node);
node              616 drivers/staging/octeon/ethernet.c 	struct device_node *node = NULL;
node              621 drivers/staging/octeon/ethernet.c 		node = of_get_next_child(parent, node);
node              622 drivers/staging/octeon/ethernet.c 		if (!node)
node              624 drivers/staging/octeon/ethernet.c 		addr = of_get_property(node, "reg", &size);
node              628 drivers/staging/octeon/ethernet.c 	return node;
node              447 drivers/staging/ralink-gdma/ralink-gdma.c 				list_del(&desc->vdesc.node);
node               78 drivers/staging/rtl8712/rtl871x_event.h 	unsigned char *node;
node              519 drivers/staging/rtl8723bs/hal/rtl8723b_cmd.c 		memcpy(pSSIDListPkt, &pwrctl->pno_ssid_list->node[i].SSID,
node               86 drivers/staging/rtl8723bs/include/rtw_event.h 	unsigned char *node;
node              179 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h 	struct pno_ssid	node[MAX_PNO_LIST_COUNT];
node              163 drivers/staging/speakup/spk_types.h 	struct list_head node;
node              372 drivers/staging/speakup/synth.c 	list_for_each_entry(tmp, &synths, node) {
node              455 drivers/staging/speakup/synth.c 	list_for_each_entry(tmp, &synths, node) {
node              466 drivers/staging/speakup/synth.c 		list_add_tail(&in_synth->node, &synths);
node              478 drivers/staging/speakup/synth.c 	list_del(&in_synth->node);
node              165 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
node              180 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
node              263 drivers/staging/uwb/beacon.c 	list_add(&bce->node, &rc->uwb_beca.list);
node              278 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
node              293 drivers/staging/uwb/beacon.c 	list_for_each_entry_safe(bce, next, &rc->uwb_beca.list, node) {
node              294 drivers/staging/uwb/beacon.c 		list_del(&bce->node);
node              352 drivers/staging/uwb/lc-dev.c 	list_del(&uwb_dev->bce->node);
node               20 drivers/staging/uwb/pal.c 	INIT_LIST_HEAD(&pal->node);
node               53 drivers/staging/uwb/pal.c 	list_add(&pal->node, &rc->pals);
node              105 drivers/staging/uwb/pal.c 	list_del(&pal->node);
node               35 drivers/staging/uwb/radio.c 	list_for_each_entry(pal, &rc->pals, node) {
node              172 drivers/staging/uwb/radio.c 	list_for_each_entry(pal, &rc->pals, node) {
node              760 drivers/staging/uwb/rsv.c 	list_for_each_entry(pal, &rc->pals, node) {
node              267 drivers/staging/uwb/uwb-internal.h 	struct list_head node;
node              440 drivers/staging/uwb/uwb.h 	struct list_head node;
node              206 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_debugfs.c 	struct vchiq_debugfs_node *node =
node              209 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_debugfs.c 	debugfs_remove_recursive(node->dentry);
node              774 drivers/target/iscsi/iscsi_target_stat.c CONFIGFS_ATTR_RO(iscsi_stat_sess_, node);
node              110 drivers/target/target_core_user.c 	struct list_head node;
node             1396 drivers/target/target_core_user.c 	INIT_LIST_HEAD(&udev->node);
node             1960 drivers/target/target_core_user.c 	list_add(&udev->node, &root_udev);
node             1997 drivers/target/target_core_user.c 	list_del(&udev->node);
node             2645 drivers/target/target_core_user.c 	list_for_each_entry(udev, &root_udev, node) {
node              141 drivers/tc/tc.c 		list_add_tail(&tdev->node, &tbus->devices);
node               89 drivers/thermal/cpu_cooling.c 	struct list_head node;
node              634 drivers/thermal/cpu_cooling.c 	list_add(&cpufreq_cdev->node, &cpufreq_cdev_list);
node              735 drivers/thermal/cpu_cooling.c 	list_del(&cpufreq_cdev->node);
node              352 drivers/thermal/qcom/qcom-spmi-temp-alarm.c 	struct device_node *node;
node              357 drivers/thermal/qcom/qcom-spmi-temp-alarm.c 	node = pdev->dev.of_node;
node              372 drivers/thermal/qcom/qcom-spmi-temp-alarm.c 	ret = of_property_read_u32(node, "reg", &res);
node              431 drivers/thermal/qcom/qcom-spmi-temp-alarm.c 					IRQF_ONESHOT, node->name, chip);
node              152 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node) {
node              193 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node) {
node              618 drivers/thermal/thermal_core.c 	list_for_each_entry(cdev, &thermal_cdev_list, node) {
node              638 drivers/thermal/thermal_core.c 	list_for_each_entry(cdev, &thermal_cdev_list, node) {
node              697 drivers/thermal/thermal_core.c 	list_for_each_entry(pos1, &thermal_tz_list, node) {
node              701 drivers/thermal/thermal_core.c 	list_for_each_entry(pos2, &thermal_cdev_list, node) {
node              902 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node) {
node              994 drivers/thermal/thermal_core.c 	list_add(&cdev->node, &thermal_cdev_list);
node             1001 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node)
node             1132 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_cdev_list, node)
node             1140 drivers/thermal/thermal_core.c 	list_del(&cdev->node);
node             1143 drivers/thermal/thermal_core.c 	list_for_each_entry(tz, &thermal_tz_list, node) {
node             1183 drivers/thermal/thermal_core.c 		list_for_each_entry(pos, &thermal_cdev_list, node) {
node             1194 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_cdev_list, node) {
node             1342 drivers/thermal/thermal_core.c 	list_add_tail(&tz->node, &thermal_tz_list);
node             1387 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node)
node             1395 drivers/thermal/thermal_core.c 	list_del(&tz->node);
node             1398 drivers/thermal/thermal_core.c 	list_for_each_entry(cdev, &thermal_cdev_list, node) {
node             1448 drivers/thermal/thermal_core.c 	list_for_each_entry(pos, &thermal_tz_list, node)
node             1574 drivers/thermal/thermal_core.c 		list_for_each_entry(tz, &thermal_tz_list, node) {
node               26 drivers/thermal/thermal_hwmon.c 	struct list_head node;
node               93 drivers/thermal/thermal_hwmon.c 	list_for_each_entry(hwmon, &thermal_hwmon_list, node) {
node              193 drivers/thermal/thermal_hwmon.c 		list_add_tail(&hwmon->node, &thermal_hwmon_list);
node              244 drivers/thermal/thermal_hwmon.c 	list_del(&hwmon->node);
node              820 drivers/thermal/ti-soc-thermal/ti-bandgap.c 	struct device_node *node = pdev->dev.of_node;
node              827 drivers/thermal/ti-soc-thermal/ti-bandgap.c 	if (!node) {
node              863 drivers/thermal/ti-soc-thermal/ti-bandgap.c 		bgp->tshut_gpio = of_get_gpio(node, 0);
node              915 drivers/tty/hvc/hvcs.c 	list_for_each_entry(pi, &head, node)
node              523 drivers/tty/serdev/core.c 	struct device_node *node;
node              528 drivers/tty/serdev/core.c 	for_each_available_child_of_node(ctrl->dev.of_node, node) {
node              529 drivers/tty/serdev/core.c 		if (!of_get_property(node, "compatible", NULL))
node              532 drivers/tty/serdev/core.c 		dev_dbg(&ctrl->dev, "adding child %pOF\n", node);
node              538 drivers/tty/serdev/core.c 		serdev->dev.of_node = node;
node               84 drivers/tty/serial/8250/8250_core.c 	struct			hlist_node node;
node              167 drivers/tty/serial/8250/8250_core.c 		hlist_del(&i->node);
node              184 drivers/tty/serial/8250/8250_core.c 		i = hlist_entry(n, struct irq_info, node);
node              197 drivers/tty/serial/8250/8250_core.c 		hlist_add_head(&i->node, h);
node              236 drivers/tty/serial/8250/8250_core.c 		i = hlist_entry(n, struct irq_info, node);
node              232 drivers/tty/serial/earlycon.c 			     unsigned long node,
node              243 drivers/tty/serial/earlycon.c 	addr = of_flat_dt_translate_address(node);
node              250 drivers/tty/serial/earlycon.c 	val = of_get_flat_dt_prop(node, "reg-offset", NULL);
node              255 drivers/tty/serial/earlycon.c 	val = of_get_flat_dt_prop(node, "reg-shift", NULL);
node              258 drivers/tty/serial/earlycon.c 	big_endian = of_get_flat_dt_prop(node, "big-endian", NULL) != NULL ||
node              260 drivers/tty/serial/earlycon.c 		 of_get_flat_dt_prop(node, "native-endian", NULL) != NULL);
node              261 drivers/tty/serial/earlycon.c 	val = of_get_flat_dt_prop(node, "reg-io-width", NULL);
node              279 drivers/tty/serial/earlycon.c 	val = of_get_flat_dt_prop(node, "current-speed", NULL);
node              283 drivers/tty/serial/earlycon.c 	val = of_get_flat_dt_prop(node, "clock-frequency", NULL);
node              821 drivers/tty/serial/lantiq.c 	struct device_node *node = pdev->dev.of_node;
node              848 drivers/tty/serial/lantiq.c 	line = of_alias_get_id(node, "serial");
node              737 drivers/tty/serial/pmac_zilog.c 			PMAC_FTR_SCC_ENABLE, uap->node, uap->port_type, 1);
node              741 drivers/tty/serial/pmac_zilog.c 				PMAC_FTR_MODEM_ENABLE, uap->node, 0, 1);
node              751 drivers/tty/serial/pmac_zilog.c 				PMAC_FTR_MODEM_ENABLE, uap->node, 0, 0);
node              754 drivers/tty/serial/pmac_zilog.c 		pmac_call_feature(PMAC_FTR_SCC_ENABLE, uap->node, uap->port_type, 0);
node             1399 drivers/tty/serial/pmac_zilog.c 	struct device_node *np = uap->node;
node             1532 drivers/tty/serial/pmac_zilog.c 	np = uap->node;
node             1536 drivers/tty/serial/pmac_zilog.c 	uap->node = NULL;
node             1552 drivers/tty/serial/pmac_zilog.c 		if (pmz_ports[i].node == mdev->ofdev.dev.of_node)
node             1569 drivers/tty/serial/pmac_zilog.c 		       uap->node);
node             1672 drivers/tty/serial/pmac_zilog.c 		pmz_ports[count].node		= node_a;
node             1673 drivers/tty/serial/pmac_zilog.c 		pmz_ports[count+1].node		= node_b;
node             1928 drivers/tty/serial/pmac_zilog.c 		if (uport->node != NULL)
node             2006 drivers/tty/serial/pmac_zilog.c 	if (uap->node == NULL)
node               27 drivers/tty/serial/pmac_zilog.h 	struct device_node		*node;
node             1662 drivers/tty/serial/serial-tegra.c 	struct device_node *node;
node             1666 drivers/tty/serial/serial-tegra.c 	node = of_find_matching_node(NULL, tegra_uart_of_match);
node             1667 drivers/tty/serial/serial-tegra.c 	if (node)
node             1668 drivers/tty/serial/serial-tegra.c 		match = of_match_node(tegra_uart_of_match, node);
node              575 drivers/tty/vcc.c 	u64 node;
node              620 drivers/tty/vcc.c 	node = vio_vdev_node(hp, vdev);
node              621 drivers/tty/vcc.c 	if (node == MDESC_NODE_NULL) {
node              627 drivers/tty/vcc.c 	domain = mdesc_get_property(hp, node, "vcc-domain-name", NULL);
node              121 drivers/tty/vt/vt.c 	int node;
node             4017 drivers/tty/vt/vt.c 			con_driver->node = i;
node             4032 drivers/tty/vt/vt.c 					  MKDEV(0, con_driver->node),
node             4034 drivers/tty/vt/vt.c 					  "vtcon%i", con_driver->node);
node             4112 drivers/tty/vt/vt.c 		device_destroy(vtconsole_class, MKDEV(0, con_driver->node));
node             4120 drivers/tty/vt/vt.c 		con_driver->node = 0;
node             4186 drivers/tty/vt/vt.c 							  MKDEV(0, con->node),
node             4188 drivers/tty/vt/vt.c 							  "vtcon%i", con->node);
node              250 drivers/uio/uio_fsl_elbc_gpcm.c static int get_of_data(struct fsl_elbc_gpcm *priv, struct device_node *node,
node              259 drivers/uio/uio_fsl_elbc_gpcm.c 	ret = of_address_to_resource(node, 0, res);
node              266 drivers/uio/uio_fsl_elbc_gpcm.c 	ret = of_property_read_u32(node, "reg", &priv->bank);
node              273 drivers/uio/uio_fsl_elbc_gpcm.c 	ret = of_property_read_u32(node, "elbc-gpcm-br", reg_br);
node              280 drivers/uio/uio_fsl_elbc_gpcm.c 	ret = of_property_read_u32(node, "elbc-gpcm-or", reg_or);
node              288 drivers/uio/uio_fsl_elbc_gpcm.c 	if (of_property_read_string(node, "device_type", &type) == 0)
node              292 drivers/uio/uio_fsl_elbc_gpcm.c 	*irq = irq_of_parse_and_map(node, 0);
node              300 drivers/uio/uio_fsl_elbc_gpcm.c 	if (of_property_read_string(node, "uio_name", &dt_name) != 0)
node              311 drivers/uio/uio_fsl_elbc_gpcm.c 	struct device_node *node = pdev->dev.of_node;
node              334 drivers/uio/uio_fsl_elbc_gpcm.c 	ret = get_of_data(priv, node, &res, &reg_br_new, &reg_or_new,
node              390 drivers/uio/uio_fsl_elbc_gpcm.c 	info->mem[0].name = kasprintf(GFP_KERNEL, "%pOFn", node);
node              105 drivers/uio/uio_pdrv_genirq.c 	struct device_node *node = pdev->dev.of_node;
node              111 drivers/uio/uio_pdrv_genirq.c 	if (node) {
node              122 drivers/uio/uio_pdrv_genirq.c 		if (!of_property_read_string(node, "linux,uio-name", &name))
node              126 drivers/uio/uio_pdrv_genirq.c 						       "%pOFn", node);
node               31 drivers/usb/c67x00/c67x00-sched.c 	struct list_head node;
node              254 drivers/usb/c67x00/c67x00-sched.c 	INIT_LIST_HEAD(&ep_data->node);
node              268 drivers/usb/c67x00/c67x00-sched.c 	if (list_empty(&ep_data->node)) {
node              269 drivers/usb/c67x00/c67x00-sched.c 		list_add(&ep_data->node, &c67x00->list[type]);
node              273 drivers/usb/c67x00/c67x00-sched.c 		list_for_each_entry(prev, &c67x00->list[type], node) {
node              276 drivers/usb/c67x00/c67x00-sched.c 				list_add(&ep_data->node, prev->node.prev);
node              297 drivers/usb/c67x00/c67x00-sched.c 	list_del(&ep_data->node);
node              783 drivers/usb/c67x00/c67x00-sched.c 	list_for_each_entry(ep_data, &c67x00->list[type], node) {
node              158 drivers/usb/chipidea/debug.c 	struct td_node *node, *tmpnode;
node              169 drivers/usb/chipidea/debug.c 			list_for_each_entry_safe(node, tmpnode, &req->tds, td) {
node              172 drivers/usb/chipidea/debug.c 					   (u32)node->dma,
node              178 drivers/usb/chipidea/debug.c 						   *((u32 *)node->ptr + j));
node              345 drivers/usb/chipidea/udc.c 	struct td_node *lastnode, *node = kzalloc(sizeof(struct td_node),
node              348 drivers/usb/chipidea/udc.c 	if (node == NULL)
node              351 drivers/usb/chipidea/udc.c 	node->ptr = dma_pool_zalloc(hwep->td_pool, GFP_ATOMIC, &node->dma);
node              352 drivers/usb/chipidea/udc.c 	if (node->ptr == NULL) {
node              353 drivers/usb/chipidea/udc.c 		kfree(node);
node              357 drivers/usb/chipidea/udc.c 	node->ptr->token = cpu_to_le32(length << __ffs(TD_TOTAL_BYTES));
node              358 drivers/usb/chipidea/udc.c 	node->ptr->token &= cpu_to_le32(TD_TOTAL_BYTES);
node              359 drivers/usb/chipidea/udc.c 	node->ptr->token |= cpu_to_le32(TD_STATUS_ACTIVE);
node              366 drivers/usb/chipidea/udc.c 		node->ptr->token |= cpu_to_le32(mul << __ffs(TD_MULTO));
node              371 drivers/usb/chipidea/udc.c 		node->ptr->page[0] = cpu_to_le32(temp);
node              375 drivers/usb/chipidea/udc.c 			node->ptr->page[i] = cpu_to_le32(page);
node              385 drivers/usb/chipidea/udc.c 		lastnode->ptr->next = cpu_to_le32(node->dma);
node              388 drivers/usb/chipidea/udc.c 	INIT_LIST_HEAD(&node->td);
node              389 drivers/usb/chipidea/udc.c 	list_add_tail(&node->td, &hwreq->tds);
node              529 drivers/usb/chipidea/udc.c 					   struct td_node *node)
node              531 drivers/usb/chipidea/udc.c 	hwep->qh.ptr->td.next = cpu_to_le32(node->dma);
node              549 drivers/usb/chipidea/udc.c 	struct td_node *node, *tmpnode;
node              559 drivers/usb/chipidea/udc.c 	list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) {
node              560 drivers/usb/chipidea/udc.c 		tmptoken = le32_to_cpu(node->ptr->token);
node              566 drivers/usb/chipidea/udc.c 					reprime_dtd(ci, hwep, node);
node              601 drivers/usb/chipidea/udc.c 		hwep->pending_td = node;
node              602 drivers/usb/chipidea/udc.c 		list_del_init(&node->td);
node              627 drivers/usb/chipidea/udc.c 	struct td_node *node, *tmpnode;
node              639 drivers/usb/chipidea/udc.c 		list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) {
node              640 drivers/usb/chipidea/udc.c 			dma_pool_free(hwep->td_pool, node->ptr, node->dma);
node              641 drivers/usb/chipidea/udc.c 			list_del_init(&node->td);
node              642 drivers/usb/chipidea/udc.c 			node->ptr = NULL;
node              643 drivers/usb/chipidea/udc.c 			kfree(node);
node             1359 drivers/usb/chipidea/udc.c 	struct td_node *node, *tmpnode;
node             1371 drivers/usb/chipidea/udc.c 	list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) {
node             1372 drivers/usb/chipidea/udc.c 		dma_pool_free(hwep->td_pool, node->ptr, node->dma);
node             1373 drivers/usb/chipidea/udc.c 		list_del_init(&node->td);
node             1374 drivers/usb/chipidea/udc.c 		node->ptr = NULL;
node             1375 drivers/usb/chipidea/udc.c 		kfree(node);
node             1418 drivers/usb/chipidea/udc.c 	struct td_node *node, *tmpnode;
node             1429 drivers/usb/chipidea/udc.c 	list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) {
node             1430 drivers/usb/chipidea/udc.c 		dma_pool_free(hwep->td_pool, node->ptr, node->dma);
node             1431 drivers/usb/chipidea/udc.c 		list_del(&node->td);
node             1432 drivers/usb/chipidea/udc.c 		kfree(node);
node              282 drivers/usb/common/common.c 	struct device_node *node;
node              285 drivers/usb/common/common.c 	node = of_parse_phandle(dev->of_node, "companion", 0);
node              286 drivers/usb/common/common.c 	if (node)
node              287 drivers/usb/common/common.c 		pdev = of_find_device_by_node(node);
node              289 drivers/usb/common/common.c 	of_node_put(node);
node               64 drivers/usb/core/driver.c 	INIT_LIST_HEAD(&dynid->node);
node               99 drivers/usb/core/driver.c 	list_add_tail(&dynid->node, &dynids->list);
node              119 drivers/usb/core/driver.c 	list_for_each_entry(dynid, &dynids->list, node)
node              164 drivers/usb/core/driver.c 	list_for_each_entry_safe(dynid, n, &usb_driver->dynids.list, node) {
node              169 drivers/usb/core/driver.c 			list_del(&dynid->node);
node              224 drivers/usb/core/driver.c 	list_for_each_entry_safe(dynid, n, &usb_drv->dynids.list, node) {
node              225 drivers/usb/core/driver.c 		list_del(&dynid->node);
node              237 drivers/usb/core/driver.c 	list_for_each_entry(dynid, &drv->dynids.list, node) {
node             2051 drivers/usb/core/message.c 	struct list_head	node;
node             2063 drivers/usb/core/message.c 	list_del(&req->node);
node             2081 drivers/usb/core/message.c 	list_for_each_entry(req, &set_config_list, node) {
node             2120 drivers/usb/core/message.c 	list_add(&req->node, &set_config_list);
node               27 drivers/usb/core/of.c 	struct device_node *node;
node               30 drivers/usb/core/of.c 	for_each_child_of_node(hub->dev.of_node, node) {
node               31 drivers/usb/core/of.c 		if (of_property_read_u32(node, "reg", &reg))
node               35 drivers/usb/core/of.c 			return node;
node               94 drivers/usb/core/of.c 	struct device_node *node;
node               97 drivers/usb/core/of.c 	for_each_child_of_node(udev->dev.of_node, node) {
node               98 drivers/usb/core/of.c 		if (of_property_read_u32_array(node, "reg", reg, 2))
node              102 drivers/usb/core/of.c 			return node;
node              134 drivers/usb/core/usb-acpi.c 	list_for_each_entry(adev, &parent->children, node) {
node             1082 drivers/usb/dwc3/core.c 	struct device_node	*node = dev->of_node;
node             1085 drivers/usb/dwc3/core.c 	if (node) {
node               53 drivers/usb/dwc3/dwc3-exynos.c 	struct device_node	*node = dev->of_node;
node              112 drivers/usb/dwc3/dwc3-exynos.c 	if (node) {
node              113 drivers/usb/dwc3/dwc3-exynos.c 		ret = of_platform_populate(node, NULL, NULL, dev);
node               82 drivers/usb/dwc3/dwc3-keystone.c 	struct device_node	*node = pdev->dev.of_node;
node              108 drivers/usb/dwc3/dwc3-keystone.c 	if (of_device_is_compatible(node, "ti,am654-dwc3"))
node              128 drivers/usb/dwc3/dwc3-keystone.c 	error = of_platform_populate(node, NULL, NULL, dev);
node              157 drivers/usb/dwc3/dwc3-keystone.c 	struct device_node *node = pdev->dev.of_node;
node              159 drivers/usb/dwc3/dwc3-keystone.c 	if (!of_device_is_compatible(node, "ti,am654-dwc3"))
node              370 drivers/usb/dwc3/dwc3-omap.c 	struct device_node	*node = omap->dev->of_node;
node              380 drivers/usb/dwc3/dwc3-omap.c 	if (of_device_is_compatible(node, "ti,am437x-dwc3")) {
node              392 drivers/usb/dwc3/dwc3-omap.c 	struct device_node	*node = omap->dev->of_node;
node              397 drivers/usb/dwc3/dwc3-omap.c 	of_property_read_u32(node, "utmi-mode", &utmi_mode);
node              416 drivers/usb/dwc3/dwc3-omap.c 	struct device_node	*node = omap->dev->of_node;
node              419 drivers/usb/dwc3/dwc3-omap.c 	if (of_property_read_bool(node, "extcon")) {
node              451 drivers/usb/dwc3/dwc3-omap.c 	struct device_node	*node = pdev->dev.of_node;
node              464 drivers/usb/dwc3/dwc3-omap.c 	if (!node) {
node              483 drivers/usb/dwc3/dwc3-omap.c 	if (of_property_read_bool(node, "vbus-supply")) {
node              513 drivers/usb/dwc3/dwc3-omap.c 	ret = of_platform_populate(node, NULL, NULL, dev);
node              200 drivers/usb/dwc3/dwc3-st.c 	struct device_node *node = dev->of_node, *child;
node              214 drivers/usb/dwc3/dwc3-st.c 	regmap = syscon_regmap_lookup_by_phandle(node, "st,syscfg");
node              254 drivers/usb/dwc3/dwc3-st.c 	child = of_get_child_by_name(node, "dwc3");
node              262 drivers/usb/dwc3/dwc3-st.c 	ret = of_platform_populate(node, NULL, NULL, dev);
node              911 drivers/usb/dwc3/gadget.c 		dma_addr_t dma, unsigned length, unsigned chain, unsigned node,
node              928 drivers/usb/dwc3/gadget.c 		if (!node) {
node             1024 drivers/usb/dwc3/gadget.c 		struct dwc3_request *req, unsigned chain, unsigned node)
node             1051 drivers/usb/dwc3/gadget.c 	__dwc3_prepare_one_trb(dep, trb, dma, length, chain, node,
node               33 drivers/usb/gadget/legacy/hid.c 	struct list_head node;
node              100 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry(e, &hidg_func_list, node) {
node              113 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry(n, &hidg_func_list, node) {
node              145 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry(n, &hidg_func_list, node) {
node              195 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry(m, &hidg_func_list, node) {
node              207 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry(n, &hidg_func_list, node) {
node              233 drivers/usb/gadget/legacy/hid.c 	list_add_tail(&entry->node, &hidg_func_list);
node              242 drivers/usb/gadget/legacy/hid.c 	list_for_each_entry_safe(e, n, &hidg_func_list, node) {
node              243 drivers/usb/gadget/legacy/hid.c 		list_del(&e->node);
node              181 drivers/usb/gadget/udc/fsl_qe_udc.h 	struct list_head node;
node              249 drivers/usb/gadget/udc/fsl_qe_udc.h 	INIT_LIST_HEAD(&(frm->node));
node              174 drivers/usb/host/fhci-hcd.c 	list_for_each_entry_safe(ed, next_ed, &fhci->empty_eds, node) {
node              175 drivers/usb/host/fhci-hcd.c 		list_del(&ed->node);
node              179 drivers/usb/host/fhci-hcd.c 	list_for_each_entry_safe(td, next_td, &fhci->empty_tds, node) {
node              180 drivers/usb/host/fhci-hcd.c 		list_del(&td->node);
node              563 drivers/usb/host/fhci-hcd.c 	struct device_node *node = dev->of_node;
node              579 drivers/usb/host/fhci-hcd.c 	sprop = of_get_property(node, "mode", NULL);
node              593 drivers/usb/host/fhci-hcd.c 	iprop = of_get_property(node, "hub-power-budget", &size);
node              598 drivers/usb/host/fhci-hcd.c 	ret = of_address_to_resource(node, 0, &usb_regs);
node              613 drivers/usb/host/fhci-hcd.c 	iprop = of_get_property(node, "reg", &size);
node              636 drivers/usb/host/fhci-hcd.c 		gpio = of_get_gpio_flags(node, i, &flags);
node              671 drivers/usb/host/fhci-hcd.c 		fhci->pins[j] = qe_pin_request(node, j);
node              695 drivers/usb/host/fhci-hcd.c 	usb_irq = irq_of_parse_and_map(node, 0);
node              703 drivers/usb/host/fhci-hcd.c 	sprop = of_get_property(node, "fsl,fullspeed-clock", NULL);
node              713 drivers/usb/host/fhci-hcd.c 	sprop = of_get_property(node, "fsl,lowspeed-clock", NULL);
node               26 drivers/usb/host/fhci-mem.c 	INIT_LIST_HEAD(&td->node);
node               34 drivers/usb/host/fhci-mem.c 	INIT_LIST_HEAD(&ed->node);
node               42 drivers/usb/host/fhci-mem.c 		td = list_entry(fhci->empty_tds.next, struct td, node);
node               58 drivers/usb/host/fhci-mem.c 	list_add(&td->node, &fhci->empty_tds);
node               66 drivers/usb/host/fhci-mem.c 		ed = list_entry(fhci->empty_eds.next, struct ed, node);
node               82 drivers/usb/host/fhci-mem.c 	list_add(&ed->node, &fhci->empty_eds);
node               64 drivers/usb/host/fhci-q.c 		list_add_tail(&td->node, &ed->td_list);
node               75 drivers/usb/host/fhci-q.c 		td = list_entry(ed->td_list.next, struct td, node);
node              112 drivers/usb/host/fhci-q.c 		td = list_entry(ed->td_list.next, struct td, node);
node              118 drivers/usb/host/fhci-q.c 						 node);
node              132 drivers/usb/host/fhci-q.c 		td = list_entry(p_list->done_list.next, struct td, node);
node              145 drivers/usb/host/fhci-q.c 	list_del_init(&td->node);
node              149 drivers/usb/host/fhci-q.c 		ed->td_head = list_entry(ed->td_list.next, struct td, node);
node              155 drivers/usb/host/fhci-q.c 	list_add_tail(&td->node, &usb->hc_list->done_list);
node              168 drivers/usb/host/fhci-q.c 		list_del_init(&urb_priv->tds[i]->node);
node              174 drivers/usb/host/fhci-q.c 		ed->td_head = list_entry(ed->td_list.next, struct td, node);
node              184 drivers/usb/host/fhci-q.c 		list_del_init(&ed->node);
node               66 drivers/usb/host/fhci-sched.c 				       node);
node              257 drivers/usb/host/fhci-sched.c 	struct list_head *node = list->next;
node              260 drivers/usb/host/fhci-sched.c 		list_move_tail(node, list);
node              284 drivers/usb/host/fhci-sched.c 	list_for_each_entry(ed, list, node) {
node              874 drivers/usb/host/fhci-sched.c 			list_add(&ed->node, &fhci->hc_list->ctrl_list);
node              877 drivers/usb/host/fhci-sched.c 			list_add(&ed->node, &fhci->hc_list->bulk_list);
node              880 drivers/usb/host/fhci-sched.c 			list_add(&ed->node, &fhci->hc_list->intr_list);
node              883 drivers/usb/host/fhci-sched.c 			list_add(&ed->node, &fhci->hc_list->iso_list);
node              331 drivers/usb/host/fhci.h 	struct list_head node;
node              352 drivers/usb/host/fhci.h 	struct list_head node;
node              221 drivers/usb/host/uhci-debug.c 				struct urb_priv, node);
node              229 drivers/usb/host/uhci-debug.c 		list_for_each_entry(urbp, &qh->queue, node) {
node              506 drivers/usb/host/uhci-debug.c 		head = &qh->node;
node              510 drivers/usb/host/uhci-debug.c 			qh = list_entry(tmp, struct uhci_qh, node);
node              159 drivers/usb/host/uhci-hcd.h 	struct list_head node;		/* Node in the list of QHs */
node              484 drivers/usb/host/uhci-hcd.h 	struct list_head node;		/* Node in the QH's urbp list */
node               55 drivers/usb/host/uhci-q.c 	lqh = list_entry(uhci->skel_async_qh->node.prev,
node               56 drivers/usb/host/uhci-q.c 			struct uhci_qh, node);
node               67 drivers/usb/host/uhci-q.c 	lqh = list_entry(uhci->skel_async_qh->node.prev,
node               68 drivers/usb/host/uhci-q.c 			struct uhci_qh, node);
node              261 drivers/usb/host/uhci-q.c 	INIT_LIST_HEAD(&qh->node);
node              298 drivers/usb/host/uhci-q.c 	list_del(&qh->node);
node              334 drivers/usb/host/uhci-q.c 	if (qh->queue.next != &urbp->node) {
node              338 drivers/usb/host/uhci-q.c 		purbp = list_entry(urbp->node.prev, struct urb_priv, node);
node              383 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
node              393 drivers/usb/host/uhci-q.c 	urbp = list_prepare_entry(urbp, &qh->queue, node);
node              394 drivers/usb/host/uhci-q.c 	list_for_each_entry_continue(urbp, &qh->queue, node) {
node              415 drivers/usb/host/uhci-q.c 	pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe;
node              426 drivers/usb/host/uhci-q.c 	list_add_tail(&qh->node, &uhci->skel_iso_qh->node);
node              439 drivers/usb/host/uhci-q.c 	list_add_tail(&qh->node, &uhci->skelqh[qh->skel]->node);
node              441 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
node              459 drivers/usb/host/uhci-q.c 	list_for_each_entry_reverse(pqh, &uhci->skel_async_qh->node, node) {
node              463 drivers/usb/host/uhci-q.c 	list_add(&qh->node, &pqh->node);
node              488 drivers/usb/host/uhci-q.c 				struct urb_priv, node);
node              506 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
node              507 drivers/usb/host/uhci-q.c 				node);
node              508 drivers/usb/host/uhci-q.c 	list_del(&qh->node);
node              525 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
node              538 drivers/usb/host/uhci-q.c 	pqh = list_entry(qh->node.prev, struct uhci_qh, node);
node              570 drivers/usb/host/uhci-q.c 	if (list_empty(&uhci->skel_unlink_qh->node) || uhci->is_stopped)
node              575 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
node              576 drivers/usb/host/uhci-q.c 				node);
node              577 drivers/usb/host/uhci-q.c 	list_move_tail(&qh->node, &uhci->skel_unlink_qh->node);
node              591 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(qh->node.next, struct uhci_qh,
node              592 drivers/usb/host/uhci-q.c 				node);
node              593 drivers/usb/host/uhci-q.c 	list_move(&qh->node, &uhci->idle_qh_list);
node              734 drivers/usb/host/uhci-q.c 	INIT_LIST_HEAD(&urbp->node);
node              745 drivers/usb/host/uhci-q.c 	if (!list_empty(&urbp->node))
node             1298 drivers/usb/host/uhci-q.c 					struct urb_priv, node)->urb;
node             1456 drivers/usb/host/uhci-q.c 	list_add_tail(&urbp->node, &qh->queue);
node             1462 drivers/usb/host/uhci-q.c 	if (qh->queue.next == &urbp->node && !qh->is_stopped) {
node             1534 drivers/usb/host/uhci-q.c 			urbp->node.prev == &qh->queue &&
node             1535 drivers/usb/host/uhci-q.c 			urbp->node.next != &qh->queue) {
node             1536 drivers/usb/host/uhci-q.c 		struct urb *nurb = list_entry(urbp->node.next,
node             1537 drivers/usb/host/uhci-q.c 				struct urb_priv, node)->urb;
node             1545 drivers/usb/host/uhci-q.c 	list_del_init(&urbp->node);
node             1582 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
node             1615 drivers/usb/host/uhci-q.c 	list_for_each_entry(urbp, &qh->queue, node) {
node             1641 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
node             1691 drivers/usb/host/uhci-q.c 		urbp = list_entry(qh->queue.next, struct urb_priv, node);
node             1762 drivers/usb/host/uhci-q.c 		uhci->next_qh = list_entry(uhci->skelqh[i]->node.next,
node             1763 drivers/usb/host/uhci-q.c 				struct uhci_qh, node);
node             1765 drivers/usb/host/uhci-q.c 			uhci->next_qh = list_entry(qh->node.next,
node             1766 drivers/usb/host/uhci-q.c 					struct uhci_qh, node);
node             1772 drivers/usb/host/uhci-q.c 	list_entry(qh->queue.next, struct urb_priv, node));
node             1789 drivers/usb/host/uhci-q.c 	if (list_empty(&uhci->skel_unlink_qh->node))
node              428 drivers/usb/host/xhci-mtk.c 	struct device_node *node = dev->of_node;
node              462 drivers/usb/host/xhci-mtk.c 	mtk->lpm_support = of_property_read_bool(node, "usb3-lpm-capable");
node              464 drivers/usb/host/xhci-mtk.c 	of_property_read_u32(node, "mediatek,u3p-dis-msk",
node              467 drivers/usb/host/xhci-mtk.c 	ret = usb_wakeup_of_property_parse(mtk, node);
node              102 drivers/usb/host/xhci-rcar.c 	struct device_node *node = dev->of_node;
node              104 drivers/usb/host/xhci-rcar.c 	return of_device_is_compatible(node, "renesas,xhci-r8a7790") ||
node              105 drivers/usb/host/xhci-rcar.c 		of_device_is_compatible(node, "renesas,xhci-r8a7791") ||
node              106 drivers/usb/host/xhci-rcar.c 		of_device_is_compatible(node, "renesas,xhci-r8a7793") ||
node              107 drivers/usb/host/xhci-rcar.c 		of_device_is_compatible(node, "renesas,rcar-gen2-xhci");
node              216 drivers/usb/mtu3/mtu3_plat.c 	struct device_node *node = pdev->dev.of_node;
node              247 drivers/usb/mtu3/mtu3_plat.c 	ssusb->num_phys = of_count_phandle_with_args(node,
node              259 drivers/usb/mtu3/mtu3_plat.c 		ssusb->phys[i] = devm_of_phy_get_by_index(dev, node, i);
node              279 drivers/usb/mtu3/mtu3_plat.c 	ret = ssusb_wakeup_of_property_parse(ssusb, node);
node              286 drivers/usb/mtu3/mtu3_plat.c 	of_property_read_u32(node, "mediatek,u3p-dis-msk",
node              299 drivers/usb/mtu3/mtu3_plat.c 	otg_sx->is_u3_drd = of_property_read_bool(node, "mediatek,usb3-drd");
node              301 drivers/usb/mtu3/mtu3_plat.c 		of_property_read_bool(node, "enable-manual-drd");
node              302 drivers/usb/mtu3/mtu3_plat.c 	otg_sx->role_sw_used = of_property_read_bool(node, "usb-role-switch");
node              304 drivers/usb/mtu3/mtu3_plat.c 	if (!otg_sx->role_sw_used && of_property_read_bool(node, "extcon")) {
node              322 drivers/usb/mtu3/mtu3_plat.c 	struct device_node *node = pdev->dev.of_node;
node              375 drivers/usb/mtu3/mtu3_plat.c 		ret = ssusb_host_init(ssusb, node);
node              388 drivers/usb/mtu3/mtu3_plat.c 		ret = ssusb_host_init(ssusb, node);
node             2058 drivers/usb/musb/musb_core.c 	struct list_head node;
node             2073 drivers/usb/musb/musb_core.c 	list_for_each_entry_safe(w, _w, &musb->pending_list, node) {
node             2082 drivers/usb/musb/musb_core.c 		list_del(&w->node);
node             2121 drivers/usb/musb/musb_core.c 		list_add_tail(&w->node, &musb->pending_list);
node              123 drivers/usb/phy/phy-am335x-control.c 	const struct device_node *node = (const struct device_node *)data;
node              124 drivers/usb/phy/phy-am335x-control.c 	return dev->of_node == node &&
node              130 drivers/usb/phy/phy-am335x-control.c 	struct device_node *node;
node              133 drivers/usb/phy/phy-am335x-control.c 	node = of_parse_phandle(dev->of_node, "ti,ctrl_mod", 0);
node              134 drivers/usb/phy/phy-am335x-control.c 	if (!node)
node              137 drivers/usb/phy/phy-am335x-control.c 	dev = bus_find_device(&platform_bus_type, NULL, node, match);
node              138 drivers/usb/phy/phy-am335x-control.c 	of_node_put(node);
node              217 drivers/usb/phy/phy-generic.c 		struct device_node *node = dev->of_node;
node              219 drivers/usb/phy/phy-generic.c 		if (of_property_read_u32(node, "clock-frequency", &clk_rate))
node              222 drivers/usb/phy/phy-generic.c 		needs_vcc = of_property_read_bool(node, "vcc-supply");
node              223 drivers/usb/phy/phy-generic.c 		needs_clk = of_property_read_bool(node, "clocks");
node              147 drivers/usb/phy/phy-isp1301.c 	const struct device_node *node = (const struct device_node *)data;
node              148 drivers/usb/phy/phy-isp1301.c 	return (dev->of_node == node) &&
node              152 drivers/usb/phy/phy-isp1301.c struct i2c_client *isp1301_get_client(struct device_node *node)
node              154 drivers/usb/phy/phy-isp1301.c 	if (node) { /* reference of ISP1301 I2C node via DT */
node              156 drivers/usb/phy/phy-isp1301.c 						     node, match);
node               52 drivers/usb/phy/phy.c static struct usb_phy *__of_usb_find_phy(struct device_node *node)
node               56 drivers/usb/phy/phy.c 	if (!of_device_is_available(node))
node               60 drivers/usb/phy/phy.c 		if (node != phy->dev->of_node)
node              487 drivers/usb/phy/phy.c 					  struct device_node *node,
node              502 drivers/usb/phy/phy.c 	phy = __of_usb_find_phy(node);
node              548 drivers/usb/phy/phy.c 	struct device_node *node;
node              556 drivers/usb/phy/phy.c 	node = of_parse_phandle(dev->of_node, phandle, index);
node              557 drivers/usb/phy/phy.c 	if (!node) {
node              562 drivers/usb/phy/phy.c 	phy = devm_usb_get_phy_by_node(dev, node, NULL);
node              563 drivers/usb/phy/phy.c 	of_node_put(node);
node               24 drivers/usb/renesas_usbhs/fifo.c 	INIT_LIST_HEAD(&pkt->node);
node               67 drivers/usb/renesas_usbhs/fifo.c 	list_move_tail(&pkt->node, &pipe->list);
node               89 drivers/usb/renesas_usbhs/fifo.c 	list_del_init(&pkt->node);
node               94 drivers/usb/renesas_usbhs/fifo.c 	return list_first_entry_or_null(&pipe->list, struct usbhs_pkt, node);
node               46 drivers/usb/renesas_usbhs/fifo.h 	struct list_head node;
node              678 drivers/usb/renesas_usbhs/mod_gadget.c 	WARN_ON(!list_empty(&ureq->pkt.node));
node              157 drivers/usb/serial/bus.c 	list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) {
node              158 drivers/usb/serial/bus.c 		list_del(&dynid->node);
node              626 drivers/usb/serial/usb-serial.c 	list_for_each_entry(dynid, &drv->dynids.list, node) {
node             1621 drivers/usb/typec/tcpm/fusb302.c 	struct device_node *node;
node             1624 drivers/usb/typec/tcpm/fusb302.c 	node = chip->dev->of_node;
node             1625 drivers/usb/typec/tcpm/fusb302.c 	chip->gpio_int_n = of_get_named_gpio(node, "fcs,int_n", 0);
node               15 drivers/usb/usbip/usbip_event.c 	struct list_head node;
node               48 drivers/usb/usbip/usbip_event.c 		ue = list_first_entry(&event_list, struct usbip_event, node);
node               49 drivers/usb/usbip/usbip_event.c 		list_del(&ue->node);
node              155 drivers/usb/usbip/usbip_event.c 	list_for_each_entry_reverse(ue, &event_list, node) {
node              166 drivers/usb/usbip/usbip_event.c 	list_add_tail(&ue->node, &event_list);
node              722 drivers/vfio/platform/vfio_platform_common.c void __vfio_platform_register_reset(struct vfio_platform_reset_node *node)
node              725 drivers/vfio/platform/vfio_platform_common.c 	list_add(&node->link, &reset_list);
node               84 drivers/vfio/vfio_iommu_type1.c 	struct rb_node		node;
node              111 drivers/vfio/vfio_iommu_type1.c 	struct rb_node		node;
node              137 drivers/vfio/vfio_iommu_type1.c 	struct rb_node *node = iommu->dma_list.rb_node;
node              139 drivers/vfio/vfio_iommu_type1.c 	while (node) {
node              140 drivers/vfio/vfio_iommu_type1.c 		struct vfio_dma *dma = rb_entry(node, struct vfio_dma, node);
node              143 drivers/vfio/vfio_iommu_type1.c 			node = node->rb_left;
node              145 drivers/vfio/vfio_iommu_type1.c 			node = node->rb_right;
node              160 drivers/vfio/vfio_iommu_type1.c 		dma = rb_entry(parent, struct vfio_dma, node);
node              168 drivers/vfio/vfio_iommu_type1.c 	rb_link_node(&new->node, parent, link);
node              169 drivers/vfio/vfio_iommu_type1.c 	rb_insert_color(&new->node, &iommu->dma_list);
node              174 drivers/vfio/vfio_iommu_type1.c 	rb_erase(&old->node, &iommu->dma_list);
node              183 drivers/vfio/vfio_iommu_type1.c 	struct rb_node *node = dma->pfn_list.rb_node;
node              185 drivers/vfio/vfio_iommu_type1.c 	while (node) {
node              186 drivers/vfio/vfio_iommu_type1.c 		vpfn = rb_entry(node, struct vfio_pfn, node);
node              189 drivers/vfio/vfio_iommu_type1.c 			node = node->rb_left;
node              191 drivers/vfio/vfio_iommu_type1.c 			node = node->rb_right;
node              207 drivers/vfio/vfio_iommu_type1.c 		vpfn = rb_entry(parent, struct vfio_pfn, node);
node              215 drivers/vfio/vfio_iommu_type1.c 	rb_link_node(&new->node, parent, link);
node              216 drivers/vfio/vfio_iommu_type1.c 	rb_insert_color(&new->node, &dma->pfn_list);
node              221 drivers/vfio/vfio_iommu_type1.c 	rb_erase(&old->node, &dma->pfn_list);
node             1057 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *node;
node             1059 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry(node, iova, list) {
node             1060 drivers/vfio/vfio_iommu_type1.c 		if (start >= node->start && end <= node->end)
node             1203 drivers/vfio/vfio_iommu_type1.c 		dma = rb_entry(n, struct vfio_dma, node);
node             1497 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *node, *next;
node             1503 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(node, next, iova, list) {
node             1504 drivers/vfio/vfio_iommu_type1.c 		if (start < node->start)
node             1506 drivers/vfio/vfio_iommu_type1.c 		if (start >= node->start && start < node->end) {
node             1507 drivers/vfio/vfio_iommu_type1.c 			node->start = start;
node             1511 drivers/vfio/vfio_iommu_type1.c 		list_del(&node->list);
node             1512 drivers/vfio/vfio_iommu_type1.c 		kfree(node);
node             1516 drivers/vfio/vfio_iommu_type1.c 	list_for_each_entry_safe(node, next, iova, list) {
node             1517 drivers/vfio/vfio_iommu_type1.c 		if (end > node->end)
node             1519 drivers/vfio/vfio_iommu_type1.c 		if (end > node->start && end <= node->end) {
node             1520 drivers/vfio/vfio_iommu_type1.c 			node->end = end;
node             1524 drivers/vfio/vfio_iommu_type1.c 		list_del(&node->list);
node             1525 drivers/vfio/vfio_iommu_type1.c 		kfree(node);
node             1856 drivers/vfio/vfio_iommu_type1.c 	struct rb_node *node;
node             1858 drivers/vfio/vfio_iommu_type1.c 	while ((node = rb_first(&iommu->dma_list)))
node             1859 drivers/vfio/vfio_iommu_type1.c 		vfio_remove_dma(iommu, rb_entry(node, struct vfio_dma, node));
node             1871 drivers/vfio/vfio_iommu_type1.c 		dma = rb_entry(n, struct vfio_dma, node);
node             1876 drivers/vfio/vfio_iommu_type1.c 							 node);
node             1893 drivers/vfio/vfio_iommu_type1.c 		dma = rb_entry(n, struct vfio_dma, node);
node             1912 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *node;
node             1929 drivers/vfio/vfio_iommu_type1.c 	node = list_first_entry(iova_copy, struct vfio_iova, list);
node             1930 drivers/vfio/vfio_iommu_type1.c 	node->start = start;
node             1931 drivers/vfio/vfio_iommu_type1.c 	node = list_last_entry(iova_copy, struct vfio_iova, list);
node             1932 drivers/vfio/vfio_iommu_type1.c 	node->end = end;
node             1946 drivers/vfio/vfio_iommu_type1.c 	struct vfio_iova *node;
node             1963 drivers/vfio/vfio_iommu_type1.c 	node = list_first_entry(iova_copy, struct vfio_iova, list);
node             1964 drivers/vfio/vfio_iommu_type1.c 	start = node->start;
node             1965 drivers/vfio/vfio_iommu_type1.c 	node = list_last_entry(iova_copy, struct vfio_iova, list);
node             1966 drivers/vfio/vfio_iommu_type1.c 	end = node->end;
node              265 drivers/vhost/vhost.c 		llist_add(&work->node, &dev->work_list);
node              335 drivers/vhost/vhost.c 	struct llist_node *node;
node              350 drivers/vhost/vhost.c 		node = llist_del_all(&dev->work_list);
node              351 drivers/vhost/vhost.c 		if (!node)
node              354 drivers/vhost/vhost.c 		node = llist_reverse_order(node);
node              357 drivers/vhost/vhost.c 		llist_for_each_entry_safe(work, work_next, node, node) {
node              617 drivers/vhost/vhost.c 			    struct vhost_umem_node *node)
node              619 drivers/vhost/vhost.c 	vhost_umem_interval_tree_remove(node, &umem->umem_tree);
node              620 drivers/vhost/vhost.c 	list_del(&node->link);
node              621 drivers/vhost/vhost.c 	kfree(node);
node              627 drivers/vhost/vhost.c 	struct vhost_umem_node *node, *tmp;
node              632 drivers/vhost/vhost.c 	list_for_each_entry_safe(node, tmp, &umem->umem_list, link)
node              633 drivers/vhost/vhost.c 		vhost_umem_free(umem, node);
node              640 drivers/vhost/vhost.c 	struct vhost_msg_node *node, *n;
node              644 drivers/vhost/vhost.c 	list_for_each_entry_safe(node, n, &dev->read_list, node) {
node              645 drivers/vhost/vhost.c 		list_del(&node->node);
node              646 drivers/vhost/vhost.c 		kfree(node);
node              649 drivers/vhost/vhost.c 	list_for_each_entry_safe(node, n, &dev->pending_list, node) {
node              650 drivers/vhost/vhost.c 		list_del(&node->node);
node              651 drivers/vhost/vhost.c 		kfree(node);
node              715 drivers/vhost/vhost.c 	struct vhost_umem_node *node;
node              720 drivers/vhost/vhost.c 	list_for_each_entry(node, &umem->umem_list, link) {
node              721 drivers/vhost/vhost.c 		unsigned long a = node->userspace_addr;
node              723 drivers/vhost/vhost.c 		if (vhost_overflow(node->userspace_addr, node->size))
node              728 drivers/vhost/vhost.c 				    node->size))
node              731 drivers/vhost/vhost.c 						   node->start,
node              732 drivers/vhost/vhost.c 						   node->size))
node              742 drivers/vhost/vhost.c 	const struct vhost_umem_node *node = vq->meta_iotlb[type];
node              744 drivers/vhost/vhost.c 	if (!node)
node              747 drivers/vhost/vhost.c 	return (void *)(uintptr_t)(node->userspace_addr + addr - node->start);
node             1021 drivers/vhost/vhost.c 	struct vhost_umem_node *tmp, *node;
node             1026 drivers/vhost/vhost.c 	node = kmalloc(sizeof(*node), GFP_ATOMIC);
node             1027 drivers/vhost/vhost.c 	if (!node)
node             1035 drivers/vhost/vhost.c 	node->start = start;
node             1036 drivers/vhost/vhost.c 	node->size = size;
node             1037 drivers/vhost/vhost.c 	node->last = end;
node             1038 drivers/vhost/vhost.c 	node->userspace_addr = userspace_addr;
node             1039 drivers/vhost/vhost.c 	node->perm = perm;
node             1040 drivers/vhost/vhost.c 	INIT_LIST_HEAD(&node->link);
node             1041 drivers/vhost/vhost.c 	list_add_tail(&node->link, &umem->umem_list);
node             1042 drivers/vhost/vhost.c 	vhost_umem_interval_tree_insert(node, &umem->umem_tree);
node             1051 drivers/vhost/vhost.c 	struct vhost_umem_node *node;
node             1053 drivers/vhost/vhost.c 	while ((node = vhost_umem_interval_tree_iter_first(&umem->umem_tree,
node             1055 drivers/vhost/vhost.c 		vhost_umem_free(umem, node);
node             1061 drivers/vhost/vhost.c 	struct vhost_msg_node *node, *n;
node             1065 drivers/vhost/vhost.c 	list_for_each_entry_safe(node, n, &d->pending_list, node) {
node             1066 drivers/vhost/vhost.c 		struct vhost_iotlb_msg *vq_msg = &node->msg.iotlb;
node             1070 drivers/vhost/vhost.c 			vhost_poll_queue(&node->vq->poll);
node             1071 drivers/vhost/vhost.c 			list_del(&node->node);
node             1072 drivers/vhost/vhost.c 			kfree(node);
node             1205 drivers/vhost/vhost.c 	struct vhost_msg_node *node;
node             1217 drivers/vhost/vhost.c 		node = vhost_dequeue_msg(dev, &dev->read_list);
node             1218 drivers/vhost/vhost.c 		if (node)
node             1239 drivers/vhost/vhost.c 	if (node) {
node             1241 drivers/vhost/vhost.c 		void *start = &node->msg;
node             1243 drivers/vhost/vhost.c 		switch (node->msg.type) {
node             1245 drivers/vhost/vhost.c 			size = sizeof(node->msg);
node             1246 drivers/vhost/vhost.c 			msg = &node->msg.iotlb;
node             1249 drivers/vhost/vhost.c 			size = sizeof(node->msg_v2);
node             1250 drivers/vhost/vhost.c 			msg = &node->msg_v2.iotlb;
node             1259 drivers/vhost/vhost.c 			kfree(node);
node             1262 drivers/vhost/vhost.c 		vhost_enqueue_msg(dev, &dev->pending_list, node);
node             1272 drivers/vhost/vhost.c 	struct vhost_msg_node *node;
node             1276 drivers/vhost/vhost.c 	node = vhost_new_msg(vq, v2 ? VHOST_IOTLB_MSG_V2 : VHOST_IOTLB_MSG);
node             1277 drivers/vhost/vhost.c 	if (!node)
node             1281 drivers/vhost/vhost.c 		node->msg_v2.type = VHOST_IOTLB_MSG_V2;
node             1282 drivers/vhost/vhost.c 		msg = &node->msg_v2.iotlb;
node             1284 drivers/vhost/vhost.c 		msg = &node->msg.iotlb;
node             1291 drivers/vhost/vhost.c 	vhost_enqueue_msg(dev, &dev->read_list, node);
node             1308 drivers/vhost/vhost.c 				 const struct vhost_umem_node *node,
node             1314 drivers/vhost/vhost.c 	if (likely(node->perm & access))
node             1315 drivers/vhost/vhost.c 		vq->meta_iotlb[type] = node;
node             1321 drivers/vhost/vhost.c 	const struct vhost_umem_node *node;
node             1329 drivers/vhost/vhost.c 		node = vhost_umem_interval_tree_iter_first(&umem->umem_tree,
node             1332 drivers/vhost/vhost.c 		if (node == NULL || node->start > addr) {
node             1335 drivers/vhost/vhost.c 		} else if (!(node->perm & access)) {
node             1342 drivers/vhost/vhost.c 		size = node->size - addr + node->start;
node             1345 drivers/vhost/vhost.c 			vhost_vq_meta_update(vq, node, type);
node             2043 drivers/vhost/vhost.c 	const struct vhost_umem_node *node;
node             2057 drivers/vhost/vhost.c 		node = vhost_umem_interval_tree_iter_first(&umem->umem_tree,
node             2059 drivers/vhost/vhost.c 		if (node == NULL || node->start > addr) {
node             2066 drivers/vhost/vhost.c 		} else if (!(node->perm & access)) {
node             2072 drivers/vhost/vhost.c 		size = node->size - addr + node->start;
node             2075 drivers/vhost/vhost.c 			(node->userspace_addr + addr - node->start);
node             2580 drivers/vhost/vhost.c 	struct vhost_msg_node *node = kmalloc(sizeof *node, GFP_KERNEL);
node             2581 drivers/vhost/vhost.c 	if (!node)
node             2585 drivers/vhost/vhost.c 	memset(&node->msg, 0, sizeof node->msg);
node             2586 drivers/vhost/vhost.c 	node->vq = vq;
node             2587 drivers/vhost/vhost.c 	node->msg.type = type;
node             2588 drivers/vhost/vhost.c 	return node;
node             2593 drivers/vhost/vhost.c 		       struct vhost_msg_node *node)
node             2596 drivers/vhost/vhost.c 	list_add_tail(&node->node, head);
node             2606 drivers/vhost/vhost.c 	struct vhost_msg_node *node = NULL;
node             2610 drivers/vhost/vhost.c 		node = list_first_entry(head, struct vhost_msg_node,
node             2611 drivers/vhost/vhost.c 					node);
node             2612 drivers/vhost/vhost.c 		list_del(&node->node);
node             2616 drivers/vhost/vhost.c 	return node;
node               21 drivers/vhost/vhost.h 	struct llist_node	  node;
node               55 drivers/vhost/vhost.h #define START(node) ((node)->start)
node               56 drivers/vhost/vhost.h #define LAST(node) ((node)->last)
node              156 drivers/vhost/vhost.h   struct list_head node;
node              220 drivers/vhost/vhost.h 		       struct vhost_msg_node *node);
node              178 drivers/video/backlight/arcxcnn_bl.c 	struct device_node *node = dev->of_node;
node              183 drivers/video/backlight/arcxcnn_bl.c 	if (!node)
node              186 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_string(node, "label", &lp->pdata->name);
node              190 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "default-brightness", &prog_val);
node              194 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,led-config-0", &prog_val);
node              198 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,led-config-1", &prog_val);
node              202 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,dim-freq", &prog_val);
node              206 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,comp-config", &prog_val);
node              210 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,filter-config", &prog_val);
node              214 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_read_u32(node, "arc,trim-config", &prog_val);
node              218 drivers/video/backlight/arcxcnn_bl.c 	ret = of_property_count_u32_elems(node, "led-sources");
node              226 drivers/video/backlight/arcxcnn_bl.c 		ret = of_property_read_u32_array(node, "led-sources", sources,
node               52 drivers/video/backlight/backlight.c 	int node = evdata->info->node;
node               66 drivers/video/backlight/backlight.c 			    !bd->fb_bl_on[node]) {
node               67 drivers/video/backlight/backlight.c 				bd->fb_bl_on[node] = true;
node               74 drivers/video/backlight/backlight.c 				   bd->fb_bl_on[node]) {
node               75 drivers/video/backlight/backlight.c 				bd->fb_bl_on[node] = false;
node              592 drivers/video/backlight/backlight.c struct backlight_device *of_find_backlight_by_node(struct device_node *node)
node              596 drivers/video/backlight/backlight.c 	dev = class_find_device(backlight_class, NULL, node, of_parent_match);
node               82 drivers/video/backlight/gpio_backlight.c 	struct device_node *node = gbl->dev->of_node;
node               85 drivers/video/backlight/gpio_backlight.c 	if (!node || !node->phandle)
node              374 drivers/video/backlight/lm3630a_bl.c static int lm3630a_parse_led_sources(struct fwnode_handle *node,
node              380 drivers/video/backlight/lm3630a_bl.c 	num_sources = fwnode_property_count_u32(node, "led-sources");
node              386 drivers/video/backlight/lm3630a_bl.c 	ret = fwnode_property_read_u32_array(node, "led-sources", sources,
node              402 drivers/video/backlight/lm3630a_bl.c 			      struct fwnode_handle *node, int *seen_led_sources)
node              409 drivers/video/backlight/lm3630a_bl.c 	ret = fwnode_property_read_u32(node, "reg", &bank);
node              416 drivers/video/backlight/lm3630a_bl.c 	led_sources = lm3630a_parse_led_sources(node, BIT(bank));
node              425 drivers/video/backlight/lm3630a_bl.c 	linear = fwnode_property_read_bool(node,
node              447 drivers/video/backlight/lm3630a_bl.c 	ret = fwnode_property_read_string(node, "label", &label);
node              455 drivers/video/backlight/lm3630a_bl.c 	ret = fwnode_property_read_u32(node, "default-brightness",
node              464 drivers/video/backlight/lm3630a_bl.c 	ret = fwnode_property_read_u32(node, "max-brightness", &val);
node              479 drivers/video/backlight/lm3630a_bl.c 	struct fwnode_handle *node;
node              481 drivers/video/backlight/lm3630a_bl.c 	device_for_each_child_node(pchip->dev, node) {
node              482 drivers/video/backlight/lm3630a_bl.c 		ret = lm3630a_parse_bank(pdata, node, &seen_led_sources);
node              348 drivers/video/backlight/lp855x_bl.c 	struct device_node *node = dev->of_node;
node              352 drivers/video/backlight/lp855x_bl.c 	if (!node) {
node              361 drivers/video/backlight/lp855x_bl.c 	of_property_read_string(node, "bl-name", &pdata->name);
node              362 drivers/video/backlight/lp855x_bl.c 	of_property_read_u8(node, "dev-ctrl", &pdata->device_control);
node              363 drivers/video/backlight/lp855x_bl.c 	of_property_read_u8(node, "init-brt", &pdata->initial_brightness);
node              364 drivers/video/backlight/lp855x_bl.c 	of_property_read_u32(node, "pwm-period", &pdata->period_ns);
node              367 drivers/video/backlight/lp855x_bl.c 	rom_length = of_get_child_count(node);
node              377 drivers/video/backlight/lp855x_bl.c 		for_each_child_of_node(node, child) {
node              228 drivers/video/backlight/pwm_bl.c 	struct device_node *node = dev->of_node;
node              238 drivers/video/backlight/pwm_bl.c 	if (!node)
node              247 drivers/video/backlight/pwm_bl.c 	of_property_read_u32(node, "post-pwm-on-delay-ms",
node              249 drivers/video/backlight/pwm_bl.c 	of_property_read_u32(node, "pwm-off-delay-ms", &data->pwm_off_delay);
node              257 drivers/video/backlight/pwm_bl.c 	prop = of_find_property(node, "brightness-levels", &length);
node              272 drivers/video/backlight/pwm_bl.c 		ret = of_property_read_u32_array(node, "brightness-levels",
node              278 drivers/video/backlight/pwm_bl.c 		ret = of_property_read_u32(node, "default-brightness-level",
node              290 drivers/video/backlight/pwm_bl.c 		of_property_read_u32(node, "num-interpolated-steps",
node              417 drivers/video/backlight/pwm_bl.c 	struct device_node *node = pb->dev->of_node;
node              420 drivers/video/backlight/pwm_bl.c 	if (!node || !node->phandle)
node              451 drivers/video/backlight/pwm_bl.c 	struct device_node *node = pdev->dev.of_node;
node              530 drivers/video/backlight/pwm_bl.c 	if (IS_ERR(pb->pwm) && PTR_ERR(pb->pwm) != -EPROBE_DEFER && !node) {
node              187 drivers/video/backlight/tps65217_bl.c 	struct device_node *node;
node              191 drivers/video/backlight/tps65217_bl.c 	node = of_get_child_by_name(tps->dev->of_node, "backlight");
node              192 drivers/video/backlight/tps65217_bl.c 	if (!node)
node              202 drivers/video/backlight/tps65217_bl.c 	if (!of_property_read_u32(node, "isel", &val)) {
node              215 drivers/video/backlight/tps65217_bl.c 	if (!of_property_read_u32(node, "fdim", &val)) {
node              241 drivers/video/backlight/tps65217_bl.c 	if (!of_property_read_u32(node, "default-brightness", &val)) {
node              252 drivers/video/backlight/tps65217_bl.c 	of_node_put(node);
node              257 drivers/video/backlight/tps65217_bl.c 	of_node_put(node);
node              556 drivers/video/fbdev/amba-clcd.c static int clcdfb_of_get_dpi_panel_mode(struct device_node *node,
node              563 drivers/video/fbdev/amba-clcd.c 	err = of_get_display_timing(node, "panel-timing", &timing);
node              565 drivers/video/fbdev/amba-clcd.c 		pr_err("%pOF: problems parsing panel-timing (%d)\n", node, err);
node              849 drivers/video/fbdev/amba-clcd.c 	struct device_node *node = dev->dev.of_node;
node              854 drivers/video/fbdev/amba-clcd.c 	board->name = of_node_full_name(node);
node              858 drivers/video/fbdev/amba-clcd.c 	if (of_find_property(node, "memory-region", NULL)) {
node              598 drivers/video/fbdev/arkfb.c 	rv = svga_check_timings (&ark_timing_regs, var, info->node);
node              785 drivers/video/fbdev/arkfb.c 			  hmul, info->node);
node             1212 drivers/video/fbdev/atmel_lcdfb.c 		       info->node, info->fix.mmio_start, sinfo->mmio, sinfo->irq_base);
node             1851 drivers/video/fbdev/aty/aty128fb.c 	snprintf(name, sizeof(name), "aty128bl%d", info->node);
node             2219 drivers/video/fbdev/aty/atyfb_base.c 	snprintf(name, sizeof(name), "atybl%d", info->node);
node             2722 drivers/video/fbdev/aty/atyfb_base.c 		info->node, info->fix.id, par->bus_type == ISA ? "ISA" : "PCI");
node              149 drivers/video/fbdev/aty/radeon_backlight.c 	snprintf(name, sizeof(name), "radeonbl%d", rinfo->info->node);
node              422 drivers/video/fbdev/chipsfb.c 		 p->node, p->fix.smem_len / 1024);
node              181 drivers/video/fbdev/core/fb_defio.c 	struct list_head *node, *next;
node              197 drivers/video/fbdev/core/fb_defio.c 	list_for_each_safe(node, next, &fbdefio->pagelist) {
node              198 drivers/video/fbdev/core/fb_defio.c 		list_del(node);
node             3090 drivers/video/fbdev/core/fbcon.c 	int idx = info->node;
node             3147 drivers/video/fbdev/core/fbcon.c 	idx = info->node;
node             3178 drivers/video/fbdev/core/fbcon.c 	int i, idx = info->node;
node             3209 drivers/video/fbdev/core/fbcon.c 		       info->fix.id, info->node);
node             3210 drivers/video/fbdev/core/fbcon.c 		primary_device = info->node;
node             3217 drivers/video/fbdev/core/fbcon.c 			       "fb%i, to tty %i-%i\n", info->node,
node             3238 drivers/video/fbdev/core/fbcon.c 	idx = info->node;
node             3321 drivers/video/fbdev/core/fbcon.c 			    info->node == con2fb_map[i]) {
node             3335 drivers/video/fbdev/core/fbcon.c 		    info->node == con2fb_map[fg_console]) {
node              727 drivers/video/fbdev/core/fbmem.c 		seq_printf(m, "%d %s\n", fi->node, fi->fix.id);
node             1595 drivers/video/fbdev/core/fbmem.c 	fb_info->node = i;
node             1664 drivers/video/fbdev/core/fbmem.c 	int i = fb_info->node;
node             1680 drivers/video/fbdev/core/fbmem.c 	i = fb_info->node;
node             1704 drivers/video/fbdev/core/fbmem.c 	registered_fb[fb_info->node] = NULL;
node              170 drivers/video/fbdev/core/svgalib.c void svga_dump_var(struct fb_var_screeninfo *var, int node)
node              172 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.vmode         : 0x%X\n", node, var->vmode);
node              173 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.xres          : %d\n", node, var->xres);
node              174 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.yres          : %d\n", node, var->yres);
node              175 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.bits_per_pixel: %d\n", node, var->bits_per_pixel);
node              176 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.xres_virtual  : %d\n", node, var->xres_virtual);
node              177 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.yres_virtual  : %d\n", node, var->yres_virtual);
node              178 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.left_margin   : %d\n", node, var->left_margin);
node              179 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.right_margin  : %d\n", node, var->right_margin);
node              180 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.upper_margin  : %d\n", node, var->upper_margin);
node              181 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.lower_margin  : %d\n", node, var->lower_margin);
node              182 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.hsync_len     : %d\n", node, var->hsync_len);
node              183 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.vsync_len     : %d\n", node, var->vsync_len);
node              184 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.sync          : 0x%X\n", node, var->sync);
node              185 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: var.pixclock      : %d\n\n", node, var->pixclock);
node              381 drivers/video/fbdev/core/svgalib.c int svga_compute_pll(const struct svga_pll *pll, u32 f_wanted, u16 *m, u16 *n, u16 *r, int node)
node              386 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: ideal frequency: %d kHz\n", node, (unsigned int) f_wanted);
node              433 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: found frequency: %d kHz (VCO %d kHz)\n", node, (int) (f_current >> ar), (int) f_current);
node              434 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: m = %d n = %d r = %d\n", node, (unsigned int) *m, (unsigned int) *n, (unsigned int) *r);
node              443 drivers/video/fbdev/core/svgalib.c int svga_check_timings(const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, int node)
node              512 drivers/video/fbdev/core/svgalib.c 		      u32 hmul, u32 hdiv, u32 vmul, u32 vdiv, u32 hborder, int node)
node              519 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal total      : %d\n", node, value);
node              524 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal display    : %d\n", node, value);
node              529 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal blank start: %d\n", node, value);
node              534 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal blank end  : %d\n", node, value);
node              539 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal sync start : %d\n", node, value);
node              544 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: horizontal sync end   : %d\n", node, value);
node              549 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical total        : %d\n", node, value);
node              554 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical display      : %d\n", node, value);
node              559 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical blank start  : %d\n", node, value);
node              564 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical blank end    : %d\n", node, value);
node              569 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical sync start   : %d\n", node, value);
node              574 drivers/video/fbdev/core/svgalib.c 	pr_debug("fb%d: vertical sync end     : %d\n", node, value);
node              581 drivers/video/fbdev/core/svgalib.c 		pr_debug("fb%d: positive horizontal sync\n", node);
node              584 drivers/video/fbdev/core/svgalib.c 		pr_debug("fb%d: negative horizontal sync\n", node);
node              588 drivers/video/fbdev/core/svgalib.c 		pr_debug("fb%d: positive vertical sync\n", node);
node              591 drivers/video/fbdev/core/svgalib.c 		pr_debug("fb%d: negative vertical sync\n\n", node);
node              520 drivers/video/fbdev/ep93xx-fb.c 	info->node		= -1;
node              297 drivers/video/fbdev/geode/gx1fb_core.c 	info->node		= -1;
node              311 drivers/video/fbdev/geode/gxfb_core.c 	info->node		= -1;
node              432 drivers/video/fbdev/geode/lxfb_core.c 	info->node		= -1;
node              476 drivers/video/fbdev/grvga.c 		 info->node, info->var.xres, info->var.yres, info->var.bits_per_pixel,
node             2084 drivers/video/fbdev/i810/i810_main.c 	       info->node,
node              113 drivers/video/fbdev/matrox/i2c-matroxfb.c 		minfo->fbcon.node);
node              215 drivers/video/fbdev/matrox/i2c-matroxfb.c 	.node =		LIST_HEAD_INIT(i2c_matroxfb.node),
node             1955 drivers/video/fbdev/matrox/matroxfb_base.c #define matroxfb_driver_l(x) list_entry(x, struct matroxfb_driver, node)
node             1959 drivers/video/fbdev/matrox/matroxfb_base.c 	list_add(&drv->node, &matroxfb_driver_list);
node             1979 drivers/video/fbdev/matrox/matroxfb_base.c 	list_del(&drv->node);
node             2003 drivers/video/fbdev/matrox/matroxfb_base.c 	     drv = matroxfb_driver_l(drv->node.next)) {
node              507 drivers/video/fbdev/matrox/matroxfb_base.h 	struct list_head	node;
node              660 drivers/video/fbdev/matrox/matroxfb_crtc2.c 		minfo->fbcon.node, m2info->fbcon.node);
node              683 drivers/video/fbdev/matrox/matroxfb_crtc2.c 		id = m2info->fbcon.node;
node              728 drivers/video/fbdev/metronomefb.c 		" memory\n", info->node, videomemorysize >> 10);
node               79 drivers/video/fbdev/mmp/core.c 	list_add_tail(&panel->node, &panel_list);
node               82 drivers/video/fbdev/mmp/core.c 	list_for_each_entry(path, &path_list, node) {
node              107 drivers/video/fbdev/mmp/core.c 	list_del(&panel->node);
node              109 drivers/video/fbdev/mmp/core.c 	list_for_each_entry(path, &path_list, node) {
node              134 drivers/video/fbdev/mmp/core.c 	list_for_each_entry(path, &path_list, node) {
node              176 drivers/video/fbdev/mmp/core.c 	list_for_each_entry(panel, &panel_list, node) {
node              204 drivers/video/fbdev/mmp/core.c 	list_add_tail(&path->node, &path_list);
node              226 drivers/video/fbdev/mmp/core.c 	list_del(&path->node);
node              509 drivers/video/fbdev/mmp/fb/mmpfb.c 	info->node = -1;
node              633 drivers/video/fbdev/mmp/fb/mmpfb.c 		info->node, info->fix.id);
node              102 drivers/video/fbdev/nvidia/nv_backlight.c 	snprintf(name, sizeof(name), "nvidiabl%d", info->node);
node              213 drivers/video/fbdev/omap2/omapfb/displays/connector-analog-tv.c 	struct device_node *node = pdev->dev.of_node;
node              216 drivers/video/fbdev/omap2/omapfb/displays/connector-analog-tv.c 	in = omapdss_of_find_source_for_first_ep(node);
node              238 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c 	struct device_node *node = pdev->dev.of_node;
node              243 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c 	in = omapdss_of_find_source_for_first_ep(node);
node              251 drivers/video/fbdev/omap2/omapfb/displays/connector-dvi.c 	adapter_node = of_parse_phandle(node, "ddc-i2c-bus", 0);
node              208 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c 	struct device_node *node = pdev->dev.of_node;
node              213 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c 	gpio = of_get_named_gpio(node, "hpd-gpios", 0);
node              219 drivers/video/fbdev/omap2/omapfb/displays/connector-hdmi.c 	in = omapdss_of_find_source_for_first_ep(node);
node              182 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c 	struct device_node *node = pdev->dev.of_node;
node              190 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c 	if (node == NULL) {
node              207 drivers/video/fbdev/omap2/omapfb/displays/encoder-opa362.c 	in = omapdss_of_find_source_for_first_ep(node);
node              168 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c 	struct device_node *node = pdev->dev.of_node;
node              172 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c 	gpio = of_get_named_gpio(node, "powerdown-gpios", 0);
node              181 drivers/video/fbdev/omap2/omapfb/displays/encoder-tfp410.c 	in = omapdss_of_find_source_for_first_ep(node);
node              203 drivers/video/fbdev/omap2/omapfb/displays/encoder-tpd12s015.c 	struct device_node *node = pdev->dev.of_node;
node              206 drivers/video/fbdev/omap2/omapfb/displays/encoder-tpd12s015.c 	in = omapdss_of_find_source_for_first_ep(node);
node              200 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c 	struct device_node *node = pdev->dev.of_node;
node              215 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c 	r = of_get_display_timing(node, "panel-timing", &timing);
node              224 drivers/video/fbdev/omap2/omapfb/displays/panel-dpi.c 	in = omapdss_of_find_source_for_first_ep(node);
node             1120 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c 	struct device_node *node = pdev->dev.of_node;
node             1125 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c 	gpio = of_get_named_gpio(node, "reset-gpios", 0);
node             1132 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c 	gpio = of_get_named_gpio(node, "te-gpios", 0);
node             1140 drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c 	in = omapdss_of_find_source_for_first_ep(node);
node              241 drivers/video/fbdev/omap2/omapfb/displays/panel-lgphilips-lb035q02.c 	struct device_node *node = spi->dev.of_node;
node              256 drivers/video/fbdev/omap2/omapfb/displays/panel-lgphilips-lb035q02.c 	in = omapdss_of_find_source_for_first_ep(node);
node              233 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c 	struct device_node *node = spi->dev.of_node;
node              238 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c 	gpio = of_get_named_gpio(node, "reset-gpios", 0);
node              248 drivers/video/fbdev/omap2/omapfb/displays/panel-nec-nl8048hl11.c 	in = omapdss_of_find_source_for_first_ep(node);
node              214 drivers/video/fbdev/omap2/omapfb/displays/panel-sharp-ls037v7dw01.c 	struct device_node *node = pdev->dev.of_node;
node              249 drivers/video/fbdev/omap2/omapfb/displays/panel-sharp-ls037v7dw01.c 	in = omapdss_of_find_source_for_first_ep(node);
node              358 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td028ttec1.c 	struct device_node *node = spi->dev.of_node;
node              362 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td028ttec1.c 	in = omapdss_of_find_source_for_first_ep(node);
node              465 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c 	struct device_node *node = spi->dev.of_node;
node              470 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c 	gpio = of_get_named_gpio(node, "reset-gpios", 0);
node              477 drivers/video/fbdev/omap2/omapfb/displays/panel-tpo-td043mtea1.c 	in = omapdss_of_find_source_for_first_ep(node);
node             5072 drivers/video/fbdev/omap2/omapfb/dss/dsi.c 	struct device_node *node = pdev->dev.of_node;
node             5081 drivers/video/fbdev/omap2/omapfb/dss/dsi.c 	ep = omapdss_of_get_first_endpoint(node);
node              120 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c static struct device_node *omapdss_of_get_remote_port(const struct device_node *node)
node              124 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 	np = of_graph_get_remote_endpoint(node);
node              152 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c omapdss_of_find_source_for_first_ep(struct device_node *node)
node              158 drivers/video/fbdev/omap2/omapfb/dss/dss-of.c 	ep = omapdss_of_get_first_endpoint(node);
node              531 drivers/video/fbdev/omap2/omapfb/dss/hdmi4.c 	struct device_node *node = pdev->dev.of_node;
node              535 drivers/video/fbdev/omap2/omapfb/dss/hdmi4.c 	ep = omapdss_of_get_first_endpoint(node);
node              561 drivers/video/fbdev/omap2/omapfb/dss/hdmi5.c 	struct device_node *node = pdev->dev.of_node;
node              565 drivers/video/fbdev/omap2/omapfb/dss/hdmi5.c 	ep = omapdss_of_get_first_endpoint(node);
node               28 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	struct device_node *node;
node               44 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_update_prop(struct device_node *node, char *compat,
node               57 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	of_update_property(node, prop);
node               80 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_omapify_node(struct device_node *node)
node               87 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	prop = of_find_property(node, "compatible", NULL);
node              108 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	omapdss_update_prop(node, new_compat, new_len);
node              111 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_add_to_list(struct device_node *node, bool root)
node              116 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 		n->node = node;
node              122 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static bool __init omapdss_list_contains(const struct device_node *node)
node              127 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 		if (n->node == node)
node              134 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c static void __init omapdss_walk_device(struct device_node *node, bool root)
node              138 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	omapdss_add_to_list(node, root);
node              144 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	n = of_get_child_by_name(node, "ports");
node              146 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 		n = of_get_child_by_name(node, "port");
node              153 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 	while ((n = of_graph_get_next_endpoint(node, n)) != NULL) {
node              208 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 			omapdss_omapify_node(n->node);
node              211 drivers/video/fbdev/omap2/omapfb/dss/omapdss-boot-init.c 		of_node_put(n->node);
node              804 drivers/video/fbdev/omap2/omapfb/dss/venc.c 	struct device_node *node = pdev->dev.of_node;
node              809 drivers/video/fbdev/omap2/omapfb/dss/venc.c 	ep = omapdss_of_get_first_endpoint(node);
node              229 drivers/video/fbdev/pmag-aa-fb.c 		info->node, info->fix.id, dev_name(dev));
node              652 drivers/video/fbdev/pxa168fb.c 	info->node = -1;
node              892 drivers/video/fbdev/pxafb.c 	ofb->fb.node			= -1;
node             1830 drivers/video/fbdev/pxafb.c 	fbi->fb.node		= -1;
node              340 drivers/video/fbdev/riva/fbdev.c 	snprintf(name, sizeof(name), "rivabl%d", info->node);
node              990 drivers/video/fbdev/s3c2410fb.c 		fbinfo->node, fbinfo->fix.id);
node              466 drivers/video/fbdev/s3fb.c 			      1000000000 / pixclock, &m, &n, &r, info->node);
node              584 drivers/video/fbdev/s3fb.c 	rv = svga_check_timings (&s3_timing_regs, var, info->node);
node              591 drivers/video/fbdev/s3fb.c 				info->node);
node              892 drivers/video/fbdev/s3fb.c 			 hmul, info->node);
node              494 drivers/video/fbdev/simplefb.c 	dev_info(&pdev->dev, "fb%d: simplefb registered!\n", info->node);
node             1926 drivers/video/fbdev/sm501fb.c 	dev_info(info->dev, "fb%d: %s frame buffer\n", fbi->node, fbi->fix.id);
node             1097 drivers/video/fbdev/smscufx.c 		info->node, user, info, dev->fb_count);
node             1133 drivers/video/fbdev/smscufx.c 	int node = info->node;
node             1150 drivers/video/fbdev/smscufx.c 	pr_debug("fb_info for /dev/fb%d has been freed", node);
node             1177 drivers/video/fbdev/smscufx.c 		  info->node, user, dev->fb_count);
node             1719 drivers/video/fbdev/smscufx.c 		" Using %dK framebuffer memory\n", info->node,
node             1811 drivers/video/fbdev/smscufx.c 	struct list_head *node;
node             1828 drivers/video/fbdev/smscufx.c 		node = dev->urbs.list.next; /* have reserved one with sem */
node             1829 drivers/video/fbdev/smscufx.c 		list_del_init(node);
node             1833 drivers/video/fbdev/smscufx.c 		unode = list_entry(node, struct urb_node, entry);
node             1840 drivers/video/fbdev/smscufx.c 		kfree(node);
node              595 drivers/video/fbdev/ssd1307fb.c 	struct device_node *node = client->dev.of_node;
node              602 drivers/video/fbdev/ssd1307fb.c 	if (!node) {
node              638 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,width", &par->width))
node              641 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,height", &par->height))
node              644 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,page-offset", &par->page_offset))
node              647 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,com-offset", &par->com_offset))
node              650 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,prechargep1", &par->prechargep1))
node              653 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,prechargep2", &par->prechargep2))
node              656 drivers/video/fbdev/ssd1307fb.c 	if (!of_property_read_u8_array(node, "solomon,lookup-table",
node              661 drivers/video/fbdev/ssd1307fb.c 	par->seg_remap = !of_property_read_bool(node, "solomon,segment-no-remap");
node              662 drivers/video/fbdev/ssd1307fb.c 	par->com_seq = of_property_read_bool(node, "solomon,com-seq");
node              663 drivers/video/fbdev/ssd1307fb.c 	par->com_lrremap = of_property_read_bool(node, "solomon,com-lrremap");
node              664 drivers/video/fbdev/ssd1307fb.c 	par->com_invdir = of_property_read_bool(node, "solomon,com-invdir");
node              666 drivers/video/fbdev/ssd1307fb.c 		of_property_read_bool(node, "solomon,area-color-enable");
node              667 drivers/video/fbdev/ssd1307fb.c 	par->low_power = of_property_read_bool(node, "solomon,low-power");
node              673 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,dclk-div", &par->dclk_div))
node              675 drivers/video/fbdev/ssd1307fb.c 	if (of_property_read_u32(node, "solomon,dclk-frq", &par->dclk_frq))
node              745 drivers/video/fbdev/ssd1307fb.c 	snprintf(bl_name, sizeof(bl_name), "ssd1307fb%d", info->node);
node              759 drivers/video/fbdev/ssd1307fb.c 	dev_info(&client->dev, "fb%d: %s framebuffer device registered, using %d bytes of video memory\n", info->node, info->fix.id, vmem_size);
node             1698 drivers/video/fbdev/tridentfb.c 	   info->node, info->fix.id, info->var.xres,
node             1847 drivers/video/fbdev/udlfb.c 	struct list_head *node;
node             1857 drivers/video/fbdev/udlfb.c 		node = dlfb->urbs.list.next; /* have reserved one with sem */
node             1858 drivers/video/fbdev/udlfb.c 		list_del_init(node);
node             1862 drivers/video/fbdev/udlfb.c 		unode = list_entry(node, struct urb_node, entry);
node             1869 drivers/video/fbdev/udlfb.c 		kfree(node);
node             1778 drivers/video/fbdev/via/viafbdev.c 	viafbinfo->node = 0;
node             1888 drivers/video/fbdev/via/viafbdev.c 		  viafbinfo->node, viafbinfo->fix.id, default_var.xres,
node              309 drivers/video/fbdev/vt8500lcdfb.c 	fbi->fb.node		= -1;
node              257 drivers/video/fbdev/vt8623fb.c 	rv = svga_compute_pll(&vt8623_pll, 1000000000 / pixclock, &m, &n, &r, info->node);
node              360 drivers/video/fbdev/vt8623fb.c 	rv = svga_check_timings (&vt8623_timing_regs, var, info->node);
node              505 drivers/video/fbdev/vt8623fb.c 			 1, info->node);
node              713 drivers/video/fbdev/w100fb.c 	info->node = -1;
node              296 drivers/video/fbdev/wm8505fb.c 	fbi->fb.node		= -1;
node              796 drivers/virt/fsl_hypervisor.c 	struct device_node *node;
node              799 drivers/virt/fsl_hypervisor.c 	node = of_find_node_by_path("/hypervisor");
node              800 drivers/virt/fsl_hypervisor.c 	if (!node)
node              803 drivers/virt/fsl_hypervisor.c 	ret = of_find_property(node, "fsl,hv-version", NULL) != NULL;
node              805 drivers/virt/fsl_hypervisor.c 	of_node_put(node);
node              100 drivers/virtio/virtio_mmio.c 	struct list_head node;
node              302 drivers/virtio/virtio_mmio.c 		list_for_each_entry(info, &vm_dev->virtqueues, node)
node              320 drivers/virtio/virtio_mmio.c 	list_del(&info->node);
node              438 drivers/virtio/virtio_mmio.c 	list_add(&info->node, &vm_dev->virtqueues);
node               67 drivers/virtio/virtio_pci_common.c 	list_for_each_entry(info, &vp_dev->virtqueues, node) {
node              197 drivers/virtio/virtio_pci_common.c 		list_add(&info->node, &vp_dev->virtqueues);
node              200 drivers/virtio/virtio_pci_common.c 		INIT_LIST_HEAD(&info->node);
node              218 drivers/virtio/virtio_pci_common.c 	list_del(&info->node);
node               36 drivers/virtio/virtio_pci_common.h 	struct list_head node;
node              444 drivers/w1/w1_netlink.c 	struct w1_cb_node *node = container_of(async_cmd, struct w1_cb_node,
node              446 drivers/w1/w1_netlink.c 	u16 mlen = node->msg->len;
node              449 drivers/w1/w1_netlink.c 	struct w1_slave *sl = node->sl;
node              450 drivers/w1/w1_netlink.c 	struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)node->msg->data;
node              453 drivers/w1/w1_netlink.c 	dev->priv = node->block;
node              456 drivers/w1/w1_netlink.c 	node->block->cur_msg = node->msg;
node              468 drivers/w1/w1_netlink.c 		w1_netlink_check_send(node->block);
node              470 drivers/w1/w1_netlink.c 		w1_netlink_queue_status(node->block, node->msg, cmd, err);
node              479 drivers/w1/w1_netlink.c 		w1_netlink_queue_status(node->block, node->msg, cmd, err);
node              495 drivers/w1/w1_netlink.c 	w1_unref_block(node->block);
node              544 drivers/w1/w1_netlink.c 	struct w1_cb_node *node = NULL;
node              615 drivers/w1/w1_netlink.c 		node = (struct w1_cb_node *)(block->request_cn.data + cn->len);
node              627 drivers/w1/w1_netlink.c 		block->first_cn = (struct cn_msg *)(node + node_count);
node              679 drivers/w1/w1_netlink.c 		node->async.cb = w1_process_cb;
node              680 drivers/w1/w1_netlink.c 		node->block = block;
node              681 drivers/w1/w1_netlink.c 		node->msg = (struct w1_netlink_msg *)((u8 *)&block->request_cn +
node              683 drivers/w1/w1_netlink.c 		node->sl = sl;
node              684 drivers/w1/w1_netlink.c 		node->dev = dev;
node              687 drivers/w1/w1_netlink.c 		list_add_tail(&node->async.async_entry, &dev->async_list);
node              690 drivers/w1/w1_netlink.c 		++node;
node              133 drivers/watchdog/octeon-wdt-main.c 	int node = cpu_to_node(cpu);
node              138 drivers/watchdog/octeon-wdt-main.c 			cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1);
node              147 drivers/watchdog/octeon-wdt-main.c 		cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1);
node              279 drivers/watchdog/octeon-wdt-main.c 		unsigned int node = cvmx_get_node_num();
node              289 drivers/watchdog/octeon-wdt-main.c 			ciu_wdog.u64 = cvmx_read_csr_node(node, CVMX_CIU_WDOGX(lcore));
node              302 drivers/watchdog/octeon-wdt-main.c 	int node;
node              306 drivers/watchdog/octeon-wdt-main.c 	node = cpu_to_node(cpu);
node              312 drivers/watchdog/octeon-wdt-main.c 		domain = octeon_irq_get_block_domain(node,
node              325 drivers/watchdog/octeon-wdt-main.c 	int node;
node              330 drivers/watchdog/octeon-wdt-main.c 	node = cpu_to_node(cpu);
node              333 drivers/watchdog/octeon-wdt-main.c 	cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1);
node              337 drivers/watchdog/octeon-wdt-main.c 	cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64);
node              348 drivers/watchdog/octeon-wdt-main.c 	int node;
node              353 drivers/watchdog/octeon-wdt-main.c 	node = cpu_to_node(cpu);
node              359 drivers/watchdog/octeon-wdt-main.c 	cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64);
node              365 drivers/watchdog/octeon-wdt-main.c 		domain = octeon_irq_get_block_domain(node, WD_BLOCK_NUMBER);
node              391 drivers/watchdog/octeon-wdt-main.c 	cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(core), 1);
node              397 drivers/watchdog/octeon-wdt-main.c 	cvmx_write_csr_node(node, CVMX_CIU_WDOGX(core), ciu_wdog.u64);
node              406 drivers/watchdog/octeon-wdt-main.c 	int node;
node              413 drivers/watchdog/octeon-wdt-main.c 		node = cpu_to_node(cpu);
node              414 drivers/watchdog/octeon-wdt-main.c 		cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1);
node              458 drivers/watchdog/octeon-wdt-main.c 	int node;
node              470 drivers/watchdog/octeon-wdt-main.c 		node = cpu_to_node(cpu);
node              471 drivers/watchdog/octeon-wdt-main.c 		cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1);
node              475 drivers/watchdog/octeon-wdt-main.c 		cvmx_write_csr_node(node, CVMX_CIU_WDOGX(coreid), ciu_wdog.u64);
node              476 drivers/watchdog/octeon-wdt-main.c 		cvmx_write_csr_node(node, CVMX_CIU_PP_POKEX(coreid), 1);
node              501 drivers/watchdog/orion_wdt.c 	struct device_node *node = pdev->dev.of_node;
node              513 drivers/watchdog/orion_wdt.c 	if (of_device_is_compatible(node, "marvell,orion-wdt")) {
node              520 drivers/watchdog/orion_wdt.c 	} else if (of_device_is_compatible(node, "marvell,armada-370-wdt") ||
node              521 drivers/watchdog/orion_wdt.c 		   of_device_is_compatible(node, "marvell,armada-xp-wdt")) {
node              528 drivers/watchdog/orion_wdt.c 	} else if (of_device_is_compatible(node, "marvell,armada-375-wdt") ||
node              529 drivers/watchdog/orion_wdt.c 		   of_device_is_compatible(node, "marvell,armada-380-wdt")) {
node               95 drivers/watchdog/sun4v_wdt.c 	u64 node;
node              114 drivers/watchdog/sun4v_wdt.c 	node = mdesc_node_by_name(handle, MDESC_NODE_NULL, "platform");
node              116 drivers/watchdog/sun4v_wdt.c 	if (node == MDESC_NODE_NULL)
node              127 drivers/watchdog/sun4v_wdt.c 	value = mdesc_get_property(handle, node, "watchdog-resolution", NULL);
node              135 drivers/watchdog/sun4v_wdt.c 	value = mdesc_get_property(handle, node, "watchdog-max-timeout", NULL);
node               27 drivers/watchdog/wdat_wdt.c 	struct list_head node;
node              128 drivers/watchdog/wdat_wdt.c 	list_for_each_entry(instr, wdat->instructions[action], node) {
node              392 drivers/watchdog/wdat_wdt.c 		INIT_LIST_HEAD(&instr->node);
node              436 drivers/watchdog/wdat_wdt.c 		list_add_tail(&instr->node, instructions);
node               90 drivers/xen/cpu_hotplug.c 		.node = "cpu",
node               84 drivers/xen/evtchn.c 	struct rb_node node;
node              116 drivers/xen/evtchn.c 		this = rb_entry(*new, struct user_evtchn, node);
node              128 drivers/xen/evtchn.c 	rb_link_node(&evtchn->node, parent, new);
node              129 drivers/xen/evtchn.c 	rb_insert_color(&evtchn->node, &u->evtchns);
node              137 drivers/xen/evtchn.c 	rb_erase(&evtchn->node, &u->evtchns);
node              143 drivers/xen/evtchn.c 	struct rb_node *node = u->evtchns.rb_node;
node              145 drivers/xen/evtchn.c 	while (node) {
node              148 drivers/xen/evtchn.c 		evtchn = rb_entry(node, struct user_evtchn, node);
node              151 drivers/xen/evtchn.c 			node = node->rb_left;
node              153 drivers/xen/evtchn.c 			node = node->rb_right;
node              673 drivers/xen/evtchn.c 	struct rb_node *node;
node              675 drivers/xen/evtchn.c 	while ((node = u->evtchns.rb_node)) {
node              678 drivers/xen/evtchn.c 		evtchn = rb_entry(node, struct user_evtchn, node);
node              314 drivers/xen/manage.c 	.node = "control/sysrq",
node              320 drivers/xen/manage.c 	.node = "control/shutdown",
node              333 drivers/xen/manage.c 	char node[FEATURE_PATH_SIZE];
node              353 drivers/xen/manage.c 		snprintf(node, FEATURE_PATH_SIZE, "feature-%s",
node              355 drivers/xen/manage.c 		err = xenbus_printf(XBT_NIL, "control", node, "%u", 1);
node              358 drivers/xen/manage.c 				err, node);
node              106 drivers/xen/xen-balloon.c 	.node = "memory/target",
node              117 drivers/xen/xenbus/xenbus.h void xenbus_dev_changed(const char *node, struct xen_bus_type *bus);
node              122 drivers/xen/xenbus/xenbus_client.c 	watch->node = path;
node              128 drivers/xen/xenbus/xenbus_client.c 		watch->node = NULL;
node              555 drivers/xen/xenbus/xenbus_client.c 	struct xenbus_map_node *node;
node              569 drivers/xen/xenbus/xenbus_client.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              570 drivers/xen/xenbus/xenbus_client.c 	if (!node)
node              573 drivers/xen/xenbus/xenbus_client.c 	err = alloc_xenballooned_pages(nr_pages, node->hvm.pages);
node              577 drivers/xen/xenbus/xenbus_client.c 	gnttab_foreach_grant(node->hvm.pages, nr_grefs,
node              581 drivers/xen/xenbus/xenbus_client.c 	err = __xenbus_map_ring(dev, gnt_ref, nr_grefs, node->handles,
node              583 drivers/xen/xenbus/xenbus_client.c 	node->nr_handles = nr_grefs;
node              588 drivers/xen/xenbus/xenbus_client.c 	addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP,
node              595 drivers/xen/xenbus/xenbus_client.c 	node->hvm.addr = addr;
node              598 drivers/xen/xenbus/xenbus_client.c 	list_add(&node->next, &xenbus_valloc_pages);
node              606 drivers/xen/xenbus/xenbus_client.c 		xenbus_unmap_ring(dev, node->handles, nr_grefs, info.addrs);
node              612 drivers/xen/xenbus/xenbus_client.c 		free_xenballooned_pages(nr_pages, node->hvm.pages);
node              614 drivers/xen/xenbus/xenbus_client.c 	kfree(node);
node              681 drivers/xen/xenbus/xenbus_client.c 	struct xenbus_map_node *node;
node              694 drivers/xen/xenbus/xenbus_client.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              695 drivers/xen/xenbus/xenbus_client.c 	if (!node)
node              700 drivers/xen/xenbus/xenbus_client.c 		kfree(node);
node              707 drivers/xen/xenbus/xenbus_client.c 	err = __xenbus_map_ring(dev, gnt_refs, nr_grefs, node->handles,
node              714 drivers/xen/xenbus/xenbus_client.c 	node->nr_handles = nr_grefs;
node              715 drivers/xen/xenbus/xenbus_client.c 	node->pv.area = area;
node              718 drivers/xen/xenbus/xenbus_client.c 	list_add(&node->next, &xenbus_valloc_pages);
node              730 drivers/xen/xenbus/xenbus_client.c 	kfree(node);
node              736 drivers/xen/xenbus/xenbus_client.c 	struct xenbus_map_node *node;
node              744 drivers/xen/xenbus/xenbus_client.c 	list_for_each_entry(node, &xenbus_valloc_pages, next) {
node              745 drivers/xen/xenbus/xenbus_client.c 		if (node->pv.area->addr == vaddr) {
node              746 drivers/xen/xenbus/xenbus_client.c 			list_del(&node->next);
node              750 drivers/xen/xenbus/xenbus_client.c 	node = NULL;
node              754 drivers/xen/xenbus/xenbus_client.c 	if (!node) {
node              760 drivers/xen/xenbus/xenbus_client.c 	for (i = 0; i < node->nr_handles; i++) {
node              768 drivers/xen/xenbus/xenbus_client.c 		unmap[i].handle = node->handles[i];
node              776 drivers/xen/xenbus/xenbus_client.c 	for (i = 0; i < node->nr_handles; i++) {
node              781 drivers/xen/xenbus/xenbus_client.c 					 node->handles[i], unmap[i].status);
node              788 drivers/xen/xenbus/xenbus_client.c 		free_vm_area(node->pv.area);
node              791 drivers/xen/xenbus/xenbus_client.c 			 node->pv.area, node->nr_handles);
node              793 drivers/xen/xenbus/xenbus_client.c 	kfree(node);
node              824 drivers/xen/xenbus/xenbus_client.c 	struct xenbus_map_node *node;
node              832 drivers/xen/xenbus/xenbus_client.c 	list_for_each_entry(node, &xenbus_valloc_pages, next) {
node              833 drivers/xen/xenbus/xenbus_client.c 		addr = node->hvm.addr;
node              835 drivers/xen/xenbus/xenbus_client.c 			list_del(&node->next);
node              839 drivers/xen/xenbus/xenbus_client.c 	node = addr = NULL;
node              843 drivers/xen/xenbus/xenbus_client.c 	if (!node) {
node              849 drivers/xen/xenbus/xenbus_client.c 	nr_pages = XENBUS_PAGES(node->nr_handles);
node              851 drivers/xen/xenbus/xenbus_client.c 	gnttab_foreach_grant(node->hvm.pages, node->nr_handles,
node              855 drivers/xen/xenbus/xenbus_client.c 	rv = xenbus_unmap_ring(dev, node->handles, node->nr_handles,
node              859 drivers/xen/xenbus/xenbus_client.c 		free_xenballooned_pages(nr_pages, node->hvm.pages);
node              864 drivers/xen/xenbus/xenbus_client.c 	kfree(node);
node              235 drivers/xen/xenbus/xenbus_dev_frontend.c 	kfree(watch->watch.node);
node              249 drivers/xen/xenbus/xenbus_dev_frontend.c 	watch->watch.node = kstrdup(path, GFP_KERNEL);
node              250 drivers/xen/xenbus/xenbus_dev_frontend.c 	if (watch->watch.node == NULL)
node              534 drivers/xen/xenbus/xenbus_dev_frontend.c 			    !strcmp(watch->watch.node, path)) {
node              113 drivers/xen/xenbus/xenbus_probe.c 	if (dev->otherend_watch.node) {
node              115 drivers/xen/xenbus/xenbus_probe.c 		kfree(dev->otherend_watch.node);
node              116 drivers/xen/xenbus/xenbus_probe.c 		dev->otherend_watch.node = NULL;
node              191 drivers/xen/xenbus/xenbus_probe.c 		state, xenbus_strstate(state), dev->otherend_watch.node, path);
node              556 drivers/xen/xenbus/xenbus_probe.c void xenbus_dev_changed(const char *node, struct xen_bus_type *bus)
node              563 drivers/xen/xenbus/xenbus_probe.c 	if (char_count(node, '/') < 2)
node              566 drivers/xen/xenbus/xenbus_probe.c 	exists = xenbus_exists(XBT_NIL, node, "");
node              568 drivers/xen/xenbus/xenbus_probe.c 		xenbus_cleanup_devices(node, &bus->bus);
node              573 drivers/xen/xenbus/xenbus_probe.c 	p = strchr(node, '/') + 1;
node              577 drivers/xen/xenbus/xenbus_probe.c 	rootlen = strsep_len(node, '/', bus->levels);
node              580 drivers/xen/xenbus/xenbus_probe.c 	root = kasprintf(GFP_KERNEL, "%.*s", rootlen, node);
node              215 drivers/xen/xenbus/xenbus_probe_backend.c 	.node = "backend",
node              167 drivers/xen/xenbus/xenbus_probe_frontend.c 	.node = "device",
node              368 drivers/xen/xenbus/xenbus_probe_frontend.c 	be_watch.node = kasprintf(GFP_NOIO | __GFP_HIGH, "%s/state", be);
node              369 drivers/xen/xenbus/xenbus_probe_frontend.c 	if (!be_watch.node)
node              397 drivers/xen/xenbus/xenbus_probe_frontend.c 	kfree(be_watch.node);
node              423 drivers/xen/xenbus/xenbus_xs.c 			const char *dir, const char *node, unsigned int *num)
node              428 drivers/xen/xenbus/xenbus_xs.c 	path = join(dir, node);
node              443 drivers/xen/xenbus/xenbus_xs.c 		  const char *dir, const char *node)
node              448 drivers/xen/xenbus/xenbus_xs.c 	d = xenbus_directory(t, dir, node, &dir_n);
node              461 drivers/xen/xenbus/xenbus_xs.c 		  const char *dir, const char *node, unsigned int *len)
node              466 drivers/xen/xenbus/xenbus_xs.c 	path = join(dir, node);
node              480 drivers/xen/xenbus/xenbus_xs.c 		 const char *dir, const char *node, const char *string)
node              486 drivers/xen/xenbus/xenbus_xs.c 	path = join(dir, node);
node              503 drivers/xen/xenbus/xenbus_xs.c 		 const char *dir, const char *node)
node              508 drivers/xen/xenbus/xenbus_xs.c 	path = join(dir, node);
node              519 drivers/xen/xenbus/xenbus_xs.c int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node)
node              524 drivers/xen/xenbus/xenbus_xs.c 	path = join(dir, node);
node              569 drivers/xen/xenbus/xenbus_xs.c 		 const char *dir, const char *node, const char *fmt, ...)
node              575 drivers/xen/xenbus/xenbus_xs.c 	val = xenbus_read(t, dir, node, NULL);
node              591 drivers/xen/xenbus/xenbus_xs.c unsigned int xenbus_read_unsigned(const char *dir, const char *node,
node              597 drivers/xen/xenbus/xenbus_xs.c 	ret = xenbus_scanf(XBT_NIL, dir, node, "%u", &val);
node              607 drivers/xen/xenbus/xenbus_xs.c 		  const char *dir, const char *node, const char *fmt, ...)
node              620 drivers/xen/xenbus/xenbus_xs.c 	ret = xenbus_write(t, dir, node, buf);
node              775 drivers/xen/xenbus/xenbus_xs.c 	err = xs_watch(watch->node, token);
node              804 drivers/xen/xenbus/xenbus_xs.c 	err = xs_unwatch(watch->node, token);
node              806 drivers/xen/xenbus/xenbus_xs.c 		pr_warn("Failed to release watch %s: %i\n", watch->node, err);
node              852 drivers/xen/xenbus/xenbus_xs.c 		xs_watch(watch->node, token);
node               88 fs/afs/afs.h   	__s8		node[6];			/* spatially unique node ID (MAC addr) */
node              203 fs/afs/afs.h   	__be32		node[6];
node              463 fs/afs/cmservice.c 			r->node[loop] = ntohl(b[loop + 5]);
node              574 fs/afs/cmservice.c 			r->node[loop] = ntohl(b[loop + 5]);
node              619 fs/afs/cmservice.c 		reply.ia.uuid[loop + 5] = htonl((s8) call->net->uuid.node[loop]);
node               83 fs/afs/vlclient.c 			uuid->node[j] = (u8)ntohl(xdr->node[j]);
node              303 fs/afs/vlclient.c 		r->uuid.node[i] = htonl(u->node[i]);
node              100 fs/befs/btree.c 			     struct befs_btree_node *node,
node              103 fs/befs/btree.c static int befs_leafnode(struct befs_btree_node *node);
node              105 fs/befs/btree.c static fs16 *befs_bt_keylen_index(struct befs_btree_node *node);
node              107 fs/befs/btree.c static fs64 *befs_bt_valarray(struct befs_btree_node *node);
node              109 fs/befs/btree.c static char *befs_bt_keydata(struct befs_btree_node *node);
node              112 fs/befs/btree.c 			 struct befs_btree_node *node,
node              116 fs/befs/btree.c 			     struct befs_btree_node *node,
node              192 fs/befs/btree.c 		  struct befs_btree_node *node, befs_off_t node_off)
node              198 fs/befs/btree.c 	if (node->bh)
node              199 fs/befs/btree.c 		brelse(node->bh);
node              201 fs/befs/btree.c 	node->bh = befs_read_datastream(sb, ds, node_off, &off);
node              202 fs/befs/btree.c 	if (!node->bh) {
node              209 fs/befs/btree.c 	node->od_node =
node              210 fs/befs/btree.c 	    (befs_btree_nodehead *) ((void *) node->bh->b_data + off);
node              212 fs/befs/btree.c 	befs_dump_index_node(sb, node->od_node);
node              214 fs/befs/btree.c 	node->head.left = fs64_to_cpu(sb, node->od_node->left);
node              215 fs/befs/btree.c 	node->head.right = fs64_to_cpu(sb, node->od_node->right);
node              216 fs/befs/btree.c 	node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow);
node              217 fs/befs/btree.c 	node->head.all_key_count =
node              218 fs/befs/btree.c 	    fs16_to_cpu(sb, node->od_node->all_key_count);
node              219 fs/befs/btree.c 	node->head.all_key_length =
node              220 fs/befs/btree.c 	    fs16_to_cpu(sb, node->od_node->all_key_length);
node              330 fs/befs/btree.c befs_find_key(struct super_block *sb, struct befs_btree_node *node,
node              345 fs/befs/btree.c 	last = node->head.all_key_count - 1;
node              346 fs/befs/btree.c 	thiskey = befs_bt_get_key(sb, node, last, &keylen);
node              354 fs/befs/btree.c 	valarray = befs_bt_valarray(node);
node              363 fs/befs/btree.c 		thiskey = befs_bt_get_key(sb, node, mid, &keylen);
node              599 fs/befs/btree.c befs_leafnode(struct befs_btree_node *node)
node              602 fs/befs/btree.c 	if (node->head.overflow == BEFS_BT_INVAL)
node              622 fs/befs/btree.c befs_bt_keylen_index(struct befs_btree_node *node)
node              626 fs/befs/btree.c 	    (sizeof (befs_btree_nodehead) + node->head.all_key_length);
node              632 fs/befs/btree.c 	return (fs16 *) ((void *) node->od_node + off);
node              643 fs/befs/btree.c befs_bt_valarray(struct befs_btree_node *node)
node              645 fs/befs/btree.c 	void *keylen_index_start = (void *) befs_bt_keylen_index(node);
node              646 fs/befs/btree.c 	size_t keylen_index_size = node->head.all_key_count * sizeof (fs16);
node              659 fs/befs/btree.c befs_bt_keydata(struct befs_btree_node *node)
node              661 fs/befs/btree.c 	return (char *) ((void *) node->od_node + sizeof (befs_btree_nodehead));
node              675 fs/befs/btree.c befs_bt_get_key(struct super_block *sb, struct befs_btree_node *node,
node              682 fs/befs/btree.c 	if (index < 0 || index > node->head.all_key_count) {
node              687 fs/befs/btree.c 	keystart = befs_bt_keydata(node);
node              688 fs/befs/btree.c 	keylen_index = befs_bt_keylen_index(node);
node              248 fs/befs/debug.c befs_dump_index_node(const struct super_block *sb, befs_btree_nodehead *node)
node              253 fs/befs/debug.c 	befs_debug(sb, "  left %016LX", fs64_to_cpu(sb, node->left));
node              254 fs/befs/debug.c 	befs_debug(sb, "  right %016LX", fs64_to_cpu(sb, node->right));
node              255 fs/befs/debug.c 	befs_debug(sb, "  overflow %016LX", fs64_to_cpu(sb, node->overflow));
node              257 fs/befs/debug.c 		   fs16_to_cpu(sb, node->all_key_count));
node              259 fs/befs/debug.c 		   fs16_to_cpu(sb, node->all_key_length));
node              539 fs/btrfs/backref.c 		root_level = btrfs_header_level(root->node);
node              586 fs/btrfs/backref.c unode_aux_to_inode_list(struct ulist_node *node)
node              588 fs/btrfs/backref.c 	if (!node)
node              590 fs/btrfs/backref.c 	return (struct extent_inode_elem *)(uintptr_t)node->aux;
node              618 fs/btrfs/backref.c 	struct ulist_node *node;
node              675 fs/btrfs/backref.c 		node = ulist_next(parents, &uiter);
node              676 fs/btrfs/backref.c 		ref->parent = node ? node->val : 0;
node              677 fs/btrfs/backref.c 		ref->inode_list = unode_aux_to_inode_list(node);
node              680 fs/btrfs/backref.c 		while ((node = ulist_next(parents, &uiter))) {
node              691 fs/btrfs/backref.c 			new_ref->parent = node->val;
node              692 fs/btrfs/backref.c 			new_ref->inode_list = unode_aux_to_inode_list(node);
node              720 fs/btrfs/backref.c 	struct rb_node *node;
node              722 fs/btrfs/backref.c 	while ((node = rb_first_cached(&tree->root))) {
node              723 fs/btrfs/backref.c 		ref = rb_entry(node, struct prelim_ref, rbnode);
node              724 fs/btrfs/backref.c 		rb_erase_cached(node, &tree->root);
node              764 fs/btrfs/backref.c 	struct btrfs_delayed_ref_node *node;
node              777 fs/btrfs/backref.c 		node = rb_entry(n, struct btrfs_delayed_ref_node,
node              779 fs/btrfs/backref.c 		if (node->seq > seq)
node              782 fs/btrfs/backref.c 		switch (node->action) {
node              788 fs/btrfs/backref.c 			count = node->ref_mod;
node              791 fs/btrfs/backref.c 			count = node->ref_mod * -1;
node              797 fs/btrfs/backref.c 		switch (node->type) {
node              802 fs/btrfs/backref.c 			ref = btrfs_delayed_node_to_tree_ref(node);
node              805 fs/btrfs/backref.c 					       node->bytenr, count, sc,
node              813 fs/btrfs/backref.c 			ref = btrfs_delayed_node_to_tree_ref(node);
node              816 fs/btrfs/backref.c 					     ref->parent, node->bytenr, count,
node              823 fs/btrfs/backref.c 			ref = btrfs_delayed_node_to_data_ref(node);
node              839 fs/btrfs/backref.c 					       &key, 0, node->bytenr, count, sc,
node              847 fs/btrfs/backref.c 			ref = btrfs_delayed_node_to_data_ref(node);
node              850 fs/btrfs/backref.c 					     node->bytenr, count, sc,
node             1126 fs/btrfs/backref.c 	struct rb_node *node;
node             1252 fs/btrfs/backref.c 	node = rb_first_cached(&preftrees.direct.root);
node             1253 fs/btrfs/backref.c 	while (node) {
node             1254 fs/btrfs/backref.c 		ref = rb_entry(node, struct prelim_ref, rbnode);
node             1255 fs/btrfs/backref.c 		node = rb_next(&ref->rbnode);
node             1341 fs/btrfs/backref.c 	struct ulist_node *node = NULL;
node             1346 fs/btrfs/backref.c 	while ((node = ulist_next(blocks, &uiter))) {
node             1347 fs/btrfs/backref.c 		if (!node->aux)
node             1349 fs/btrfs/backref.c 		eie = unode_aux_to_inode_list(node);
node             1351 fs/btrfs/backref.c 		node->aux = 0;
node             1405 fs/btrfs/backref.c 	struct ulist_node *node = NULL;
node             1427 fs/btrfs/backref.c 		node = ulist_next(tmp, &uiter);
node             1428 fs/btrfs/backref.c 		if (!node)
node             1430 fs/btrfs/backref.c 		bytenr = node->val;
node             1474 fs/btrfs/backref.c 	struct ulist_node *node;
node             1510 fs/btrfs/backref.c 		node = ulist_next(tmp, &uiter);
node             1511 fs/btrfs/backref.c 		if (!node)
node             1513 fs/btrfs/backref.c 		bytenr = node->val;
node              254 fs/btrfs/block-group.c 	struct rb_node *node;
node              266 fs/btrfs/block-group.c 	node = rb_next(&cache->cache_node);
node              268 fs/btrfs/block-group.c 	if (node) {
node              269 fs/btrfs/block-group.c 		cache = rb_entry(node, struct btrfs_block_group_cache,
node              132 fs/btrfs/ctree.c 		eb = rcu_dereference(root->node);
node              161 fs/btrfs/ctree.c 		if (eb == root->node)
node              180 fs/btrfs/ctree.c 		if (eb == root->node)
node              288 fs/btrfs/ctree.c 	struct rb_node node;
node              346 fs/btrfs/ctree.c 	struct rb_node *node;
node              379 fs/btrfs/ctree.c 	for (node = rb_first(tm_root); node; node = next) {
node              380 fs/btrfs/ctree.c 		next = rb_next(node);
node              381 fs/btrfs/ctree.c 		tm = rb_entry(node, struct tree_mod_elem, node);
node              384 fs/btrfs/ctree.c 		rb_erase(node, tm_root);
node              413 fs/btrfs/ctree.c 		cur = rb_entry(*new, struct tree_mod_elem, node);
node              427 fs/btrfs/ctree.c 	rb_link_node(&tm->node, parent, new);
node              428 fs/btrfs/ctree.c 	rb_insert_color(&tm->node, tm_root);
node              486 fs/btrfs/ctree.c 	RB_CLEAR_NODE(&tm->node);
node              578 fs/btrfs/ctree.c 		if (tm_list[i] && !RB_EMPTY_NODE(&tm_list[i]->node))
node              579 fs/btrfs/ctree.c 			rb_erase(&tm_list[i]->node, &eb->fs_info->tree_mod_log);
node              602 fs/btrfs/ctree.c 				rb_erase(&tm_list[j]->node,
node              685 fs/btrfs/ctree.c 	struct rb_node *node;
node              691 fs/btrfs/ctree.c 	node = tm_root->rb_node;
node              692 fs/btrfs/ctree.c 	while (node) {
node              693 fs/btrfs/ctree.c 		cur = rb_entry(node, struct tree_mod_elem, node);
node              695 fs/btrfs/ctree.c 			node = node->rb_left;
node              697 fs/btrfs/ctree.c 			node = node->rb_right;
node              699 fs/btrfs/ctree.c 			node = node->rb_left;
node              705 fs/btrfs/ctree.c 			node = node->rb_left;
node              711 fs/btrfs/ctree.c 			node = node->rb_right;
node              805 fs/btrfs/ctree.c 		if (tm_list[i] && !RB_EMPTY_NODE(&tm_list[i]->node))
node              806 fs/btrfs/ctree.c 			rb_erase(&tm_list[i]->node, &fs_info->tree_mod_log);
node              875 fs/btrfs/ctree.c 	    buf != root->node && buf != root->commit_root &&
node             1118 fs/btrfs/ctree.c 	if (buf == root->node) {
node             1125 fs/btrfs/ctree.c 		ret = tree_mod_log_insert_root(root->node, cow, 1);
node             1127 fs/btrfs/ctree.c 		rcu_assign_pointer(root->node, cow);
node             1279 fs/btrfs/ctree.c 		next = rb_next(&tm->node);
node             1282 fs/btrfs/ctree.c 		tm = rb_entry(next, struct tree_mod_elem, node);
node             1899 fs/btrfs/ctree.c 		ret = tree_mod_log_insert_root(root->node, child, 1);
node             1901 fs/btrfs/ctree.c 		rcu_assign_pointer(root->node, child);
node             2215 fs/btrfs/ctree.c 	struct extent_buffer *node;
node             2232 fs/btrfs/ctree.c 	node = path->nodes[level];
node             2234 fs/btrfs/ctree.c 	search = btrfs_node_blockptr(node, slot);
node             2244 fs/btrfs/ctree.c 	nritems = btrfs_header_nritems(node);
node             2258 fs/btrfs/ctree.c 			btrfs_node_key(node, &disk_key, nr);
node             2262 fs/btrfs/ctree.c 		search = btrfs_node_blockptr(node, nr);
node             3398 fs/btrfs/ctree.c 	BUG_ON(path->nodes[level-1] != root->node);
node             3407 fs/btrfs/ctree.c 					 root->node->start, 0);
node             3423 fs/btrfs/ctree.c 	old = root->node;
node             3424 fs/btrfs/ctree.c 	ret = tree_mod_log_insert_root(root->node, c, 0);
node             3426 fs/btrfs/ctree.c 	rcu_assign_pointer(root->node, c);
node             3508 fs/btrfs/ctree.c 	if (c == root->node) {
node             4926 fs/btrfs/ctree.c 	if (nritems == 0 && parent == root->node) {
node             4927 fs/btrfs/ctree.c 		BUG_ON(btrfs_header_level(root->node) != 1);
node             4929 fs/btrfs/ctree.c 		btrfs_set_header_level(root->node, 0);
node             5023 fs/btrfs/ctree.c 		if (leaf == root->node) {
node              475 fs/btrfs/ctree.h 	struct rb_node node;
node              961 fs/btrfs/ctree.h 	struct extent_buffer *node;
node             2622 fs/btrfs/ctree.h 			struct extent_buffer *node,
node             2692 fs/btrfs/ctree.h 			 struct extent_buffer *node);
node               71 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *node;
node               73 fs/btrfs/delayed-inode.c 	node = READ_ONCE(btrfs_inode->delayed_node);
node               74 fs/btrfs/delayed-inode.c 	if (node) {
node               75 fs/btrfs/delayed-inode.c 		refcount_inc(&node->refs);
node               76 fs/btrfs/delayed-inode.c 		return node;
node               80 fs/btrfs/delayed-inode.c 	node = radix_tree_lookup(&root->delayed_nodes_tree, ino);
node               82 fs/btrfs/delayed-inode.c 	if (node) {
node               84 fs/btrfs/delayed-inode.c 			refcount_inc(&node->refs);	/* can be accessed */
node               85 fs/btrfs/delayed-inode.c 			BUG_ON(btrfs_inode->delayed_node != node);
node               87 fs/btrfs/delayed-inode.c 			return node;
node              106 fs/btrfs/delayed-inode.c 		if (refcount_inc_not_zero(&node->refs)) {
node              107 fs/btrfs/delayed-inode.c 			refcount_inc(&node->refs);
node              108 fs/btrfs/delayed-inode.c 			btrfs_inode->delayed_node = node;
node              110 fs/btrfs/delayed-inode.c 			node = NULL;
node              114 fs/btrfs/delayed-inode.c 		return node;
node              125 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *node;
node              131 fs/btrfs/delayed-inode.c 	node = btrfs_get_delayed_node(btrfs_inode);
node              132 fs/btrfs/delayed-inode.c 	if (node)
node              133 fs/btrfs/delayed-inode.c 		return node;
node              135 fs/btrfs/delayed-inode.c 	node = kmem_cache_zalloc(delayed_node_cache, GFP_NOFS);
node              136 fs/btrfs/delayed-inode.c 	if (!node)
node              138 fs/btrfs/delayed-inode.c 	btrfs_init_delayed_node(node, root, ino);
node              141 fs/btrfs/delayed-inode.c 	refcount_set(&node->refs, 2);
node              145 fs/btrfs/delayed-inode.c 		kmem_cache_free(delayed_node_cache, node);
node              150 fs/btrfs/delayed-inode.c 	ret = radix_tree_insert(&root->delayed_nodes_tree, ino, node);
node              153 fs/btrfs/delayed-inode.c 		kmem_cache_free(delayed_node_cache, node);
node              157 fs/btrfs/delayed-inode.c 	btrfs_inode->delayed_node = node;
node              161 fs/btrfs/delayed-inode.c 	return node;
node              170 fs/btrfs/delayed-inode.c 				     struct btrfs_delayed_node *node,
node              174 fs/btrfs/delayed-inode.c 	if (test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) {
node              175 fs/btrfs/delayed-inode.c 		if (!list_empty(&node->p_list))
node              176 fs/btrfs/delayed-inode.c 			list_move_tail(&node->p_list, &root->prepare_list);
node              178 fs/btrfs/delayed-inode.c 			list_add_tail(&node->p_list, &root->prepare_list);
node              180 fs/btrfs/delayed-inode.c 		list_add_tail(&node->n_list, &root->node_list);
node              181 fs/btrfs/delayed-inode.c 		list_add_tail(&node->p_list, &root->prepare_list);
node              182 fs/btrfs/delayed-inode.c 		refcount_inc(&node->refs);	/* inserted into list */
node              184 fs/btrfs/delayed-inode.c 		set_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags);
node              191 fs/btrfs/delayed-inode.c 				       struct btrfs_delayed_node *node)
node              194 fs/btrfs/delayed-inode.c 	if (test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) {
node              196 fs/btrfs/delayed-inode.c 		refcount_dec(&node->refs);	/* not in the list */
node              197 fs/btrfs/delayed-inode.c 		list_del_init(&node->n_list);
node              198 fs/btrfs/delayed-inode.c 		if (!list_empty(&node->p_list))
node              199 fs/btrfs/delayed-inode.c 			list_del_init(&node->p_list);
node              200 fs/btrfs/delayed-inode.c 		clear_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags);
node              209 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *node = NULL;
node              216 fs/btrfs/delayed-inode.c 	node = list_entry(p, struct btrfs_delayed_node, n_list);
node              217 fs/btrfs/delayed-inode.c 	refcount_inc(&node->refs);
node              221 fs/btrfs/delayed-inode.c 	return node;
node              225 fs/btrfs/delayed-inode.c 						struct btrfs_delayed_node *node)
node              231 fs/btrfs/delayed-inode.c 	delayed_root = node->root->fs_info->delayed_root;
node              233 fs/btrfs/delayed-inode.c 	if (!test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) {
node              238 fs/btrfs/delayed-inode.c 	} else if (list_is_last(&node->n_list, &delayed_root->node_list))
node              241 fs/btrfs/delayed-inode.c 		p = node->n_list.next;
node              285 fs/btrfs/delayed-inode.c static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node)
node              287 fs/btrfs/delayed-inode.c 	__btrfs_release_delayed_node(node, 0);
node              294 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *node = NULL;
node              302 fs/btrfs/delayed-inode.c 	node = list_entry(p, struct btrfs_delayed_node, p_list);
node              303 fs/btrfs/delayed-inode.c 	refcount_inc(&node->refs);
node              307 fs/btrfs/delayed-inode.c 	return node;
node              311 fs/btrfs/delayed-inode.c 					struct btrfs_delayed_node *node)
node              313 fs/btrfs/delayed-inode.c 	__btrfs_release_delayed_node(node, 1);
node              346 fs/btrfs/delayed-inode.c 	struct rb_node *node, *prev_node = NULL;
node              350 fs/btrfs/delayed-inode.c 	node = root->rb_node;
node              352 fs/btrfs/delayed-inode.c 	while (node) {
node              353 fs/btrfs/delayed-inode.c 		delayed_item = rb_entry(node, struct btrfs_delayed_item,
node              355 fs/btrfs/delayed-inode.c 		prev_node = node;
node              358 fs/btrfs/delayed-inode.c 			node = node->rb_right;
node              360 fs/btrfs/delayed-inode.c 			node = node->rb_left;
node              370 fs/btrfs/delayed-inode.c 		else if ((node = rb_prev(prev_node)) != NULL) {
node              371 fs/btrfs/delayed-inode.c 			*prev = rb_entry(node, struct btrfs_delayed_item,
node              382 fs/btrfs/delayed-inode.c 		else if ((node = rb_next(prev_node)) != NULL) {
node              383 fs/btrfs/delayed-inode.c 			*next = rb_entry(node, struct btrfs_delayed_item,
node              403 fs/btrfs/delayed-inode.c 	struct rb_node **p, *node;
node              417 fs/btrfs/delayed-inode.c 	node = &ins->rb_node;
node              435 fs/btrfs/delayed-inode.c 	rb_link_node(node, parent_node, p);
node              436 fs/btrfs/delayed-inode.c 	rb_insert_color_cached(node, root, leftmost);
node              450 fs/btrfs/delayed-inode.c static int __btrfs_add_delayed_insertion_item(struct btrfs_delayed_node *node,
node              453 fs/btrfs/delayed-inode.c 	return __btrfs_add_delayed_item(node, item,
node              457 fs/btrfs/delayed-inode.c static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node,
node              460 fs/btrfs/delayed-inode.c 	return __btrfs_add_delayed_item(node, item,
node              606 fs/btrfs/delayed-inode.c 					struct btrfs_delayed_node *node)
node              647 fs/btrfs/delayed-inode.c 			node->bytes_reserved = num_bytes;
node              662 fs/btrfs/delayed-inode.c 		node->bytes_reserved = num_bytes;
node              669 fs/btrfs/delayed-inode.c 						struct btrfs_delayed_node *node,
node              674 fs/btrfs/delayed-inode.c 	if (!node->bytes_reserved)
node              679 fs/btrfs/delayed-inode.c 				      node->inode_id, node->bytes_reserved, 0);
node              681 fs/btrfs/delayed-inode.c 				node->bytes_reserved);
node              683 fs/btrfs/delayed-inode.c 		btrfs_qgroup_free_meta_prealloc(node->root,
node              684 fs/btrfs/delayed-inode.c 				node->bytes_reserved);
node              686 fs/btrfs/delayed-inode.c 		btrfs_qgroup_convert_reserved_meta(node->root,
node              687 fs/btrfs/delayed-inode.c 				node->bytes_reserved);
node              688 fs/btrfs/delayed-inode.c 	node->bytes_reserved = 0;
node              838 fs/btrfs/delayed-inode.c 				      struct btrfs_delayed_node *node)
node              844 fs/btrfs/delayed-inode.c 	mutex_lock(&node->mutex);
node              845 fs/btrfs/delayed-inode.c 	curr = __btrfs_first_delayed_insertion_item(node);
node              866 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node              870 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node              940 fs/btrfs/delayed-inode.c 				      struct btrfs_delayed_node *node)
node              947 fs/btrfs/delayed-inode.c 	mutex_lock(&node->mutex);
node              948 fs/btrfs/delayed-inode.c 	curr = __btrfs_first_delayed_deletion_item(node);
node              968 fs/btrfs/delayed-inode.c 			mutex_unlock(&node->mutex);
node              976 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node              981 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node             1015 fs/btrfs/delayed-inode.c 					struct btrfs_delayed_node *node)
node             1025 fs/btrfs/delayed-inode.c 	key.objectid = node->inode_id;
node             1029 fs/btrfs/delayed-inode.c 	if (test_bit(BTRFS_DELAYED_NODE_DEL_IREF, &node->flags))
node             1047 fs/btrfs/delayed-inode.c 	write_extent_buffer(leaf, &node->inode_item, (unsigned long)inode_item,
node             1051 fs/btrfs/delayed-inode.c 	if (!test_bit(BTRFS_DELAYED_NODE_DEL_IREF, &node->flags))
node             1059 fs/btrfs/delayed-inode.c 	if (key.objectid != node->inode_id)
node             1073 fs/btrfs/delayed-inode.c 	btrfs_release_delayed_iref(node);
node             1077 fs/btrfs/delayed-inode.c 	btrfs_delayed_inode_release_metadata(fs_info, node, (ret < 0));
node             1078 fs/btrfs/delayed-inode.c 	btrfs_release_delayed_inode(node);
node             1104 fs/btrfs/delayed-inode.c 					     struct btrfs_delayed_node *node)
node             1108 fs/btrfs/delayed-inode.c 	mutex_lock(&node->mutex);
node             1109 fs/btrfs/delayed-inode.c 	if (!test_bit(BTRFS_DELAYED_NODE_INODE_DIRTY, &node->flags)) {
node             1110 fs/btrfs/delayed-inode.c 		mutex_unlock(&node->mutex);
node             1114 fs/btrfs/delayed-inode.c 	ret = __btrfs_update_delayed_inode(trans, root, path, node);
node             1115 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node             1122 fs/btrfs/delayed-inode.c 				   struct btrfs_delayed_node *node)
node             1126 fs/btrfs/delayed-inode.c 	ret = btrfs_insert_delayed_items(trans, path, node->root, node);
node             1130 fs/btrfs/delayed-inode.c 	ret = btrfs_delete_delayed_items(trans, path, node->root, node);
node             1134 fs/btrfs/delayed-inode.c 	ret = btrfs_update_delayed_inode(trans, node->root, path, node);
node             1493 fs/btrfs/delayed-inode.c 					       struct btrfs_delayed_node *node,
node             1498 fs/btrfs/delayed-inode.c 	mutex_lock(&node->mutex);
node             1499 fs/btrfs/delayed-inode.c 	item = __btrfs_lookup_delayed_insertion_item(node, key);
node             1501 fs/btrfs/delayed-inode.c 		mutex_unlock(&node->mutex);
node             1505 fs/btrfs/delayed-inode.c 	btrfs_delayed_item_release_metadata(node->root, item);
node             1507 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node             1514 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *node;
node             1519 fs/btrfs/delayed-inode.c 	node = btrfs_get_or_create_delayed_node(dir);
node             1520 fs/btrfs/delayed-inode.c 	if (IS_ERR(node))
node             1521 fs/btrfs/delayed-inode.c 		return PTR_ERR(node);
node             1527 fs/btrfs/delayed-inode.c 	ret = btrfs_delete_delayed_insertion_item(trans->fs_info, node,
node             1552 fs/btrfs/delayed-inode.c 	mutex_lock(&node->mutex);
node             1553 fs/btrfs/delayed-inode.c 	ret = __btrfs_add_delayed_deletion_item(node, item);
node             1557 fs/btrfs/delayed-inode.c 			  index, node->root->root_key.objectid,
node             1558 fs/btrfs/delayed-inode.c 			  node->inode_id, ret);
node             1562 fs/btrfs/delayed-inode.c 	mutex_unlock(&node->mutex);
node             1564 fs/btrfs/delayed-inode.c 	btrfs_release_delayed_node(node);
node              207 fs/btrfs/delayed-ref.c 	if (ref1->node.type == BTRFS_TREE_BLOCK_REF_KEY) {
node              227 fs/btrfs/delayed-ref.c 	if (ref1->node.type == BTRFS_EXTENT_DATA_REF_KEY) {
node              279 fs/btrfs/delayed-ref.c 						   struct rb_node *node)
node              288 fs/btrfs/delayed-ref.c 	ins = rb_entry(node, struct btrfs_delayed_ref_head, href_node);
node              305 fs/btrfs/delayed-ref.c 	rb_link_node(node, parent_node, p);
node              306 fs/btrfs/delayed-ref.c 	rb_insert_color_cached(node, root, leftmost);
node              314 fs/btrfs/delayed-ref.c 	struct rb_node *node = &ins->ref_node;
node              336 fs/btrfs/delayed-ref.c 	rb_link_node(node, parent_node, p);
node              337 fs/btrfs/delayed-ref.c 	rb_insert_color_cached(node, root, leftmost);
node              437 fs/btrfs/delayed-ref.c 	struct rb_node *node = rb_next(&ref->ref_node);
node              440 fs/btrfs/delayed-ref.c 	while (!done && node) {
node              443 fs/btrfs/delayed-ref.c 		next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
node              444 fs/btrfs/delayed-ref.c 		node = rb_next(node);
node              483 fs/btrfs/delayed-ref.c 	struct rb_node *node;
node              506 fs/btrfs/delayed-ref.c 	for (node = rb_first_cached(&head->ref_tree); node;
node              507 fs/btrfs/delayed-ref.c 	     node = rb_next(node)) {
node              508 fs/btrfs/delayed-ref.c 		ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
node              554 fs/btrfs/delayed-ref.c 		struct rb_node *node;
node              556 fs/btrfs/delayed-ref.c 		node = rb_next(&head->href_node);
node              557 fs/btrfs/delayed-ref.c 		if (!node) {
node              563 fs/btrfs/delayed-ref.c 		head = rb_entry(node, struct btrfs_delayed_ref_head,
node              962 fs/btrfs/delayed-ref.c 	init_delayed_ref_common(fs_info, &ref->node, bytenr, num_bytes,
node              984 fs/btrfs/delayed-ref.c 	ret = insert_delayed_ref(trans, delayed_refs, head_ref, &ref->node);
node              993 fs/btrfs/delayed-ref.c 	trace_add_delayed_tree_ref(fs_info, &ref->node, ref,
node             1038 fs/btrfs/delayed-ref.c 	init_delayed_ref_common(fs_info, &ref->node, bytenr, num_bytes,
node             1080 fs/btrfs/delayed-ref.c 	ret = insert_delayed_ref(trans, delayed_refs, head_ref, &ref->node);
node             1089 fs/btrfs/delayed-ref.c 	trace_add_delayed_data_ref(trans->fs_info, &ref->node, ref,
node              124 fs/btrfs/delayed-ref.h 	struct btrfs_delayed_ref_node node;
node              131 fs/btrfs/delayed-ref.h 	struct btrfs_delayed_ref_node node;
node              381 fs/btrfs/delayed-ref.h btrfs_delayed_node_to_tree_ref(struct btrfs_delayed_ref_node *node)
node              383 fs/btrfs/delayed-ref.h 	return container_of(node, struct btrfs_delayed_tree_ref, node);
node              387 fs/btrfs/delayed-ref.h btrfs_delayed_node_to_data_ref(struct btrfs_delayed_ref_node *node)
node              389 fs/btrfs/delayed-ref.h 	return container_of(node, struct btrfs_delayed_data_ref, node);
node             1125 fs/btrfs/disk-io.c 	root->node = NULL;
node             1256 fs/btrfs/disk-io.c 	root->node = leaf;
node             1330 fs/btrfs/disk-io.c 	root->node = leaf;
node             1332 fs/btrfs/disk-io.c 	btrfs_mark_buffer_dirty(root->node);
node             1333 fs/btrfs/disk-io.c 	btrfs_tree_unlock(root->node);
node             1372 fs/btrfs/disk-io.c 	btrfs_set_root_node(&log_root->root_item, log_root->node);
node             1414 fs/btrfs/disk-io.c 	root->node = read_tree_block(fs_info,
node             1417 fs/btrfs/disk-io.c 	if (IS_ERR(root->node)) {
node             1418 fs/btrfs/disk-io.c 		ret = PTR_ERR(root->node);
node             1420 fs/btrfs/disk-io.c 	} else if (!btrfs_buffer_uptodate(root->node, generation, 0)) {
node             1422 fs/btrfs/disk-io.c 		free_extent_buffer(root->node);
node             1856 fs/btrfs/disk-io.c 	    btrfs_header_generation(info->tree_root->node))
node             1869 fs/btrfs/disk-io.c 	btrfs_set_backup_tree_root(root_backup, info->tree_root->node->start);
node             1871 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->tree_root->node));
node             1874 fs/btrfs/disk-io.c 			       btrfs_header_level(info->tree_root->node));
node             1876 fs/btrfs/disk-io.c 	btrfs_set_backup_chunk_root(root_backup, info->chunk_root->node->start);
node             1878 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->chunk_root->node));
node             1880 fs/btrfs/disk-io.c 			       btrfs_header_level(info->chunk_root->node));
node             1882 fs/btrfs/disk-io.c 	btrfs_set_backup_extent_root(root_backup, info->extent_root->node->start);
node             1884 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->extent_root->node));
node             1886 fs/btrfs/disk-io.c 			       btrfs_header_level(info->extent_root->node));
node             1892 fs/btrfs/disk-io.c 	if (info->fs_root && info->fs_root->node) {
node             1894 fs/btrfs/disk-io.c 					 info->fs_root->node->start);
node             1896 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->fs_root->node));
node             1898 fs/btrfs/disk-io.c 			       btrfs_header_level(info->fs_root->node));
node             1901 fs/btrfs/disk-io.c 	btrfs_set_backup_dev_root(root_backup, info->dev_root->node->start);
node             1903 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->dev_root->node));
node             1905 fs/btrfs/disk-io.c 				       btrfs_header_level(info->dev_root->node));
node             1907 fs/btrfs/disk-io.c 	btrfs_set_backup_csum_root(root_backup, info->csum_root->node->start);
node             1909 fs/btrfs/disk-io.c 			       btrfs_header_generation(info->csum_root->node));
node             1911 fs/btrfs/disk-io.c 			       btrfs_header_level(info->csum_root->node));
node             2011 fs/btrfs/disk-io.c 		free_extent_buffer(root->node);
node             2013 fs/btrfs/disk-io.c 		root->node = NULL;
node             2047 fs/btrfs/disk-io.c 			free_extent_buffer(gang[0]->node);
node             2264 fs/btrfs/disk-io.c 	log_tree_root->node = read_tree_block(fs_info, bytenr,
node             2267 fs/btrfs/disk-io.c 	if (IS_ERR(log_tree_root->node)) {
node             2269 fs/btrfs/disk-io.c 		ret = PTR_ERR(log_tree_root->node);
node             2272 fs/btrfs/disk-io.c 	} else if (!extent_buffer_uptodate(log_tree_root->node)) {
node             2274 fs/btrfs/disk-io.c 		free_extent_buffer(log_tree_root->node);
node             2283 fs/btrfs/disk-io.c 		free_extent_buffer(log_tree_root->node);
node             2987 fs/btrfs/disk-io.c 	chunk_root->node = read_tree_block(fs_info,
node             2990 fs/btrfs/disk-io.c 	if (IS_ERR(chunk_root->node) ||
node             2991 fs/btrfs/disk-io.c 	    !extent_buffer_uptodate(chunk_root->node)) {
node             2993 fs/btrfs/disk-io.c 		if (!IS_ERR(chunk_root->node))
node             2994 fs/btrfs/disk-io.c 			free_extent_buffer(chunk_root->node);
node             2995 fs/btrfs/disk-io.c 		chunk_root->node = NULL;
node             2998 fs/btrfs/disk-io.c 	btrfs_set_root_node(&chunk_root->root_item, chunk_root->node);
node             3001 fs/btrfs/disk-io.c 	read_extent_buffer(chunk_root->node, fs_info->chunk_tree_uuid,
node             3002 fs/btrfs/disk-io.c 	   btrfs_header_chunk_tree_uuid(chunk_root->node), BTRFS_UUID_SIZE);
node             3025 fs/btrfs/disk-io.c 	tree_root->node = read_tree_block(fs_info,
node             3028 fs/btrfs/disk-io.c 	if (IS_ERR(tree_root->node) ||
node             3029 fs/btrfs/disk-io.c 	    !extent_buffer_uptodate(tree_root->node)) {
node             3031 fs/btrfs/disk-io.c 		if (!IS_ERR(tree_root->node))
node             3032 fs/btrfs/disk-io.c 			free_extent_buffer(tree_root->node);
node             3033 fs/btrfs/disk-io.c 		tree_root->node = NULL;
node             3037 fs/btrfs/disk-io.c 	btrfs_set_root_node(&tree_root->root_item, tree_root->node);
node             3876 fs/btrfs/disk-io.c 			free_extent_buffer(root->reloc_root->node);
node             3898 fs/btrfs/disk-io.c 	free_extent_buffer(root->node);
node             4268 fs/btrfs/disk-io.c 	struct rb_node *node;
node             4282 fs/btrfs/disk-io.c 	while ((node = rb_first_cached(&delayed_refs->href_root)) != NULL) {
node             4287 fs/btrfs/disk-io.c 		head = rb_entry(node, struct btrfs_delayed_ref_head,
node               40 fs/btrfs/extent-tree.c 			       struct btrfs_delayed_ref_node *node, u64 parent,
node               52 fs/btrfs/extent-tree.c 				     struct btrfs_delayed_ref_node *node,
node             1432 fs/btrfs/extent-tree.c 				  struct btrfs_delayed_ref_node *node,
node             1441 fs/btrfs/extent-tree.c 	u64 bytenr = node->bytenr;
node             1442 fs/btrfs/extent-tree.c 	u64 num_bytes = node->num_bytes;
node             1488 fs/btrfs/extent-tree.c 				struct btrfs_delayed_ref_node *node,
node             1499 fs/btrfs/extent-tree.c 	ins.objectid = node->bytenr;
node             1500 fs/btrfs/extent-tree.c 	ins.offset = node->num_bytes;
node             1503 fs/btrfs/extent-tree.c 	ref = btrfs_delayed_node_to_data_ref(node);
node             1504 fs/btrfs/extent-tree.c 	trace_run_delayed_data_ref(trans->fs_info, node, ref, node->action);
node             1506 fs/btrfs/extent-tree.c 	if (node->type == BTRFS_SHARED_DATA_REF_KEY)
node             1510 fs/btrfs/extent-tree.c 	if (node->action == BTRFS_ADD_DELAYED_REF && insert_reserved) {
node             1516 fs/btrfs/extent-tree.c 						 node->ref_mod);
node             1517 fs/btrfs/extent-tree.c 	} else if (node->action == BTRFS_ADD_DELAYED_REF) {
node             1518 fs/btrfs/extent-tree.c 		ret = __btrfs_inc_extent_ref(trans, node, parent, ref_root,
node             1520 fs/btrfs/extent-tree.c 					     node->ref_mod, extent_op);
node             1521 fs/btrfs/extent-tree.c 	} else if (node->action == BTRFS_DROP_DELAYED_REF) {
node             1522 fs/btrfs/extent-tree.c 		ret = __btrfs_free_extent(trans, node, parent,
node             1524 fs/btrfs/extent-tree.c 					  ref->offset, node->ref_mod,
node             1638 fs/btrfs/extent-tree.c 				struct btrfs_delayed_ref_node *node,
node             1647 fs/btrfs/extent-tree.c 	ref = btrfs_delayed_node_to_tree_ref(node);
node             1648 fs/btrfs/extent-tree.c 	trace_run_delayed_tree_ref(trans->fs_info, node, ref, node->action);
node             1650 fs/btrfs/extent-tree.c 	if (node->type == BTRFS_SHARED_BLOCK_REF_KEY)
node             1654 fs/btrfs/extent-tree.c 	if (node->ref_mod != 1) {
node             1657 fs/btrfs/extent-tree.c 			  node->bytenr, node->ref_mod, node->action, ref_root,
node             1661 fs/btrfs/extent-tree.c 	if (node->action == BTRFS_ADD_DELAYED_REF && insert_reserved) {
node             1663 fs/btrfs/extent-tree.c 		ret = alloc_reserved_tree_block(trans, node, extent_op);
node             1664 fs/btrfs/extent-tree.c 	} else if (node->action == BTRFS_ADD_DELAYED_REF) {
node             1665 fs/btrfs/extent-tree.c 		ret = __btrfs_inc_extent_ref(trans, node, parent, ref_root,
node             1667 fs/btrfs/extent-tree.c 	} else if (node->action == BTRFS_DROP_DELAYED_REF) {
node             1668 fs/btrfs/extent-tree.c 		ret = __btrfs_free_extent(trans, node, parent, ref_root,
node             1678 fs/btrfs/extent-tree.c 			       struct btrfs_delayed_ref_node *node,
node             1686 fs/btrfs/extent-tree.c 			btrfs_pin_extent(trans->fs_info, node->bytenr,
node             1687 fs/btrfs/extent-tree.c 					 node->num_bytes, 1);
node             1691 fs/btrfs/extent-tree.c 	if (node->type == BTRFS_TREE_BLOCK_REF_KEY ||
node             1692 fs/btrfs/extent-tree.c 	    node->type == BTRFS_SHARED_BLOCK_REF_KEY)
node             1693 fs/btrfs/extent-tree.c 		ret = run_delayed_tree_ref(trans, node, extent_op,
node             1695 fs/btrfs/extent-tree.c 	else if (node->type == BTRFS_EXTENT_DATA_REF_KEY ||
node             1696 fs/btrfs/extent-tree.c 		 node->type == BTRFS_SHARED_DATA_REF_KEY)
node             1697 fs/btrfs/extent-tree.c 		ret = run_delayed_data_ref(trans, node, extent_op,
node             1702 fs/btrfs/extent-tree.c 		btrfs_pin_extent(trans->fs_info, node->bytenr,
node             1703 fs/btrfs/extent-tree.c 				 node->num_bytes, 1);
node             2165 fs/btrfs/extent-tree.c 	struct rb_node *node;
node             2196 fs/btrfs/extent-tree.c 		node = rb_first_cached(&delayed_refs->href_root);
node             2197 fs/btrfs/extent-tree.c 		if (!node) {
node             2201 fs/btrfs/extent-tree.c 		head = rb_entry(node, struct btrfs_delayed_ref_head,
node             2250 fs/btrfs/extent-tree.c 	struct rb_node *node;
node             2293 fs/btrfs/extent-tree.c 	for (node = rb_first_cached(&head->ref_tree); node;
node             2294 fs/btrfs/extent-tree.c 	     node = rb_next(node)) {
node             2295 fs/btrfs/extent-tree.c 		ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
node             2949 fs/btrfs/extent-tree.c 			       struct btrfs_delayed_ref_node *node, u64 parent,
node             2968 fs/btrfs/extent-tree.c 	u64 bytenr = node->bytenr;
node             2969 fs/btrfs/extent-tree.c 	u64 num_bytes = node->num_bytes;
node             4264 fs/btrfs/extent-tree.c 				     struct btrfs_delayed_ref_node *node,
node             4281 fs/btrfs/extent-tree.c 	ref = btrfs_delayed_node_to_tree_ref(node);
node             4283 fs/btrfs/extent-tree.c 	extent_key.objectid = node->bytenr;
node             4289 fs/btrfs/extent-tree.c 		extent_key.offset = node->num_bytes;
node             4292 fs/btrfs/extent-tree.c 		num_bytes = node->num_bytes;
node             4324 fs/btrfs/extent-tree.c 	if (node->type == BTRFS_SHARED_BLOCK_REF_KEY) {
node             5088 fs/btrfs/extent-tree.c 	if (eb == root->node) {
node             5247 fs/btrfs/extent-tree.c 		level = btrfs_header_level(root->node);
node             5276 fs/btrfs/extent-tree.c 		level = btrfs_header_level(root->node);
node             5400 fs/btrfs/extent-tree.c 		free_extent_buffer(root->node);
node             5433 fs/btrfs/extent-tree.c 			struct extent_buffer *node,
node             5462 fs/btrfs/extent-tree.c 	btrfs_assert_tree_locked(node);
node             5463 fs/btrfs/extent-tree.c 	level = btrfs_header_level(node);
node             5464 fs/btrfs/extent-tree.c 	path->nodes[level] = node;
node              273 fs/btrfs/extent_io.c 		struct rb_node *node;
node              276 fs/btrfs/extent_io.c 		node = rb_first(&tree->state);
node              277 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node              329 fs/btrfs/extent_io.c 				   struct rb_node *node,
node              357 fs/btrfs/extent_io.c 	rb_link_node(node, parent, p);
node              358 fs/btrfs/extent_io.c 	rb_insert_color(node, root);
node              523 fs/btrfs/extent_io.c 	struct rb_node *node;
node              535 fs/btrfs/extent_io.c 	node = tree_insert(&tree->state, NULL, end, &state->rb_node, p, parent);
node              536 fs/btrfs/extent_io.c 	if (node) {
node              538 fs/btrfs/extent_io.c 		found = rb_entry(node, struct extent_state, rb_node);
node              565 fs/btrfs/extent_io.c 	struct rb_node *node;
node              575 fs/btrfs/extent_io.c 	node = tree_insert(&tree->state, &orig->rb_node, prealloc->end,
node              577 fs/btrfs/extent_io.c 	if (node) {
node              676 fs/btrfs/extent_io.c 	struct rb_node *node;
node              727 fs/btrfs/extent_io.c 	node = tree_search(tree, start);
node              728 fs/btrfs/extent_io.c 	if (!node)
node              730 fs/btrfs/extent_io.c 	state = rb_entry(node, struct extent_state, rb_node);
node              845 fs/btrfs/extent_io.c 	struct rb_node *node;
node              856 fs/btrfs/extent_io.c 		node = tree_search(tree, start);
node              858 fs/btrfs/extent_io.c 		if (!node)
node              861 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node              879 fs/btrfs/extent_io.c 			node = rb_next(node);
node              944 fs/btrfs/extent_io.c 	struct rb_node *node;
node              971 fs/btrfs/extent_io.c 			node = &state->rb_node;
node              979 fs/btrfs/extent_io.c 	node = tree_search_for_insert(tree, start, &p, &parent);
node              980 fs/btrfs/extent_io.c 	if (!node) {
node              992 fs/btrfs/extent_io.c 	state = rb_entry(node, struct extent_state, rb_node);
node             1176 fs/btrfs/extent_io.c 	struct rb_node *node;
node             1207 fs/btrfs/extent_io.c 			node = &state->rb_node;
node             1216 fs/btrfs/extent_io.c 	node = tree_search_for_insert(tree, start, &p, &parent);
node             1217 fs/btrfs/extent_io.c 	if (!node) {
node             1231 fs/btrfs/extent_io.c 	state = rb_entry(node, struct extent_state, rb_node);
node             1493 fs/btrfs/extent_io.c 	struct rb_node *node;
node             1500 fs/btrfs/extent_io.c 	node = tree_search(tree, start);
node             1501 fs/btrfs/extent_io.c 	if (!node)
node             1505 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             1509 fs/btrfs/extent_io.c 		node = rb_next(node);
node             1510 fs/btrfs/extent_io.c 		if (!node)
node             1579 fs/btrfs/extent_io.c 	struct rb_node *node, *prev = NULL, *next;
node             1585 fs/btrfs/extent_io.c 		node = __etree_search(tree, start, &next, &prev, NULL, NULL);
node             1586 fs/btrfs/extent_io.c 		if (!node && !next && !prev) {
node             1594 fs/btrfs/extent_io.c 		} else if (!node && !next) {
node             1603 fs/btrfs/extent_io.c 		} else if (!node) {
node             1604 fs/btrfs/extent_io.c 			node = next;
node             1610 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             1661 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             1669 fs/btrfs/extent_io.c 		node = rb_next(node);
node             1670 fs/btrfs/extent_io.c 		if (!node)
node             1687 fs/btrfs/extent_io.c 	struct rb_node *node;
node             1699 fs/btrfs/extent_io.c 	node = tree_search(tree, cur_start);
node             1700 fs/btrfs/extent_io.c 	if (!node) {
node             1706 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             1724 fs/btrfs/extent_io.c 		node = rb_next(node);
node             1728 fs/btrfs/extent_io.c 		if (!node)
node             1966 fs/btrfs/extent_io.c 	struct rb_node *node;
node             1985 fs/btrfs/extent_io.c 	node = tree_search(tree, cur_start);
node             1986 fs/btrfs/extent_io.c 	if (!node)
node             1990 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             2008 fs/btrfs/extent_io.c 		node = rb_next(node);
node             2009 fs/btrfs/extent_io.c 		if (!node)
node             2024 fs/btrfs/extent_io.c 	struct rb_node *node;
node             2033 fs/btrfs/extent_io.c 	node = tree_search(tree, start);
node             2034 fs/btrfs/extent_io.c 	if (!node) {
node             2038 fs/btrfs/extent_io.c 	state = rb_entry(node, struct extent_state, rb_node);
node             2052 fs/btrfs/extent_io.c 	struct rb_node *node;
node             2061 fs/btrfs/extent_io.c 	node = tree_search(tree, start);
node             2062 fs/btrfs/extent_io.c 	if (!node) {
node             2066 fs/btrfs/extent_io.c 	state = rb_entry(node, struct extent_state, rb_node);
node             2087 fs/btrfs/extent_io.c 	struct rb_node *node;
node             2093 fs/btrfs/extent_io.c 		node = &cached->rb_node;
node             2095 fs/btrfs/extent_io.c 		node = tree_search(tree, start);
node             2096 fs/btrfs/extent_io.c 	while (node && start <= end) {
node             2097 fs/btrfs/extent_io.c 		state = rb_entry(node, struct extent_state, rb_node);
node             2122 fs/btrfs/extent_io.c 		node = rb_next(node);
node             2123 fs/btrfs/extent_io.c 		if (!node) {
node              254 fs/btrfs/file.c 	struct rb_node *node;
node              257 fs/btrfs/file.c 	node = rb_first(&fs_info->defrag_inodes);
node              258 fs/btrfs/file.c 	while (node) {
node              259 fs/btrfs/file.c 		rb_erase(node, &fs_info->defrag_inodes);
node              260 fs/btrfs/file.c 		defrag = rb_entry(node, struct inode_defrag, rb_node);
node              265 fs/btrfs/file.c 		node = rb_first(&fs_info->defrag_inodes);
node              929 fs/btrfs/free-space-cache.c 	struct rb_node *node = rb_first(&ctl->free_space_offset);
node              939 fs/btrfs/free-space-cache.c 	if (!node && cluster) {
node              942 fs/btrfs/free-space-cache.c 		node = rb_first(&cluster->root);
node              947 fs/btrfs/free-space-cache.c 	while (node) {
node              950 fs/btrfs/free-space-cache.c 		e = rb_entry(node, struct btrfs_free_space, offset_index);
node              962 fs/btrfs/free-space-cache.c 		node = rb_next(node);
node              963 fs/btrfs/free-space-cache.c 		if (!node && cluster) {
node              964 fs/btrfs/free-space-cache.c 			node = rb_first(&cluster->root);
node             1450 fs/btrfs/free-space-cache.c 			      struct rb_node *node, int bitmap)
node             1494 fs/btrfs/free-space-cache.c 	rb_link_node(node, parent, p);
node             1495 fs/btrfs/free-space-cache.c 	rb_insert_color(node, root);
node             1812 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             1824 fs/btrfs/free-space-cache.c 	for (node = &entry->offset_index; node; node = rb_next(node)) {
node             1825 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2073 fs/btrfs/free-space-cache.c 		struct rb_node *node;
node             2080 fs/btrfs/free-space-cache.c 		node = rb_first(&cluster->root);
node             2081 fs/btrfs/free-space-cache.c 		if (!node) {
node             2086 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2546 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             2556 fs/btrfs/free-space-cache.c 	node = rb_first(&cluster->root);
node             2557 fs/btrfs/free-space-cache.c 	while (node) {
node             2560 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2561 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
node             2585 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             2587 fs/btrfs/free-space-cache.c 	while ((node = rb_last(&ctl->free_space_offset)) != NULL) {
node             2588 fs/btrfs/free-space-cache.c 		info = rb_entry(node, struct btrfs_free_space, offset_index);
node             2756 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             2766 fs/btrfs/free-space-cache.c 	node = rb_first(&cluster->root);
node             2767 fs/btrfs/free-space-cache.c 	if (!node)
node             2770 fs/btrfs/free-space-cache.c 	entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2778 fs/btrfs/free-space-cache.c 			node = rb_next(&entry->offset_index);
node             2779 fs/btrfs/free-space-cache.c 			if (!node)
node             2781 fs/btrfs/free-space-cache.c 			entry = rb_entry(node, struct btrfs_free_space,
node             2792 fs/btrfs/free-space-cache.c 				node = rb_next(&entry->offset_index);
node             2793 fs/btrfs/free-space-cache.c 				if (!node)
node             2795 fs/btrfs/free-space-cache.c 				entry = rb_entry(node, struct btrfs_free_space,
node             2927 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             2943 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
node             2944 fs/btrfs/free-space-cache.c 		if (!node)
node             2946 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2954 fs/btrfs/free-space-cache.c 	for (node = rb_next(&entry->offset_index); node;
node             2955 fs/btrfs/free-space-cache.c 	     node = rb_next(&entry->offset_index)) {
node             2956 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2978 fs/btrfs/free-space-cache.c 	node = &first->offset_index;
node             2987 fs/btrfs/free-space-cache.c 		entry = rb_entry(node, struct btrfs_free_space, offset_index);
node             2988 fs/btrfs/free-space-cache.c 		node = rb_next(&entry->offset_index);
node             2997 fs/btrfs/free-space-cache.c 	} while (node && entry != last);
node             3200 fs/btrfs/free-space-cache.c 	struct rb_node *node;
node             3227 fs/btrfs/free-space-cache.c 			node = rb_next(&entry->offset_index);
node             3228 fs/btrfs/free-space-cache.c 			if (!node) {
node             3233 fs/btrfs/free-space-cache.c 			entry = rb_entry(node, struct btrfs_free_space,
node             1144 fs/btrfs/free-space-tree.c 	struct rb_node *node;
node             1160 fs/btrfs/free-space-tree.c 	node = rb_first(&fs_info->block_group_cache_tree);
node             1161 fs/btrfs/free-space-tree.c 	while (node) {
node             1162 fs/btrfs/free-space-tree.c 		block_group = rb_entry(node, struct btrfs_block_group_cache,
node             1167 fs/btrfs/free-space-tree.c 		node = rb_next(node);
node             1249 fs/btrfs/free-space-tree.c 	btrfs_tree_lock(free_space_root->node);
node             1250 fs/btrfs/free-space-tree.c 	btrfs_clean_tree_block(free_space_root->node);
node             1251 fs/btrfs/free-space-tree.c 	btrfs_tree_unlock(free_space_root->node);
node             1252 fs/btrfs/free-space-tree.c 	btrfs_free_tree_block(trans, free_space_root, free_space_root->node,
node             1255 fs/btrfs/free-space-tree.c 	free_extent_buffer(free_space_root->node);
node             2447 fs/btrfs/inode.c 	struct rb_node node;
node             2523 fs/btrfs/inode.c 		entry = rb_entry(parent, struct sa_defrag_extent_backref, node);
node             2532 fs/btrfs/inode.c 	rb_link_node(&backref->node, parent, p);
node             2533 fs/btrfs/inode.c 	rb_insert_color(&backref->node, root);
node             2932 fs/btrfs/inode.c 	struct rb_node *node;
node             2946 fs/btrfs/inode.c 		node = rb_first(&new->root);
node             2947 fs/btrfs/inode.c 		if (!node)
node             2949 fs/btrfs/inode.c 		rb_erase(node, &new->root);
node             2951 fs/btrfs/inode.c 		backref = rb_entry(node, struct sa_defrag_extent_backref, node);
node             4445 fs/btrfs/inode.c 	struct rb_node *node;
node             4456 fs/btrfs/inode.c 	node = root->inode_tree.rb_node;
node             4458 fs/btrfs/inode.c 	while (node) {
node             4459 fs/btrfs/inode.c 		prev = node;
node             4460 fs/btrfs/inode.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
node             4463 fs/btrfs/inode.c 			node = node->rb_left;
node             4465 fs/btrfs/inode.c 			node = node->rb_right;
node             4469 fs/btrfs/inode.c 	if (!node) {
node             4473 fs/btrfs/inode.c 				node = prev;
node             4479 fs/btrfs/inode.c 	while (node) {
node             4480 fs/btrfs/inode.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
node             4500 fs/btrfs/inode.c 		node = rb_next(node);
node             5428 fs/btrfs/inode.c 	struct rb_node *node;
node             5437 fs/btrfs/inode.c 		node = rb_first_cached(&map_tree->map);
node             5438 fs/btrfs/inode.c 		em = rb_entry(node, struct extent_map, rb_node);
node             5475 fs/btrfs/inode.c 		node = rb_first(&io_tree->state);
node             5476 fs/btrfs/inode.c 		state = rb_entry(node, struct extent_state, rb_node);
node             10753 fs/btrfs/inode.c 		entry = rb_entry(parent, struct btrfs_swapfile_pin, node);
node             10766 fs/btrfs/inode.c 	rb_link_node(&sp->node, parent, p);
node             10767 fs/btrfs/inode.c 	rb_insert_color(&sp->node, &fs_info->swapfile_pins);
node             10777 fs/btrfs/inode.c 	struct rb_node *node, *next;
node             10780 fs/btrfs/inode.c 	node = rb_first(&fs_info->swapfile_pins);
node             10781 fs/btrfs/inode.c 	while (node) {
node             10782 fs/btrfs/inode.c 		next = rb_next(node);
node             10783 fs/btrfs/inode.c 		sp = rb_entry(node, struct btrfs_swapfile_pin, node);
node             10785 fs/btrfs/inode.c 			rb_erase(&sp->node, &fs_info->swapfile_pins);
node             10790 fs/btrfs/inode.c 		node = next;
node               32 fs/btrfs/ordered-data.c 				   struct rb_node *node)
node               50 fs/btrfs/ordered-data.c 	rb_link_node(node, parent, p);
node               51 fs/btrfs/ordered-data.c 	rb_insert_color(node, root);
node              181 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              219 fs/btrfs/ordered-data.c 	node = tree_insert(&tree->tree, file_offset,
node              221 fs/btrfs/ordered-data.c 	if (node)
node              308 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              318 fs/btrfs/ordered-data.c 	node = tree_search(tree, *file_offset);
node              319 fs/btrfs/ordered-data.c 	if (!node) {
node              324 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node              378 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              390 fs/btrfs/ordered-data.c 	node = tree_search(tree, file_offset);
node              391 fs/btrfs/ordered-data.c 	if (!node) {
node              396 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node              467 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              482 fs/btrfs/ordered-data.c 	node = &entry->rb_node;
node              483 fs/btrfs/ordered-data.c 	rb_erase(node, &tree->tree);
node              484 fs/btrfs/ordered-data.c 	RB_CLEAR_NODE(node);
node              485 fs/btrfs/ordered-data.c 	if (tree->last == node)
node              716 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              721 fs/btrfs/ordered-data.c 	node = tree_search(tree, file_offset);
node              722 fs/btrfs/ordered-data.c 	if (!node)
node              725 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node              742 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              747 fs/btrfs/ordered-data.c 	node = tree_search(tree, file_offset);
node              748 fs/btrfs/ordered-data.c 	if (!node) {
node              749 fs/btrfs/ordered-data.c 		node = tree_search(tree, file_offset + len);
node              750 fs/btrfs/ordered-data.c 		if (!node)
node              755 fs/btrfs/ordered-data.c 		entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node              764 fs/btrfs/ordered-data.c 		node = rb_next(node);
node              765 fs/btrfs/ordered-data.c 		if (!node)
node              783 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              788 fs/btrfs/ordered-data.c 	node = tree_search(tree, file_offset);
node              789 fs/btrfs/ordered-data.c 	if (!node)
node              792 fs/btrfs/ordered-data.c 	entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node              810 fs/btrfs/ordered-data.c 	struct rb_node *node;
node              866 fs/btrfs/ordered-data.c 		node = rb_prev(&ordered->rb_node);
node              878 fs/btrfs/ordered-data.c 		node = prev;
node              880 fs/btrfs/ordered-data.c 	for (; node; node = rb_prev(node)) {
node              881 fs/btrfs/ordered-data.c 		test = rb_entry(node, struct btrfs_ordered_extent, rb_node);
node               18 fs/btrfs/props.c 	struct hlist_node node;
node               51 fs/btrfs/props.c 	hlist_for_each_entry(h, handlers, node)
node              446 fs/btrfs/props.c 		hash_add(prop_handlers_ht, &p->node, h);
node              177 fs/btrfs/qgroup.c 		qgroup = rb_entry(n, struct btrfs_qgroup, node);
node              198 fs/btrfs/qgroup.c 		qgroup = rb_entry(parent, struct btrfs_qgroup, node);
node              217 fs/btrfs/qgroup.c 	rb_link_node(&qgroup->node, parent, p);
node              218 fs/btrfs/qgroup.c 	rb_insert_color(&qgroup->node, &fs_info->qgroup_tree);
node              254 fs/btrfs/qgroup.c 	rb_erase(&qgroup->node, &fs_info->qgroup_tree);
node              520 fs/btrfs/qgroup.c 		qgroup = rb_entry(n, struct btrfs_qgroup, node);
node             1042 fs/btrfs/qgroup.c 		free_extent_buffer(quota_root->node);
node             1103 fs/btrfs/qgroup.c 	btrfs_tree_lock(quota_root->node);
node             1104 fs/btrfs/qgroup.c 	btrfs_clean_tree_block(quota_root->node);
node             1105 fs/btrfs/qgroup.c 	btrfs_tree_unlock(quota_root->node);
node             1106 fs/btrfs/qgroup.c 	btrfs_free_tree_block(trans, quota_root, quota_root->node, 0, 1);
node             1108 fs/btrfs/qgroup.c 	free_extent_buffer(quota_root->node);
node             1559 fs/btrfs/qgroup.c 				 node);
node             1574 fs/btrfs/qgroup.c 	rb_link_node(&record->node, parent_node, p);
node             1575 fs/btrfs/qgroup.c 	rb_insert_color(&record->node, &delayed_refs->dirty_extent_root);
node             2513 fs/btrfs/qgroup.c 	struct rb_node *node;
node             2520 fs/btrfs/qgroup.c 	while ((node = rb_first(&delayed_refs->dirty_extent_root))) {
node             2521 fs/btrfs/qgroup.c 		record = rb_entry(node, struct btrfs_qgroup_extent_record,
node             2522 fs/btrfs/qgroup.c 				  node);
node             2571 fs/btrfs/qgroup.c 		rb_erase(node, &delayed_refs->dirty_extent_root);
node             3300 fs/btrfs/qgroup.c 		qgroup = rb_entry(n, struct btrfs_qgroup, node);
node             3818 fs/btrfs/qgroup.c 						     node)
node             3903 fs/btrfs/qgroup.c 				 node);
node             3929 fs/btrfs/qgroup.c 	rb_link_node(&block->node, parent, cur);
node             3930 fs/btrfs/qgroup.c 	rb_insert_color(&block->node, &blocks->blocks[level]);
node             3955 fs/btrfs/qgroup.c 	struct rb_node *node;
node             3972 fs/btrfs/qgroup.c 	node = blocks->blocks[level].rb_node;
node             3974 fs/btrfs/qgroup.c 	while (node) {
node             3975 fs/btrfs/qgroup.c 		block = rb_entry(node, struct btrfs_qgroup_swapped_block, node);
node             3977 fs/btrfs/qgroup.c 			node = node->rb_left;
node             3979 fs/btrfs/qgroup.c 			node = node->rb_right;
node             3990 fs/btrfs/qgroup.c 	rb_erase(&block->node, &blocks->blocks[level]);
node             4036 fs/btrfs/qgroup.c 	rbtree_postorder_for_each_entry_safe(entry, next, root, node) {
node              107 fs/btrfs/qgroup.h 	struct rb_node node;
node              125 fs/btrfs/qgroup.h 	struct rb_node node;
node              218 fs/btrfs/qgroup.h 	struct rb_node node;	  /* tree of qgroups */
node              935 fs/btrfs/reada.c 	struct extent_buffer *node;
node              954 fs/btrfs/reada.c 	node = btrfs_root_node(root);
node              955 fs/btrfs/reada.c 	start = node->start;
node              956 fs/btrfs/reada.c 	generation = btrfs_header_generation(node);
node              957 fs/btrfs/reada.c 	free_extent_buffer(node);
node               22 fs/btrfs/ref-verify.c 	struct rb_node node;
node               36 fs/btrfs/ref-verify.c 	struct rb_node node;
node               71 fs/btrfs/ref-verify.c 	struct rb_node node;
node               84 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct block_entry, node);
node               93 fs/btrfs/ref-verify.c 	rb_link_node(&be->node, parent_node, p);
node               94 fs/btrfs/ref-verify.c 	rb_insert_color(&be->node, root);
node              105 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct block_entry, node);
node              125 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct root_entry, node);
node              134 fs/btrfs/ref-verify.c 	rb_link_node(&re->node, parent_node, p);
node              135 fs/btrfs/ref-verify.c 	rb_insert_color(&re->node, root);
node              171 fs/btrfs/ref-verify.c 		entry = rb_entry(parent_node, struct ref_entry, node);
node              181 fs/btrfs/ref-verify.c 	rb_link_node(&ref->node, parent_node, p);
node              182 fs/btrfs/ref-verify.c 	rb_insert_color(&ref->node, root);
node              194 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct root_entry, node);
node              240 fs/btrfs/ref-verify.c 		re = rb_entry(n, struct root_entry, node);
node              241 fs/btrfs/ref-verify.c 		rb_erase(&re->node, &be->roots);
node              246 fs/btrfs/ref-verify.c 		ref = rb_entry(n, struct ref_entry, node);
node              247 fs/btrfs/ref-verify.c 		rb_erase(&ref->node, &be->refs);
node              644 fs/btrfs/ref-verify.c 		ref = rb_entry(n, struct ref_entry, node);
node              652 fs/btrfs/ref-verify.c 		re = rb_entry(n, struct root_entry, node);
node              831 fs/btrfs/ref-verify.c 				rb_erase(&exist->node, &be->refs);
node              904 fs/btrfs/ref-verify.c 		be = rb_entry(n, struct block_entry, node);
node              905 fs/btrfs/ref-verify.c 		rb_erase(&be->node, &fs_info->block_tree);
node              924 fs/btrfs/ref-verify.c 		entry = rb_entry(n, struct block_entry, node);
node              949 fs/btrfs/ref-verify.c 	n = &be->node;
node              951 fs/btrfs/ref-verify.c 		be = rb_entry(n, struct block_entry, node);
node              970 fs/btrfs/ref-verify.c 		rb_erase(&be->node, &fs_info->block_tree);
node               83 fs/btrfs/relocation.c 	struct backref_node *node[2];
node              190 fs/btrfs/relocation.c 				struct backref_node *node);
node              192 fs/btrfs/relocation.c 				   struct backref_node *node);
node              213 fs/btrfs/relocation.c 	struct backref_node *node;
node              217 fs/btrfs/relocation.c 		node = list_entry(cache->detached.next,
node              219 fs/btrfs/relocation.c 		remove_backref_node(cache, node);
node              223 fs/btrfs/relocation.c 		node = list_entry(cache->leaves.next,
node              225 fs/btrfs/relocation.c 		remove_backref_node(cache, node);
node              241 fs/btrfs/relocation.c 	struct backref_node *node;
node              243 fs/btrfs/relocation.c 	node = kzalloc(sizeof(*node), GFP_NOFS);
node              244 fs/btrfs/relocation.c 	if (node) {
node              245 fs/btrfs/relocation.c 		INIT_LIST_HEAD(&node->list);
node              246 fs/btrfs/relocation.c 		INIT_LIST_HEAD(&node->upper);
node              247 fs/btrfs/relocation.c 		INIT_LIST_HEAD(&node->lower);
node              248 fs/btrfs/relocation.c 		RB_CLEAR_NODE(&node->rb_node);
node              251 fs/btrfs/relocation.c 	return node;
node              255 fs/btrfs/relocation.c 			      struct backref_node *node)
node              257 fs/btrfs/relocation.c 	if (node) {
node              259 fs/btrfs/relocation.c 		kfree(node);
node              283 fs/btrfs/relocation.c 				   struct rb_node *node)
node              301 fs/btrfs/relocation.c 	rb_link_node(node, parent, p);
node              302 fs/btrfs/relocation.c 	rb_insert_color(node, root);
node              340 fs/btrfs/relocation.c static struct backref_node *walk_up_backref(struct backref_node *node,
node              347 fs/btrfs/relocation.c 	while (!list_empty(&node->upper)) {
node              348 fs/btrfs/relocation.c 		edge = list_entry(node->upper.next,
node              351 fs/btrfs/relocation.c 		node = edge->node[UPPER];
node              353 fs/btrfs/relocation.c 	BUG_ON(node->detached);
node              355 fs/btrfs/relocation.c 	return node;
node              370 fs/btrfs/relocation.c 		lower = edge->node[LOWER];
node              379 fs/btrfs/relocation.c 		return edge->node[UPPER];
node              385 fs/btrfs/relocation.c static void unlock_node_buffer(struct backref_node *node)
node              387 fs/btrfs/relocation.c 	if (node->locked) {
node              388 fs/btrfs/relocation.c 		btrfs_tree_unlock(node->eb);
node              389 fs/btrfs/relocation.c 		node->locked = 0;
node              393 fs/btrfs/relocation.c static void drop_node_buffer(struct backref_node *node)
node              395 fs/btrfs/relocation.c 	if (node->eb) {
node              396 fs/btrfs/relocation.c 		unlock_node_buffer(node);
node              397 fs/btrfs/relocation.c 		free_extent_buffer(node->eb);
node              398 fs/btrfs/relocation.c 		node->eb = NULL;
node              403 fs/btrfs/relocation.c 			      struct backref_node *node)
node              405 fs/btrfs/relocation.c 	BUG_ON(!list_empty(&node->upper));
node              407 fs/btrfs/relocation.c 	drop_node_buffer(node);
node              408 fs/btrfs/relocation.c 	list_del(&node->list);
node              409 fs/btrfs/relocation.c 	list_del(&node->lower);
node              410 fs/btrfs/relocation.c 	if (!RB_EMPTY_NODE(&node->rb_node))
node              411 fs/btrfs/relocation.c 		rb_erase(&node->rb_node, &tree->rb_root);
node              412 fs/btrfs/relocation.c 	free_backref_node(tree, node);
node              419 fs/btrfs/relocation.c 				struct backref_node *node)
node              424 fs/btrfs/relocation.c 	if (!node)
node              427 fs/btrfs/relocation.c 	BUG_ON(!node->lowest && !node->detached);
node              428 fs/btrfs/relocation.c 	while (!list_empty(&node->upper)) {
node              429 fs/btrfs/relocation.c 		edge = list_entry(node->upper.next, struct backref_edge,
node              431 fs/btrfs/relocation.c 		upper = edge->node[UPPER];
node              437 fs/btrfs/relocation.c 			BUG_ON(!list_empty(&node->upper));
node              438 fs/btrfs/relocation.c 			drop_backref_node(cache, node);
node              439 fs/btrfs/relocation.c 			node = upper;
node              440 fs/btrfs/relocation.c 			node->lowest = 1;
node              453 fs/btrfs/relocation.c 	drop_backref_node(cache, node);
node              457 fs/btrfs/relocation.c 				struct backref_node *node, u64 bytenr)
node              460 fs/btrfs/relocation.c 	rb_erase(&node->rb_node, &cache->rb_root);
node              461 fs/btrfs/relocation.c 	node->bytenr = bytenr;
node              462 fs/btrfs/relocation.c 	rb_node = tree_insert(&cache->rb_root, node->bytenr, &node->rb_node);
node              473 fs/btrfs/relocation.c 	struct backref_node *node;
node              490 fs/btrfs/relocation.c 		node = list_entry(cache->detached.next,
node              492 fs/btrfs/relocation.c 		remove_backref_node(cache, node);
node              496 fs/btrfs/relocation.c 		node = list_entry(cache->changed.next,
node              498 fs/btrfs/relocation.c 		list_del_init(&node->list);
node              499 fs/btrfs/relocation.c 		BUG_ON(node->pending);
node              500 fs/btrfs/relocation.c 		update_backref_node(cache, node, node->new_bytenr);
node              508 fs/btrfs/relocation.c 		list_for_each_entry(node, &cache->pending[level], list) {
node              509 fs/btrfs/relocation.c 			BUG_ON(!node->pending);
node              510 fs/btrfs/relocation.c 			if (node->bytenr == node->new_bytenr)
node              512 fs/btrfs/relocation.c 			update_backref_node(cache, node, node->new_bytenr);
node              582 fs/btrfs/relocation.c 	struct mapping_node *node;
node              588 fs/btrfs/relocation.c 		node = rb_entry(rb_node, struct mapping_node, rb_node);
node              589 fs/btrfs/relocation.c 		root = (struct btrfs_root *)node->data;
node              694 fs/btrfs/relocation.c 	struct backref_node *node = NULL;
node              717 fs/btrfs/relocation.c 	node = alloc_backref_node(cache);
node              718 fs/btrfs/relocation.c 	if (!node) {
node              723 fs/btrfs/relocation.c 	node->bytenr = bytenr;
node              724 fs/btrfs/relocation.c 	node->level = level;
node              725 fs/btrfs/relocation.c 	node->lowest = 1;
node              726 fs/btrfs/relocation.c 	cur = node;
node              757 fs/btrfs/relocation.c 		exist = edge->node[UPPER];
node              870 fs/btrfs/relocation.c 			edge->node[LOWER] = cur;
node              871 fs/btrfs/relocation.c 			edge->node[UPPER] = upper;
node             1006 fs/btrfs/relocation.c 			edge->node[LOWER] = lower;
node             1007 fs/btrfs/relocation.c 			edge->node[UPPER] = upper;
node             1036 fs/btrfs/relocation.c 		cur = edge->node[UPPER];
node             1044 fs/btrfs/relocation.c 	ASSERT(node->checked);
node             1045 fs/btrfs/relocation.c 	cowonly = node->cowonly;
node             1047 fs/btrfs/relocation.c 		rb_node = tree_insert(&cache->rb_root, node->bytenr,
node             1048 fs/btrfs/relocation.c 				      &node->rb_node);
node             1050 fs/btrfs/relocation.c 			backref_tree_panic(rb_node, -EEXIST, node->bytenr);
node             1051 fs/btrfs/relocation.c 		list_add_tail(&node->lower, &cache->leaves);
node             1054 fs/btrfs/relocation.c 	list_for_each_entry(edge, &node->upper, list[LOWER])
node             1060 fs/btrfs/relocation.c 		upper = edge->node[UPPER];
node             1063 fs/btrfs/relocation.c 			lower = edge->node[LOWER];
node             1118 fs/btrfs/relocation.c 		if (upper == node)
node             1119 fs/btrfs/relocation.c 			node = NULL;
node             1129 fs/btrfs/relocation.c 			lower = edge->node[LOWER];
node             1158 fs/btrfs/relocation.c 			lower = edge->node[LOWER];
node             1159 fs/btrfs/relocation.c 			upper = edge->node[UPPER];
node             1184 fs/btrfs/relocation.c 			if (lower == node)
node             1185 fs/btrfs/relocation.c 				node = NULL;
node             1189 fs/btrfs/relocation.c 		remove_backref_node(cache, node);
node             1192 fs/btrfs/relocation.c 	ASSERT(!node || !node->detached);
node             1193 fs/btrfs/relocation.c 	return node;
node             1208 fs/btrfs/relocation.c 	struct backref_node *node = NULL;
node             1219 fs/btrfs/relocation.c 		node = rb_entry(rb_node, struct backref_node, rb_node);
node             1220 fs/btrfs/relocation.c 		if (node->detached)
node             1221 fs/btrfs/relocation.c 			node = NULL;
node             1223 fs/btrfs/relocation.c 			BUG_ON(node->new_bytenr != reloc_root->node->start);
node             1226 fs/btrfs/relocation.c 	if (!node) {
node             1230 fs/btrfs/relocation.c 			node = rb_entry(rb_node, struct backref_node,
node             1232 fs/btrfs/relocation.c 			BUG_ON(node->detached);
node             1236 fs/btrfs/relocation.c 	if (!node)
node             1243 fs/btrfs/relocation.c 	new_node->bytenr = dest->node->start;
node             1244 fs/btrfs/relocation.c 	new_node->level = node->level;
node             1245 fs/btrfs/relocation.c 	new_node->lowest = node->lowest;
node             1249 fs/btrfs/relocation.c 	if (!node->lowest) {
node             1250 fs/btrfs/relocation.c 		list_for_each_entry(edge, &node->lower, list[UPPER]) {
node             1255 fs/btrfs/relocation.c 			new_edge->node[UPPER] = new_node;
node             1256 fs/btrfs/relocation.c 			new_edge->node[LOWER] = edge->node[LOWER];
node             1272 fs/btrfs/relocation.c 				      &new_edge->node[LOWER]->upper);
node             1294 fs/btrfs/relocation.c 	struct mapping_node *node;
node             1297 fs/btrfs/relocation.c 	node = kmalloc(sizeof(*node), GFP_NOFS);
node             1298 fs/btrfs/relocation.c 	if (!node)
node             1301 fs/btrfs/relocation.c 	node->bytenr = root->commit_root->start;
node             1302 fs/btrfs/relocation.c 	node->data = root;
node             1306 fs/btrfs/relocation.c 			      node->bytenr, &node->rb_node);
node             1311 fs/btrfs/relocation.c 			    node->bytenr);
node             1326 fs/btrfs/relocation.c 	struct mapping_node *node = NULL;
node             1329 fs/btrfs/relocation.c 	if (rc && root->node) {
node             1334 fs/btrfs/relocation.c 			node = rb_entry(rb_node, struct mapping_node, rb_node);
node             1335 fs/btrfs/relocation.c 			rb_erase(&node->rb_node, &rc->reloc_root_tree.rb_root);
node             1336 fs/btrfs/relocation.c 			RB_CLEAR_NODE(&node->rb_node);
node             1339 fs/btrfs/relocation.c 		if (!node)
node             1341 fs/btrfs/relocation.c 		BUG_ON((struct btrfs_root *)node->data != root);
node             1347 fs/btrfs/relocation.c 	kfree(node);
node             1358 fs/btrfs/relocation.c 	struct mapping_node *node = NULL;
node             1365 fs/btrfs/relocation.c 		node = rb_entry(rb_node, struct mapping_node, rb_node);
node             1366 fs/btrfs/relocation.c 		rb_erase(&node->rb_node, &rc->reloc_root_tree.rb_root);
node             1370 fs/btrfs/relocation.c 	if (!node)
node             1372 fs/btrfs/relocation.c 	BUG_ON((struct btrfs_root *)node->data != root);
node             1375 fs/btrfs/relocation.c 	node->bytenr = root->node->start;
node             1377 fs/btrfs/relocation.c 			      node->bytenr, &node->rb_node);
node             1380 fs/btrfs/relocation.c 		backref_tree_panic(rb_node, -EEXIST, node->bytenr);
node             1426 fs/btrfs/relocation.c 		ret = btrfs_copy_root(trans, root, root->node, &eb,
node             1532 fs/btrfs/relocation.c 	if (reloc_root->commit_root != reloc_root->node) {
node             1534 fs/btrfs/relocation.c 		btrfs_set_root_node(root_item, reloc_root->node);
node             1553 fs/btrfs/relocation.c 	struct rb_node *node;
node             1560 fs/btrfs/relocation.c 	node = root->inode_tree.rb_node;
node             1562 fs/btrfs/relocation.c 	while (node) {
node             1563 fs/btrfs/relocation.c 		prev = node;
node             1564 fs/btrfs/relocation.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
node             1567 fs/btrfs/relocation.c 			node = node->rb_left;
node             1569 fs/btrfs/relocation.c 			node = node->rb_right;
node             1573 fs/btrfs/relocation.c 	if (!node) {
node             1577 fs/btrfs/relocation.c 				node = prev;
node             1583 fs/btrfs/relocation.c 	while (node) {
node             1584 fs/btrfs/relocation.c 		entry = rb_entry(node, struct btrfs_inode, rb_node);
node             1595 fs/btrfs/relocation.c 		node = rb_next(node);
node             2292 fs/btrfs/relocation.c 		extent_buffer_get(reloc_root->node);
node             2293 fs/btrfs/relocation.c 		path->nodes[level] = reloc_root->node;
node             2494 fs/btrfs/relocation.c 		free_extent_buffer(reloc_root->node);
node             2496 fs/btrfs/relocation.c 		reloc_root->node = NULL;
node             2614 fs/btrfs/relocation.c 				     struct backref_node *node,
node             2621 fs/btrfs/relocation.c 	next = node;
node             2637 fs/btrfs/relocation.c 		if (next->new_bytenr != root->node->start) {
node             2640 fs/btrfs/relocation.c 			next->new_bytenr = root->node->start;
node             2651 fs/btrfs/relocation.c 		if (!next || next->level <= node->level)
node             2657 fs/btrfs/relocation.c 	next = node;
node             2663 fs/btrfs/relocation.c 		next = edges[index]->node[UPPER];
node             2675 fs/btrfs/relocation.c struct btrfs_root *select_one_root(struct backref_node *node)
node             2683 fs/btrfs/relocation.c 	next = node;
node             2697 fs/btrfs/relocation.c 		if (next != node)
node             2701 fs/btrfs/relocation.c 		if (!next || next->level <= node->level)
node             2712 fs/btrfs/relocation.c 			struct backref_node *node, int reserve)
node             2715 fs/btrfs/relocation.c 	struct backref_node *next = node;
node             2721 fs/btrfs/relocation.c 	BUG_ON(reserve && node->processed);
node             2726 fs/btrfs/relocation.c 			if (next->processed && (reserve || next != node))
node             2737 fs/btrfs/relocation.c 			next = edge->node[UPPER];
node             2746 fs/btrfs/relocation.c 				  struct backref_node *node)
node             2754 fs/btrfs/relocation.c 	num_bytes = calcu_metadata_size(rc, node, 1) * 2;
node             2794 fs/btrfs/relocation.c 			 struct backref_node *node,
node             2811 fs/btrfs/relocation.c 	BUG_ON(lowest && node->eb);
node             2813 fs/btrfs/relocation.c 	path->lowest_level = node->level + 1;
node             2814 fs/btrfs/relocation.c 	rc->backref_cache.path[node->level] = node;
node             2815 fs/btrfs/relocation.c 	list_for_each_entry(edge, &node->upper, list[LOWER]) {
node             2821 fs/btrfs/relocation.c 		upper = edge->node[UPPER];
node             2835 fs/btrfs/relocation.c 				if (node->eb->start == bytenr)
node             2877 fs/btrfs/relocation.c 			if (bytenr != node->bytenr) {
node             2880 fs/btrfs/relocation.c 					  bytenr, node->bytenr, slot,
node             2886 fs/btrfs/relocation.c 			if (node->eb->start == bytenr)
node             2906 fs/btrfs/relocation.c 		if (!node->eb) {
node             2915 fs/btrfs/relocation.c 			BUG_ON(node->eb != eb);
node             2918 fs/btrfs/relocation.c 						node->eb->start);
node             2924 fs/btrfs/relocation.c 					       node->eb->start, blocksize,
node             2927 fs/btrfs/relocation.c 			btrfs_init_tree_ref(&ref, node->level,
node             2944 fs/btrfs/relocation.c 	if (!err && node->pending) {
node             2945 fs/btrfs/relocation.c 		drop_node_buffer(node);
node             2946 fs/btrfs/relocation.c 		list_move_tail(&node->list, &rc->backref_cache.changed);
node             2947 fs/btrfs/relocation.c 		node->pending = 0;
node             2957 fs/btrfs/relocation.c 			 struct backref_node *node,
node             2962 fs/btrfs/relocation.c 	btrfs_node_key_to_cpu(node->eb, &key, 0);
node             2963 fs/btrfs/relocation.c 	return do_relocation(trans, rc, node, &key, path, 0);
node             2972 fs/btrfs/relocation.c 	struct backref_node *node;
node             2978 fs/btrfs/relocation.c 			node = list_entry(cache->pending[level].next,
node             2980 fs/btrfs/relocation.c 			list_move_tail(&node->list, &list);
node             2981 fs/btrfs/relocation.c 			BUG_ON(!node->pending);
node             2984 fs/btrfs/relocation.c 				ret = link_to_upper(trans, rc, node, path);
node             3002 fs/btrfs/relocation.c 				   struct backref_node *node)
node             3005 fs/btrfs/relocation.c 	if (node->level == 0 ||
node             3006 fs/btrfs/relocation.c 	    in_block_group(node->bytenr, rc->block_group)) {
node             3008 fs/btrfs/relocation.c 		mark_block_processed(rc, node->bytenr, blocksize);
node             3010 fs/btrfs/relocation.c 	node->processed = 1;
node             3018 fs/btrfs/relocation.c 				    struct backref_node *node)
node             3020 fs/btrfs/relocation.c 	struct backref_node *next = node;
node             3039 fs/btrfs/relocation.c 			next = edge->node[UPPER];
node             3083 fs/btrfs/relocation.c 				struct backref_node *node,
node             3090 fs/btrfs/relocation.c 	if (!node)
node             3093 fs/btrfs/relocation.c 	BUG_ON(node->processed);
node             3094 fs/btrfs/relocation.c 	root = select_one_root(node);
node             3096 fs/btrfs/relocation.c 		update_processed_blocks(rc, node);
node             3101 fs/btrfs/relocation.c 		ret = reserve_metadata_space(trans, rc, node);
node             3108 fs/btrfs/relocation.c 			BUG_ON(node->new_bytenr);
node             3109 fs/btrfs/relocation.c 			BUG_ON(!list_empty(&node->list));
node             3112 fs/btrfs/relocation.c 			node->new_bytenr = root->node->start;
node             3113 fs/btrfs/relocation.c 			node->root = root;
node             3114 fs/btrfs/relocation.c 			list_add_tail(&node->list, &rc->backref_cache.changed);
node             3116 fs/btrfs/relocation.c 			path->lowest_level = node->level;
node             3123 fs/btrfs/relocation.c 			update_processed_blocks(rc, node);
node             3125 fs/btrfs/relocation.c 		ret = do_relocation(trans, rc, node, key, path, 1);
node             3128 fs/btrfs/relocation.c 	if (ret || node->level == 0 || node->cowonly)
node             3129 fs/btrfs/relocation.c 		remove_backref_node(&rc->backref_cache, node);
node             3141 fs/btrfs/relocation.c 	struct backref_node *node;
node             3171 fs/btrfs/relocation.c 		node = build_backref_tree(rc, &block->key,
node             3173 fs/btrfs/relocation.c 		if (IS_ERR(node)) {
node             3174 fs/btrfs/relocation.c 			err = PTR_ERR(node);
node             3178 fs/btrfs/relocation.c 		ret = relocate_tree_block(trans, rc, node, &block->key,
node             4712 fs/btrfs/relocation.c 	struct backref_node *node;
node             4733 fs/btrfs/relocation.c 		node = rc->backref_cache.path[level];
node             4734 fs/btrfs/relocation.c 		BUG_ON(node->bytenr != buf->start &&
node             4735 fs/btrfs/relocation.c 		       node->new_bytenr != buf->start);
node             4737 fs/btrfs/relocation.c 		drop_node_buffer(node);
node             4739 fs/btrfs/relocation.c 		node->eb = cow;
node             4740 fs/btrfs/relocation.c 		node->new_bytenr = cow->start;
node             4742 fs/btrfs/relocation.c 		if (!node->pending) {
node             4743 fs/btrfs/relocation.c 			list_move_tail(&node->list,
node             4745 fs/btrfs/relocation.c 			node->pending = 1;
node             4749 fs/btrfs/relocation.c 			__mark_block_processed(rc, node);
node              110 fs/btrfs/root-tree.c 			 struct extent_buffer *node)
node              112 fs/btrfs/root-tree.c 	btrfs_set_root_bytenr(item, node->start);
node              113 fs/btrfs/root-tree.c 	btrfs_set_root_level(item, btrfs_header_level(node));
node              114 fs/btrfs/root-tree.c 	btrfs_set_root_generation(item, btrfs_header_generation(node));
node              203 fs/btrfs/scrub.c 	struct rb_node node;
node              333 fs/btrfs/scrub.c 		entry = rb_entry(parent, struct full_stripe_lock, node);
node              354 fs/btrfs/scrub.c 	rb_link_node(&ret->node, parent, p);
node              355 fs/btrfs/scrub.c 	rb_insert_color(&ret->node, &locks_root->root);
node              369 fs/btrfs/scrub.c 	struct rb_node *node;
node              374 fs/btrfs/scrub.c 	node = locks_root->root.rb_node;
node              375 fs/btrfs/scrub.c 	while (node) {
node              376 fs/btrfs/scrub.c 		entry = rb_entry(node, struct full_stripe_lock, node);
node              378 fs/btrfs/scrub.c 			node = node->rb_left;
node              380 fs/btrfs/scrub.c 			node = node->rb_right;
node              514 fs/btrfs/scrub.c 		rb_erase(&fstripe_lock->node, &locks_root->root);
node              223 fs/btrfs/send.c 	struct rb_node node;
node              232 fs/btrfs/send.c 	struct rb_node node;
node              244 fs/btrfs/send.c 	struct rb_node node;
node             2867 fs/btrfs/send.c 		entry = rb_entry(parent, struct orphan_dir_info, node);
node             2884 fs/btrfs/send.c 	rb_link_node(&odi->node, parent, p);
node             2885 fs/btrfs/send.c 	rb_insert_color(&odi->node, &sctx->orphan_dirs);
node             2896 fs/btrfs/send.c 		entry = rb_entry(n, struct orphan_dir_info, node);
node             2919 fs/btrfs/send.c 	rb_erase(&odi->node, &sctx->orphan_dirs);
node             3042 fs/btrfs/send.c 		entry = rb_entry(parent, struct waiting_dir_move, node);
node             3053 fs/btrfs/send.c 	rb_link_node(&dm->node, parent, p);
node             3054 fs/btrfs/send.c 	rb_insert_color(&dm->node, &sctx->waiting_dir_moves);
node             3065 fs/btrfs/send.c 		entry = rb_entry(n, struct waiting_dir_move, node);
node             3081 fs/btrfs/send.c 	rb_erase(&dm->node, &sctx->waiting_dir_moves);
node             3108 fs/btrfs/send.c 	RB_CLEAR_NODE(&pm->node);
node             3112 fs/btrfs/send.c 		entry = rb_entry(parent, struct pending_dir_move, node);
node             3141 fs/btrfs/send.c 		rb_link_node(&pm->node, parent, p);
node             3142 fs/btrfs/send.c 		rb_insert_color(&pm->node, &sctx->pending_dir_moves);
node             3160 fs/btrfs/send.c 		entry = rb_entry(n, struct pending_dir_move, node);
node             3357 fs/btrfs/send.c 	if (!RB_EMPTY_NODE(&m->node))
node             3358 fs/btrfs/send.c 		rb_erase(&m->node, &sctx->pending_dir_moves);
node             3375 fs/btrfs/send.c 	if (!RB_EMPTY_NODE(&moves->node)) {
node             3376 fs/btrfs/send.c 		rb_erase(&moves->node, &sctx->pending_dir_moves);
node             3377 fs/btrfs/send.c 		RB_CLEAR_NODE(&moves->node);
node             6968 fs/btrfs/send.c 	    sctx->parent_root->node != sctx->parent_root->commit_root)
node             6972 fs/btrfs/send.c 		if (sctx->clone_roots[i].root->node !=
node             7313 fs/btrfs/send.c 		pm = rb_entry(n, struct pending_dir_move, node);
node             7330 fs/btrfs/send.c 		dm = rb_entry(n, struct waiting_dir_move, node);
node             7331 fs/btrfs/send.c 		rb_erase(&dm->node, &sctx->waiting_dir_moves);
node             7341 fs/btrfs/send.c 		odi = rb_entry(n, struct orphan_dir_info, node);
node              197 fs/btrfs/tests/btrfs-tests.c 	if (root->node) {
node              199 fs/btrfs/tests/btrfs-tests.c 		free_extent_buffer(root->node);
node               13 fs/btrfs/tests/extent-map-tests.c 	struct rb_node *node;
node               16 fs/btrfs/tests/extent-map-tests.c 		node = rb_first_cached(&em_tree->map);
node               17 fs/btrfs/tests/extent-map-tests.c 		em = rb_entry(node, struct extent_map, rb_node);
node              465 fs/btrfs/tests/free-space-tree-tests.c 	root->node = alloc_test_extent_buffer(root->fs_info, nodesize);
node              466 fs/btrfs/tests/free-space-tree-tests.c 	if (IS_ERR(root->node)) {
node              468 fs/btrfs/tests/free-space-tree-tests.c 		ret = PTR_ERR(root->node);
node              471 fs/btrfs/tests/free-space-tree-tests.c 	btrfs_set_header_level(root->node, 0);
node              472 fs/btrfs/tests/free-space-tree-tests.c 	btrfs_set_header_nritems(root->node, 0);
node              519 fs/btrfs/tests/free-space-tree-tests.c 	if (btrfs_header_nritems(root->node) != 0) {
node               21 fs/btrfs/tests/inode-tests.c 	struct extent_buffer *leaf = root->node;
node               54 fs/btrfs/tests/inode-tests.c 	struct extent_buffer *leaf = root->node;
node              254 fs/btrfs/tests/inode-tests.c 	root->node = alloc_dummy_extent_buffer(fs_info, nodesize);
node              255 fs/btrfs/tests/inode-tests.c 	if (!root->node) {
node              260 fs/btrfs/tests/inode-tests.c 	btrfs_set_header_nritems(root->node, 0);
node              261 fs/btrfs/tests/inode-tests.c 	btrfs_set_header_level(root->node, 0);
node              857 fs/btrfs/tests/inode-tests.c 	root->node = alloc_dummy_extent_buffer(fs_info, nodesize);
node              858 fs/btrfs/tests/inode-tests.c 	if (!root->node) {
node              863 fs/btrfs/tests/inode-tests.c 	btrfs_set_header_nritems(root->node, 0);
node              864 fs/btrfs/tests/inode-tests.c 	btrfs_set_header_level(root->node, 0);
node              486 fs/btrfs/tests/qgroup-tests.c 	root->node = alloc_test_extent_buffer(root->fs_info, nodesize);
node              487 fs/btrfs/tests/qgroup-tests.c 	if (IS_ERR(root->node)) {
node              489 fs/btrfs/tests/qgroup-tests.c 		ret = PTR_ERR(root->node);
node              492 fs/btrfs/tests/qgroup-tests.c 	btrfs_set_header_level(root->node, 0);
node              493 fs/btrfs/tests/qgroup-tests.c 	btrfs_set_header_nritems(root->node, 0);
node              293 fs/btrfs/transaction.c 		WARN_ON(!force && root->commit_root != root->node);
node             1097 fs/btrfs/transaction.c 		if (old_root_bytenr == root->node->start &&
node             1101 fs/btrfs/transaction.c 		btrfs_set_root_node(&root->root_item, root->node);
node             1252 fs/btrfs/transaction.c 			if (root->commit_root != root->node) {
node             1256 fs/btrfs/transaction.c 						    root->node);
node             2252 fs/btrfs/transaction.c 			    fs_info->tree_root->node);
node             2257 fs/btrfs/transaction.c 			    fs_info->chunk_root->node);
node             2383 fs/btrfs/transaction.c 	if (btrfs_header_backref_rev(root->node) <
node             1445 fs/btrfs/tree-checker.c int btrfs_check_node(struct extent_buffer *node)
node             1447 fs/btrfs/tree-checker.c 	struct btrfs_fs_info *fs_info = node->fs_info;
node             1448 fs/btrfs/tree-checker.c 	unsigned long nr = btrfs_header_nritems(node);
node             1451 fs/btrfs/tree-checker.c 	int level = btrfs_header_level(node);
node             1456 fs/btrfs/tree-checker.c 		generic_err(node, 0,
node             1464 fs/btrfs/tree-checker.c 			   btrfs_header_owner(node), node->start,
node             1471 fs/btrfs/tree-checker.c 		bytenr = btrfs_node_blockptr(node, slot);
node             1472 fs/btrfs/tree-checker.c 		btrfs_node_key_to_cpu(node, &key, slot);
node             1473 fs/btrfs/tree-checker.c 		btrfs_node_key_to_cpu(node, &next_key, slot + 1);
node             1476 fs/btrfs/tree-checker.c 			generic_err(node, slot,
node             1482 fs/btrfs/tree-checker.c 			generic_err(node, slot,
node             1490 fs/btrfs/tree-checker.c 			generic_err(node, slot,
node               24 fs/btrfs/tree-checker.h int btrfs_check_node(struct extent_buffer *node);
node               45 fs/btrfs/tree-defrag.c 	level = btrfs_header_level(root->node);
node             2807 fs/btrfs/tree-log.c 			if (path->nodes[*level] == root->node)
node             2870 fs/btrfs/tree-log.c 	level = btrfs_header_level(log->node);
node             2872 fs/btrfs/tree-log.c 	path->nodes[level] = log->node;
node             2873 fs/btrfs/tree-log.c 	extent_buffer_get(log->node);
node             3123 fs/btrfs/tree-log.c 	btrfs_set_root_node(&log->root_item, log->node);
node             3242 fs/btrfs/tree-log.c 				 log_root_tree->node->start);
node             3244 fs/btrfs/tree-log.c 				       btrfs_header_level(log_root_tree->node));
node             3317 fs/btrfs/tree-log.c 	free_extent_buffer(log->node);
node             6247 fs/btrfs/tree-log.c 							log->node->start,
node             6248 fs/btrfs/tree-log.c 							log->node->len);
node             6249 fs/btrfs/tree-log.c 			free_extent_buffer(log->node);
node             6287 fs/btrfs/tree-log.c 		free_extent_buffer(log->node);
node             6320 fs/btrfs/tree-log.c 	free_extent_buffer(log_root_tree->node);
node               63 fs/btrfs/ulist.c 	struct ulist_node *node;
node               66 fs/btrfs/ulist.c 	list_for_each_entry_safe(node, next, &ulist->nodes, list) {
node               67 fs/btrfs/ulist.c 		kfree(node);
node              135 fs/btrfs/ulist.c static void ulist_rbtree_erase(struct ulist *ulist, struct ulist_node *node)
node              137 fs/btrfs/ulist.c 	rb_erase(&node->rb_node, &ulist->root);
node              138 fs/btrfs/ulist.c 	list_del(&node->list);
node              139 fs/btrfs/ulist.c 	kfree(node);
node              195 fs/btrfs/ulist.c 	struct ulist_node *node;
node              197 fs/btrfs/ulist.c 	node = ulist_rbtree_search(ulist, val);
node              198 fs/btrfs/ulist.c 	if (node) {
node              200 fs/btrfs/ulist.c 			*old_aux = node->aux;
node              203 fs/btrfs/ulist.c 	node = kmalloc(sizeof(*node), gfp_mask);
node              204 fs/btrfs/ulist.c 	if (!node)
node              207 fs/btrfs/ulist.c 	node->val = val;
node              208 fs/btrfs/ulist.c 	node->aux = aux;
node              210 fs/btrfs/ulist.c 	ret = ulist_rbtree_insert(ulist, node);
node              212 fs/btrfs/ulist.c 	list_add_tail(&node->list, &ulist->nodes);
node              230 fs/btrfs/ulist.c 	struct ulist_node *node;
node              232 fs/btrfs/ulist.c 	node = ulist_rbtree_search(ulist, val);
node              234 fs/btrfs/ulist.c 	if (!node)
node              237 fs/btrfs/ulist.c 	if (node->aux != aux)
node              241 fs/btrfs/ulist.c 	ulist_rbtree_erase(ulist, node);
node              263 fs/btrfs/ulist.c 	struct ulist_node *node;
node              274 fs/btrfs/ulist.c 	node = list_entry(uiter->cur_list, struct ulist_node, list);
node              275 fs/btrfs/ulist.c 	return node;
node             7796 fs/btrfs/volumes.c 	struct rb_node *node;
node             7800 fs/btrfs/volumes.c 	for (node = rb_first_cached(&em_tree->map); node; node = rb_next(node)) {
node             7801 fs/btrfs/volumes.c 		em = rb_entry(node, struct extent_map, rb_node);
node             7914 fs/btrfs/volumes.c 	struct rb_node *node;
node             7917 fs/btrfs/volumes.c 	node = fs_info->swapfile_pins.rb_node;
node             7918 fs/btrfs/volumes.c 	while (node) {
node             7919 fs/btrfs/volumes.c 		sp = rb_entry(node, struct btrfs_swapfile_pin, node);
node             7921 fs/btrfs/volumes.c 			node = node->rb_left;
node             7923 fs/btrfs/volumes.c 			node = node->rb_right;
node             7928 fs/btrfs/volumes.c 	return node != NULL;
node             1829 fs/ceph/addr.c 		perm = rb_entry(*p, struct ceph_pool_perm, node);
node             1863 fs/ceph/addr.c 		perm = rb_entry(parent, struct ceph_pool_perm, node);
node             1971 fs/ceph/addr.c 	rb_link_node(&perm->node, parent, p);
node             1972 fs/ceph/addr.c 	rb_insert_color(&perm->node, &mdsc->pool_perm_tree);
node             2060 fs/ceph/addr.c 		perm = rb_entry(n, struct ceph_pool_perm, node);
node              126 fs/ceph/inode.c 		frag = rb_entry(parent, struct ceph_inode_frag, node);
node              145 fs/ceph/inode.c 	rb_link_node(&frag->node, parent, p);
node              146 fs/ceph/inode.c 	rb_insert_color(&frag->node, &ci->i_fragtree);
node              162 fs/ceph/inode.c 			rb_entry(n, struct ceph_inode_frag, node);
node              266 fs/ceph/inode.c 			rb_erase(&frag->node, &ci->i_fragtree);
node              341 fs/ceph/inode.c 		frag = rb_entry(rb_node, struct ceph_inode_frag, node);
node              371 fs/ceph/inode.c 			frag = rb_entry(rb_node, struct ceph_inode_frag, node);
node              383 fs/ceph/inode.c 				rb_erase(&frag->node, &ci->i_fragtree);
node              402 fs/ceph/inode.c 		frag = rb_entry(rb_node, struct ceph_inode_frag, node);
node              407 fs/ceph/inode.c 			rb_erase(&frag->node, &ci->i_fragtree);
node              574 fs/ceph/inode.c 		frag = rb_entry(n, struct ceph_inode_frag, node);
node             3522 fs/ceph/mds_client.c 		       rb_entry(p, struct ceph_snap_realm, node);
node              315 fs/ceph/mds_client.h 	struct rb_node node;
node              323 fs/ceph/mds_client.h 	struct rb_node node;
node              336 fs/ceph/mds_client.h 	struct rb_node node;
node               85 fs/ceph/quota.c 	struct rb_node **node, *parent = NULL;
node               88 fs/ceph/quota.c 	node = &(mdsc->quotarealms_inodes.rb_node);
node               89 fs/ceph/quota.c 	while (*node) {
node               90 fs/ceph/quota.c 		parent = *node;
node               91 fs/ceph/quota.c 		qri = container_of(*node, struct ceph_quotarealm_inode, node);
node               94 fs/ceph/quota.c 			node = &((*node)->rb_left);
node               96 fs/ceph/quota.c 			node = &((*node)->rb_right);
node              108 fs/ceph/quota.c 			rb_link_node(&qri->node, parent, node);
node              109 fs/ceph/quota.c 			rb_insert_color(&qri->node, &mdsc->quotarealms_inodes);
node              177 fs/ceph/quota.c 	struct rb_node *node;
node              185 fs/ceph/quota.c 		node = rb_first(&mdsc->quotarealms_inodes);
node              186 fs/ceph/quota.c 		qri = rb_entry(node, struct ceph_quotarealm_inode, node);
node              187 fs/ceph/quota.c 		rb_erase(node, &mdsc->quotarealms_inodes);
node               92 fs/ceph/snap.c 		r = rb_entry(parent, struct ceph_snap_realm, node);
node              101 fs/ceph/snap.c 	rb_link_node(&new->node, parent, p);
node              102 fs/ceph/snap.c 	rb_insert_color(&new->node, root);
node              147 fs/ceph/snap.c 		r = rb_entry(n, struct ceph_snap_realm, node);
node              181 fs/ceph/snap.c 	rb_erase(&realm->node, &mdsc->snap_realms);
node             1021 fs/ceph/snap.c 		exist = rb_entry(*p, struct ceph_snapid_map, node);
node             1059 fs/ceph/snap.c 		exist = rb_entry(*p, struct ceph_snapid_map, node);
node             1072 fs/ceph/snap.c 		rb_link_node(&sm->node, parent, p);
node             1073 fs/ceph/snap.c 		rb_insert_color(&sm->node, &mdsc->snapid_map_tree);
node             1093 fs/ceph/snap.c 		if (!RB_EMPTY_NODE(&sm->node)) {
node             1121 fs/ceph/snap.c 		rb_erase(&sm->node, &mdsc->snapid_map_tree);
node             1143 fs/ceph/snap.c 		sm = rb_entry(p, struct ceph_snapid_map, node);
node              238 fs/ceph/super.h 	struct rb_node node;
node              255 fs/ceph/super.h 	struct rb_node node;
node              801 fs/ceph/super.h 	struct rb_node node;
node              448 fs/ceph/xattr.c 		xattr = rb_entry(parent, struct ceph_inode_xattr, node);
node              521 fs/ceph/xattr.c 		rb_link_node(&xattr->node, parent, p);
node              522 fs/ceph/xattr.c 		rb_insert_color(&xattr->node, &ci->i_xattrs.index);
node              544 fs/ceph/xattr.c 		xattr = rb_entry(parent, struct ceph_inode_xattr, node);
node              582 fs/ceph/xattr.c 	rb_erase(&xattr->node, &ci->i_xattrs.index);
node              607 fs/ceph/xattr.c 		xattr = rb_entry(p, struct ceph_inode_xattr, node);
node              631 fs/ceph/xattr.c 		xattr = rb_entry(p, struct ceph_inode_xattr, node);
node              780 fs/ceph/xattr.c 			xattr = rb_entry(p, struct ceph_inode_xattr, node);
node             5167 fs/cifs/connect.c 	struct rb_node *node;
node             5173 fs/cifs/connect.c 	while ((node = rb_first(root))) {
node             5174 fs/cifs/connect.c 		tlink = rb_entry(node, struct tcon_link, tl_rbnode);
node             5177 fs/cifs/connect.c 		rb_erase(node, root);
node             5337 fs/cifs/connect.c 	struct rb_node *node = root->rb_node;
node             5340 fs/cifs/connect.c 	while (node) {
node             5341 fs/cifs/connect.c 		tlink = rb_entry(node, struct tcon_link, tl_rbnode);
node             5344 fs/cifs/connect.c 			node = node->rb_left;
node             5346 fs/cifs/connect.c 			node = node->rb_right;
node             5473 fs/cifs/connect.c 	struct rb_node *node;
node             5485 fs/cifs/connect.c 	node = rb_first(root);
node             5486 fs/cifs/connect.c 	while (node != NULL) {
node             5487 fs/cifs/connect.c 		tmp = node;
node             5488 fs/cifs/connect.c 		node = rb_next(tmp);
node              732 fs/cifs/misc.c 		struct dfs_info3_param *node = (*target_nodes)+i;
node              734 fs/cifs/misc.c 		node->flags = le32_to_cpu(rsp->DFSFlags);
node              744 fs/cifs/misc.c 			node->path_consumed = cifs_utf16_bytes(tmp,
node              749 fs/cifs/misc.c 			node->path_consumed = le16_to_cpu(rsp->PathConsumed);
node              751 fs/cifs/misc.c 		node->server_type = le16_to_cpu(ref->ServerType);
node              752 fs/cifs/misc.c 		node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags);
node              757 fs/cifs/misc.c 		node->path_name = cifs_strndup_from_utf16(temp, max_len,
node              759 fs/cifs/misc.c 		if (!node->path_name) {
node              767 fs/cifs/misc.c 		node->node_name = cifs_strndup_from_utf16(temp, max_len,
node              769 fs/cifs/misc.c 		if (!node->node_name) {
node              774 fs/cifs/misc.c 		node->ttl = le32_to_cpu(ref->TimeToLive);
node             2219 fs/dcache.c    	struct hlist_bl_node *node;
node             2242 fs/dcache.c    	hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) {
node             2341 fs/dcache.c    	struct hlist_bl_node *node;
node             2367 fs/dcache.c    	hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) {
node             2518 fs/dcache.c    	struct hlist_bl_node *node;
node             2568 fs/dcache.c    	hlist_bl_for_each_entry(dentry, node, b, d_u.d_in_lookup_hash) {
node              780 fs/dlm/config.c 	struct dlm_config_node *nodes, *node;
node              802 fs/dlm/config.c 	node = nodes;
node              804 fs/dlm/config.c 		node->nodeid = nd->nodeid;
node              805 fs/dlm/config.c 		node->weight = nd->weight;
node              806 fs/dlm/config.c 		node->new = nd->new;
node              807 fs/dlm/config.c 		node->comm_seq = nd->comm_seq;
node              808 fs/dlm/config.c 		node++;
node              422 fs/dlm/debug_fs.c 	struct rb_node *node;
node              454 fs/dlm/debug_fs.c 		for (node = rb_first(tree); node; node = rb_next(node)) {
node              455 fs/dlm/debug_fs.c 			r = rb_entry(node, struct dlm_rsb, res_hashnode);
node              486 fs/dlm/debug_fs.c 			node = rb_first(tree);
node              487 fs/dlm/debug_fs.c 			r = rb_entry(node, struct dlm_rsb, res_hashnode);
node               35 fs/dlm/dir.c   	uint32_t node;
node               40 fs/dlm/dir.c   		node = (hash >> 16) % ls->ls_total_weight;
node               41 fs/dlm/dir.c   		return ls->ls_node_array[node];
node              448 fs/dlm/lock.c  	struct rb_node *node = tree->rb_node;
node              452 fs/dlm/lock.c  	while (node) {
node              453 fs/dlm/lock.c  		r = rb_entry(node, struct dlm_rsb, res_hashnode);
node              456 fs/dlm/lock.c  			node = node->rb_left;
node              458 fs/dlm/lock.c  			node = node->rb_right;
node              314 fs/dlm/member.c static int dlm_add_member(struct dlm_ls *ls, struct dlm_config_node *node)
node              323 fs/dlm/member.c 	error = dlm_lowcomms_connect_node(node->nodeid);
node              329 fs/dlm/member.c 	memb->nodeid = node->nodeid;
node              330 fs/dlm/member.c 	memb->weight = node->weight;
node              331 fs/dlm/member.c 	memb->comm_seq = node->comm_seq;
node              527 fs/dlm/member.c 	struct dlm_config_node *node;
node              541 fs/dlm/member.c 		node = find_config_node(rv, memb->nodeid);
node              542 fs/dlm/member.c 		if (node && !node->new)
node              545 fs/dlm/member.c 		if (!node) {
node              550 fs/dlm/member.c 				  memb->nodeid, memb->comm_seq, node->comm_seq);
node              562 fs/dlm/member.c 		node = &rv->nodes[i];
node              563 fs/dlm/member.c 		if (dlm_is_member(ls, node->nodeid))
node              565 fs/dlm/member.c 		dlm_add_member(ls, node);
node              566 fs/dlm/member.c 		log_rinfo(ls, "add member %d", node->nodeid);
node              400 fs/ecryptfs/ecryptfs_kernel.h 	struct list_head node;
node               53 fs/ecryptfs/messaging.c 		*msg_ctx = list_entry(p, struct ecryptfs_msg_ctx, node);
node               73 fs/ecryptfs/messaging.c 	list_move(&msg_ctx->node, &ecryptfs_msg_ctx_alloc_list);
node               86 fs/ecryptfs/messaging.c 	list_move(&(msg_ctx->node), &ecryptfs_msg_ctx_free_list);
node              390 fs/ecryptfs/messaging.c 		INIT_LIST_HEAD(&ecryptfs_msg_ctx_arr[i].node);
node              399 fs/ecryptfs/messaging.c 		list_add_tail(&ecryptfs_msg_ctx_arr[i].node,
node              335 fs/ext2/balloc.c 	struct rb_node *node = &rsv->rsv_node;
node              357 fs/ext2/balloc.c 	rb_link_node(node, parent, p);
node              358 fs/ext2/balloc.c 	rb_insert_color(node, root);
node               24 fs/ext4/block_validity.c 	struct rb_node	node;
node               58 fs/ext4/block_validity.c 				&system_blks->root, node)
node               72 fs/ext4/block_validity.c 	struct rb_node **n = &system_blks->root.rb_node, *node;
node               77 fs/ext4/block_validity.c 		entry = rb_entry(parent, struct ext4_system_zone, node);
node               89 fs/ext4/block_validity.c 					     node);
node              101 fs/ext4/block_validity.c 		new_node = &new_entry->node;
node              108 fs/ext4/block_validity.c 	node = rb_prev(new_node);
node              109 fs/ext4/block_validity.c 	if (node) {
node              110 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
node              114 fs/ext4/block_validity.c 			rb_erase(node, &system_blks->root);
node              120 fs/ext4/block_validity.c 	node = rb_next(new_node);
node              121 fs/ext4/block_validity.c 	if (node) {
node              122 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
node              125 fs/ext4/block_validity.c 			rb_erase(node, &system_blks->root);
node              134 fs/ext4/block_validity.c 	struct rb_node *node;
node              139 fs/ext4/block_validity.c 	node = rb_first(&sbi->system_blks->root);
node              140 fs/ext4/block_validity.c 	while (node) {
node              141 fs/ext4/block_validity.c 		entry = rb_entry(node, struct ext4_system_zone, node);
node              145 fs/ext4/block_validity.c 		node = rb_next(node);
node              175 fs/ext4/block_validity.c 		entry = rb_entry(n, struct ext4_system_zone, node);
node              181 fs/ext4/extents_status.c 	struct rb_node *node;
node              185 fs/ext4/extents_status.c 	node = rb_first(&tree->root);
node              186 fs/ext4/extents_status.c 	while (node) {
node              188 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node              192 fs/ext4/extents_status.c 		node = rb_next(node);
node              213 fs/ext4/extents_status.c 	struct rb_node *node = root->rb_node;
node              216 fs/ext4/extents_status.c 	while (node) {
node              217 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node              219 fs/ext4/extents_status.c 			node = node->rb_left;
node              221 fs/ext4/extents_status.c 			node = node->rb_right;
node              230 fs/ext4/extents_status.c 		node = rb_next(&es->rb_node);
node              231 fs/ext4/extents_status.c 		return node ? rb_entry(node, struct extent_status, rb_node) :
node              263 fs/ext4/extents_status.c 	struct rb_node *node;
node              286 fs/ext4/extents_status.c 		while ((node = rb_next(&es1->rb_node)) != NULL) {
node              287 fs/ext4/extents_status.c 			es1 = rb_entry(node, struct extent_status, rb_node);
node              533 fs/ext4/extents_status.c 	struct rb_node *node;
node              535 fs/ext4/extents_status.c 	node = rb_prev(&es->rb_node);
node              536 fs/ext4/extents_status.c 	if (!node)
node              539 fs/ext4/extents_status.c 	es1 = rb_entry(node, struct extent_status, rb_node);
node              557 fs/ext4/extents_status.c 	struct rb_node *node;
node              559 fs/ext4/extents_status.c 	node = rb_next(&es->rb_node);
node              560 fs/ext4/extents_status.c 	if (!node)
node              563 fs/ext4/extents_status.c 	es1 = rb_entry(node, struct extent_status, rb_node);
node              568 fs/ext4/extents_status.c 		rb_erase(node, &tree->root);
node              908 fs/ext4/extents_status.c 	struct rb_node *node;
node              929 fs/ext4/extents_status.c 	node = tree->root.rb_node;
node              930 fs/ext4/extents_status.c 	while (node) {
node              931 fs/ext4/extents_status.c 		es1 = rb_entry(node, struct extent_status, rb_node);
node              933 fs/ext4/extents_status.c 			node = node->rb_left;
node              935 fs/ext4/extents_status.c 			node = node->rb_right;
node              953 fs/ext4/extents_status.c 			node = rb_next(&es1->rb_node);
node              954 fs/ext4/extents_status.c 			if (node) {
node              955 fs/ext4/extents_status.c 				es1 = rb_entry(node, struct extent_status,
node              996 fs/ext4/extents_status.c 	struct rb_node *node;
node             1011 fs/ext4/extents_status.c 			node = rb_prev(&es->rb_node);
node             1012 fs/ext4/extents_status.c 			rc->left_es = node ? rb_entry(node,
node             1119 fs/ext4/extents_status.c 	struct rb_node *node = root->rb_node;
node             1122 fs/ext4/extents_status.c 	while (node) {
node             1123 fs/ext4/extents_status.c 		pr = rb_entry(node, struct pending_reservation, rb_node);
node             1125 fs/ext4/extents_status.c 			node = node->rb_left;
node             1127 fs/ext4/extents_status.c 			node = node->rb_right;
node             1134 fs/ext4/extents_status.c 		node = rb_next(&pr->rb_node);
node             1135 fs/ext4/extents_status.c 		return node ? rb_entry(node, struct pending_reservation,
node             1164 fs/ext4/extents_status.c 	struct rb_node *node;
node             1195 fs/ext4/extents_status.c 			node = rb_prev(&es->rb_node);
node             1196 fs/ext4/extents_status.c 			if (!node)
node             1198 fs/ext4/extents_status.c 			es = rb_entry(node, struct extent_status, rb_node);
node             1204 fs/ext4/extents_status.c 				node = rb_next(&right_es->rb_node);
node             1205 fs/ext4/extents_status.c 				es = node ? rb_entry(node, struct extent_status,
node             1215 fs/ext4/extents_status.c 				node = rb_next(&es->rb_node);
node             1216 fs/ext4/extents_status.c 				if (!node)
node             1218 fs/ext4/extents_status.c 				es = rb_entry(node, struct extent_status,
node             1258 fs/ext4/extents_status.c 				node = rb_next(&pr->rb_node);
node             1261 fs/ext4/extents_status.c 				if (!node)
node             1263 fs/ext4/extents_status.c 				pr = rb_entry(node, struct pending_reservation,
node             1289 fs/ext4/extents_status.c 	struct rb_node *node;
node             1364 fs/ext4/extents_status.c 		node = rb_next(&es->rb_node);
node             1365 fs/ext4/extents_status.c 		if (node)
node             1366 fs/ext4/extents_status.c 			es = rb_entry(node, struct extent_status, rb_node);
node             1374 fs/ext4/extents_status.c 		node = rb_next(&es->rb_node);
node             1377 fs/ext4/extents_status.c 		if (!node) {
node             1381 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node             1677 fs/ext4/extents_status.c 	struct rb_node *node;
node             1690 fs/ext4/extents_status.c 		node = rb_next(&es->rb_node);
node             1706 fs/ext4/extents_status.c 		if (!node)
node             1708 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node             1750 fs/ext4/extents_status.c 	struct rb_node *node;
node             1755 fs/ext4/extents_status.c 	node = rb_first(&tree->root);
node             1756 fs/ext4/extents_status.c 	while (node) {
node             1757 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node             1758 fs/ext4/extents_status.c 		node = rb_next(node);
node             1772 fs/ext4/extents_status.c 	struct rb_node *node;
node             1777 fs/ext4/extents_status.c 	node = rb_first(&tree->root);
node             1778 fs/ext4/extents_status.c 	while (node) {
node             1779 fs/ext4/extents_status.c 		pr = rb_entry(node, struct pending_reservation, rb_node);
node             1781 fs/ext4/extents_status.c 		node = rb_next(node);
node             1822 fs/ext4/extents_status.c 	struct rb_node *node;
node             1826 fs/ext4/extents_status.c 	node = (&tree->root)->rb_node;
node             1828 fs/ext4/extents_status.c 	while (node) {
node             1829 fs/ext4/extents_status.c 		pr = rb_entry(node, struct pending_reservation, rb_node);
node             1831 fs/ext4/extents_status.c 			node = node->rb_left;
node             1833 fs/ext4/extents_status.c 			node = node->rb_right;
node             2026 fs/ext4/extents_status.c 	struct rb_node *node;
node             2054 fs/ext4/extents_status.c 		node = rb_next(&es->rb_node);
node             2055 fs/ext4/extents_status.c 		if (!node)
node             2057 fs/ext4/extents_status.c 		es = rb_entry(node, struct extent_status, rb_node);
node             4656 fs/ext4/mballoc.c 	struct rb_node **n = &db->bb_free_root.rb_node, *node;
node             4695 fs/ext4/mballoc.c 	node = rb_prev(new_node);
node             4696 fs/ext4/mballoc.c 	if (node) {
node             4697 fs/ext4/mballoc.c 		entry = rb_entry(node, struct ext4_free_data, efd_node);
node             4702 fs/ext4/mballoc.c 	node = rb_next(new_node);
node             4703 fs/ext4/mballoc.c 	if (node) {
node             4704 fs/ext4/mballoc.c 		entry = rb_entry(node, struct ext4_free_data, efd_node);
node               33 fs/f2fs/extent_cache.c 	struct rb_node *node = root->rb_root.rb_node;
node               36 fs/f2fs/extent_cache.c 	while (node) {
node               37 fs/f2fs/extent_cache.c 		re = rb_entry(node, struct rb_entry, rb_node);
node               40 fs/f2fs/extent_cache.c 			node = node->rb_left;
node               42 fs/f2fs/extent_cache.c 			node = node->rb_right;
node              302 fs/f2fs/extent_cache.c 	struct rb_node *node, *next;
node              306 fs/f2fs/extent_cache.c 	node = rb_first_cached(&et->root);
node              307 fs/f2fs/extent_cache.c 	while (node) {
node              308 fs/f2fs/extent_cache.c 		next = rb_next(node);
node              309 fs/f2fs/extent_cache.c 		en = rb_entry(node, struct extent_node, rb_node);
node              311 fs/f2fs/extent_cache.c 		node = next;
node              576 fs/f2fs/extent_cache.c 			struct rb_node *node = rb_next(&en->rb_node);
node              578 fs/f2fs/extent_cache.c 			next_en = rb_entry_safe(node, struct extent_node,
node             2280 fs/f2fs/f2fs.h static inline __le32 *blkaddr_in_node(struct f2fs_node *node)
node             2282 fs/f2fs/f2fs.h 	return RAW_IS_INODE(node) ? node->i.i_addr : node->dn.addr;
node              141 fs/f2fs/inode.c 	struct f2fs_node *node = F2FS_NODE(page);
node              142 fs/f2fs/inode.c 	struct f2fs_inode *ri = &node->i;
node              143 fs/f2fs/inode.c 	__le32 ino = node->footer.ino;
node             1322 fs/f2fs/segment.c 		struct rb_node *node;
node             1378 fs/f2fs/segment.c 		node = rb_next(&prev_dc->rb_node);
node             1379 fs/f2fs/segment.c 		next_dc = rb_entry_safe(node, struct discard_cmd, rb_node);
node             1428 fs/f2fs/segment.c 		struct rb_node *node;
node             1445 fs/f2fs/segment.c 		node = rb_next(&dc->rb_node);
node             1448 fs/f2fs/segment.c 		dc = rb_entry_safe(node, struct discard_cmd, rb_node);
node             2774 fs/f2fs/segment.c 		struct rb_node *node;
node             2800 fs/f2fs/segment.c 		node = rb_next(&dc->rb_node);
node             2803 fs/f2fs/segment.c 		dc = rb_entry_safe(node, struct discard_cmd, rb_node);
node              304 fs/file_table.c 	struct llist_node *node = llist_del_all(&delayed_fput_list);
node              307 fs/file_table.c 	llist_for_each_entry_safe(f, t, node, f_u.fu_llist)
node             1234 fs/fs-writeback.c 	struct list_head *pos, *node;
node             1271 fs/fs-writeback.c 		list_for_each_prev_safe(pos, node, &tmp) {
node               18 fs/hfs/bnode.c void hfs_bnode_read(struct hfs_bnode *node, void *buf,
node               23 fs/hfs/bnode.c 	off += node->page_offset;
node               24 fs/hfs/bnode.c 	page = node->page[0];
node               30 fs/hfs/bnode.c u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off)
node               34 fs/hfs/bnode.c 	hfs_bnode_read(node, &data, off, 2);
node               38 fs/hfs/bnode.c u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off)
node               42 fs/hfs/bnode.c 	hfs_bnode_read(node, &data, off, 1);
node               46 fs/hfs/bnode.c void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off)
node               51 fs/hfs/bnode.c 	tree = node->tree;
node               52 fs/hfs/bnode.c 	if (node->type == HFS_NODE_LEAF ||
node               54 fs/hfs/bnode.c 		key_len = hfs_bnode_read_u8(node, off) + 1;
node               58 fs/hfs/bnode.c 	hfs_bnode_read(node, key, off, key_len);
node               61 fs/hfs/bnode.c void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len)
node               65 fs/hfs/bnode.c 	off += node->page_offset;
node               66 fs/hfs/bnode.c 	page = node->page[0];
node               73 fs/hfs/bnode.c void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data)
node               77 fs/hfs/bnode.c 	hfs_bnode_write(node, &v, off, 2);
node               80 fs/hfs/bnode.c void hfs_bnode_write_u8(struct hfs_bnode *node, int off, u8 data)
node               83 fs/hfs/bnode.c 	hfs_bnode_write(node, &data, off, 1);
node               86 fs/hfs/bnode.c void hfs_bnode_clear(struct hfs_bnode *node, int off, int len)
node               90 fs/hfs/bnode.c 	off += node->page_offset;
node               91 fs/hfs/bnode.c 	page = node->page[0];
node              117 fs/hfs/bnode.c void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
node              125 fs/hfs/bnode.c 	src += node->page_offset;
node              126 fs/hfs/bnode.c 	dst += node->page_offset;
node              127 fs/hfs/bnode.c 	page = node->page[0];
node              134 fs/hfs/bnode.c void hfs_bnode_dump(struct hfs_bnode *node)
node              140 fs/hfs/bnode.c 	hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this);
node              141 fs/hfs/bnode.c 	hfs_bnode_read(node, &desc, 0, sizeof(desc));
node              146 fs/hfs/bnode.c 	off = node->tree->node_size - 2;
node              148 fs/hfs/bnode.c 		key_off = hfs_bnode_read_u16(node, off);
node              150 fs/hfs/bnode.c 		if (i && node->type == HFS_NODE_INDEX) {
node              153 fs/hfs/bnode.c 			if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
node              154 fs/hfs/bnode.c 				tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1;
node              156 fs/hfs/bnode.c 				tmp = node->tree->max_key_len + 1;
node              158 fs/hfs/bnode.c 				     tmp, hfs_bnode_read_u8(node, key_off));
node              159 fs/hfs/bnode.c 			hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node              161 fs/hfs/bnode.c 		} else if (i && node->type == HFS_NODE_LEAF) {
node              164 fs/hfs/bnode.c 			tmp = hfs_bnode_read_u8(node, key_off);
node              171 fs/hfs/bnode.c void hfs_bnode_unlink(struct hfs_bnode *node)
node              177 fs/hfs/bnode.c 	tree = node->tree;
node              178 fs/hfs/bnode.c 	if (node->prev) {
node              179 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
node              182 fs/hfs/bnode.c 		tmp->next = node->next;
node              186 fs/hfs/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              187 fs/hfs/bnode.c 		tree->leaf_head = node->next;
node              189 fs/hfs/bnode.c 	if (node->next) {
node              190 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
node              193 fs/hfs/bnode.c 		tmp->prev = node->prev;
node              197 fs/hfs/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              198 fs/hfs/bnode.c 		tree->leaf_tail = node->prev;
node              201 fs/hfs/bnode.c 	if (!node->prev && !node->next) {
node              204 fs/hfs/bnode.c 	if (!node->parent) {
node              208 fs/hfs/bnode.c 	set_bit(HFS_BNODE_DELETED, &node->flags);
node              220 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              227 fs/hfs/bnode.c 	for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node              228 fs/hfs/bnode.c 	     node; node = node->next_hash) {
node              229 fs/hfs/bnode.c 		if (node->this == cnid) {
node              230 fs/hfs/bnode.c 			return node;
node              238 fs/hfs/bnode.c 	struct hfs_bnode *node, *node2;
node              251 fs/hfs/bnode.c 	node = kzalloc(size, GFP_KERNEL);
node              252 fs/hfs/bnode.c 	if (!node)
node              254 fs/hfs/bnode.c 	node->tree = tree;
node              255 fs/hfs/bnode.c 	node->this = cnid;
node              256 fs/hfs/bnode.c 	set_bit(HFS_BNODE_NEW, &node->flags);
node              257 fs/hfs/bnode.c 	atomic_set(&node->refcnt, 1);
node              259 fs/hfs/bnode.c 		node->tree->cnid, node->this);
node              260 fs/hfs/bnode.c 	init_waitqueue_head(&node->lock_wq);
node              265 fs/hfs/bnode.c 		node->next_hash = tree->node_hash[hash];
node              266 fs/hfs/bnode.c 		tree->node_hash[hash] = node;
node              270 fs/hfs/bnode.c 		kfree(node);
node              279 fs/hfs/bnode.c 	node->page_offset = off & ~PAGE_MASK;
node              288 fs/hfs/bnode.c 		node->page[i] = page;
node              291 fs/hfs/bnode.c 	return node;
node              293 fs/hfs/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              294 fs/hfs/bnode.c 	return node;
node              297 fs/hfs/bnode.c void hfs_bnode_unhash(struct hfs_bnode *node)
node              302 fs/hfs/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              303 fs/hfs/bnode.c 	for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node              304 fs/hfs/bnode.c 	     *p && *p != node; p = &(*p)->next_hash)
node              307 fs/hfs/bnode.c 	*p = node->next_hash;
node              308 fs/hfs/bnode.c 	node->tree->node_hash_cnt--;
node              314 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              320 fs/hfs/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              321 fs/hfs/bnode.c 	if (node) {
node              322 fs/hfs/bnode.c 		hfs_bnode_get(node);
node              324 fs/hfs/bnode.c 		wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags));
node              325 fs/hfs/bnode.c 		if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              327 fs/hfs/bnode.c 		return node;
node              330 fs/hfs/bnode.c 	node = __hfs_bnode_create(tree, num);
node              331 fs/hfs/bnode.c 	if (!node)
node              333 fs/hfs/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              335 fs/hfs/bnode.c 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
node              336 fs/hfs/bnode.c 		return node;
node              338 fs/hfs/bnode.c 	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset);
node              339 fs/hfs/bnode.c 	node->prev = be32_to_cpu(desc->prev);
node              340 fs/hfs/bnode.c 	node->next = be32_to_cpu(desc->next);
node              341 fs/hfs/bnode.c 	node->num_recs = be16_to_cpu(desc->num_recs);
node              342 fs/hfs/bnode.c 	node->type = desc->type;
node              343 fs/hfs/bnode.c 	node->height = desc->height;
node              344 fs/hfs/bnode.c 	kunmap(node->page[0]);
node              346 fs/hfs/bnode.c 	switch (node->type) {
node              349 fs/hfs/bnode.c 		if (node->height != 0)
node              353 fs/hfs/bnode.c 		if (node->height != 1)
node              357 fs/hfs/bnode.c 		if (node->height <= 1 || node->height > tree->depth)
node              365 fs/hfs/bnode.c 	off = hfs_bnode_read_u16(node, rec_off);
node              368 fs/hfs/bnode.c 	for (i = 1; i <= node->num_recs; off = next_off, i++) {
node              370 fs/hfs/bnode.c 		next_off = hfs_bnode_read_u16(node, rec_off);
node              376 fs/hfs/bnode.c 		if (node->type != HFS_NODE_INDEX &&
node              377 fs/hfs/bnode.c 		    node->type != HFS_NODE_LEAF)
node              379 fs/hfs/bnode.c 		key_size = hfs_bnode_read_u8(node, off) + 1;
node              383 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              384 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              385 fs/hfs/bnode.c 	return node;
node              388 fs/hfs/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              389 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              390 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              391 fs/hfs/bnode.c 	hfs_bnode_put(node);
node              395 fs/hfs/bnode.c void hfs_bnode_free(struct hfs_bnode *node)
node              399 fs/hfs/bnode.c 	for (i = 0; i < node->tree->pages_per_bnode; i++)
node              400 fs/hfs/bnode.c 		if (node->page[i])
node              401 fs/hfs/bnode.c 			put_page(node->page[i]);
node              402 fs/hfs/bnode.c 	kfree(node);
node              407 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              412 fs/hfs/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              414 fs/hfs/bnode.c 	if (node) {
node              417 fs/hfs/bnode.c 		return node;
node              419 fs/hfs/bnode.c 	node = __hfs_bnode_create(tree, num);
node              420 fs/hfs/bnode.c 	if (!node)
node              422 fs/hfs/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node              423 fs/hfs/bnode.c 		hfs_bnode_put(node);
node              427 fs/hfs/bnode.c 	pagep = node->page;
node              428 fs/hfs/bnode.c 	memset(kmap(*pagep) + node->page_offset, 0,
node              437 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              438 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              440 fs/hfs/bnode.c 	return node;
node              443 fs/hfs/bnode.c void hfs_bnode_get(struct hfs_bnode *node)
node              445 fs/hfs/bnode.c 	if (node) {
node              446 fs/hfs/bnode.c 		atomic_inc(&node->refcnt);
node              448 fs/hfs/bnode.c 			node->tree->cnid, node->this,
node              449 fs/hfs/bnode.c 			atomic_read(&node->refcnt));
node              454 fs/hfs/bnode.c void hfs_bnode_put(struct hfs_bnode *node)
node              456 fs/hfs/bnode.c 	if (node) {
node              457 fs/hfs/bnode.c 		struct hfs_btree *tree = node->tree;
node              461 fs/hfs/bnode.c 			node->tree->cnid, node->this,
node              462 fs/hfs/bnode.c 			atomic_read(&node->refcnt));
node              463 fs/hfs/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
node              464 fs/hfs/bnode.c 		if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node              467 fs/hfs/bnode.c 			if (!node->page[i])
node              469 fs/hfs/bnode.c 			mark_page_accessed(node->page[i]);
node              472 fs/hfs/bnode.c 		if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node              473 fs/hfs/bnode.c 			hfs_bnode_unhash(node);
node              475 fs/hfs/bnode.c 			hfs_bmap_free(node);
node              476 fs/hfs/bnode.c 			hfs_bnode_free(node);
node               19 fs/hfs/brec.c  u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off)
node               24 fs/hfs/brec.c  	dataoff = node->tree->node_size - (rec + 2) * 2;
node               25 fs/hfs/brec.c  	hfs_bnode_read(node, retval, dataoff, 4);
node               31 fs/hfs/brec.c  u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec)
node               35 fs/hfs/brec.c  	if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node               38 fs/hfs/brec.c  	if ((node->type == HFS_NODE_INDEX) &&
node               39 fs/hfs/brec.c  	   !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
node               40 fs/hfs/brec.c  		if (node->tree->attributes & HFS_TREE_BIGKEYS)
node               41 fs/hfs/brec.c  			retval = node->tree->max_key_len + 2;
node               43 fs/hfs/brec.c  			retval = node->tree->max_key_len + 1;
node               45 fs/hfs/brec.c  		recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
node               48 fs/hfs/brec.c  		if (node->tree->attributes & HFS_TREE_BIGKEYS) {
node               49 fs/hfs/brec.c  			retval = hfs_bnode_read_u16(node, recoff) + 2;
node               50 fs/hfs/brec.c  			if (retval > node->tree->max_key_len + 2) {
node               55 fs/hfs/brec.c  			retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1;
node               56 fs/hfs/brec.c  			if (retval > node->tree->max_key_len + 1) {
node               68 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node;
node               78 fs/hfs/brec.c  		node = hfs_bnode_find(tree, tree->leaf_head);
node               79 fs/hfs/brec.c  		if (IS_ERR(node))
node               80 fs/hfs/brec.c  			return PTR_ERR(node);
node               81 fs/hfs/brec.c  		fd->bnode = node;
node               91 fs/hfs/brec.c  	node = fd->bnode;
node               92 fs/hfs/brec.c  	hfs_bnode_dump(node);
node               94 fs/hfs/brec.c  	end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node               95 fs/hfs/brec.c  	end_off = hfs_bnode_read_u16(node, end_rec_off);
node              107 fs/hfs/brec.c  	if (node->type == HFS_NODE_LEAF) {
node              111 fs/hfs/brec.c  	node->num_recs++;
node              113 fs/hfs/brec.c  	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              114 fs/hfs/brec.c  	hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node              122 fs/hfs/brec.c  		data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node              123 fs/hfs/brec.c  		hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node              128 fs/hfs/brec.c  	hfs_bnode_move(node, data_off + size, data_off,
node              132 fs/hfs/brec.c  	hfs_bnode_write(node, fd->search_key, data_off, key_len);
node              133 fs/hfs/brec.c  	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node              134 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              140 fs/hfs/brec.c  	if (!rec && new_node != node) {
node              141 fs/hfs/brec.c  		hfs_bnode_read_key(node, fd->search_key, data_off + size);
node              180 fs/hfs/brec.c  	struct hfs_bnode *node, *parent;
node              184 fs/hfs/brec.c  	node = fd->bnode;
node              187 fs/hfs/brec.c  	end_off = tree->node_size - (node->num_recs + 1) * 2;
node              189 fs/hfs/brec.c  	if (node->type == HFS_NODE_LEAF) {
node              193 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              196 fs/hfs/brec.c  	if (!--node->num_recs) {
node              197 fs/hfs/brec.c  		hfs_bnode_unlink(node);
node              198 fs/hfs/brec.c  		if (!node->parent)
node              200 fs/hfs/brec.c  		parent = hfs_bnode_find(tree, node->parent);
node              203 fs/hfs/brec.c  		hfs_bnode_put(node);
node              204 fs/hfs/brec.c  		node = fd->bnode = parent;
node              206 fs/hfs/brec.c  		__hfs_brec_find(node, fd);
node              209 fs/hfs/brec.c  	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              216 fs/hfs/brec.c  		data_off = hfs_bnode_read_u16(node, rec_off);
node              217 fs/hfs/brec.c  		hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node              222 fs/hfs/brec.c  	hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node              225 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              234 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node, *next_node;
node              240 fs/hfs/brec.c  	node = fd->bnode;
node              244 fs/hfs/brec.c  	hfs_bnode_get(node);
node              246 fs/hfs/brec.c  		node->this, new_node->this, node->next);
node              247 fs/hfs/brec.c  	new_node->next = node->next;
node              248 fs/hfs/brec.c  	new_node->prev = node->this;
node              249 fs/hfs/brec.c  	new_node->parent = node->parent;
node              250 fs/hfs/brec.c  	new_node->type = node->type;
node              251 fs/hfs/brec.c  	new_node->height = node->height;
node              253 fs/hfs/brec.c  	if (node->next)
node              254 fs/hfs/brec.c  		next_node = hfs_bnode_find(tree, node->next);
node              259 fs/hfs/brec.c  		hfs_bnode_put(node);
node              264 fs/hfs/brec.c  	size = tree->node_size / 2 - node->num_recs * 2 - 14;
node              268 fs/hfs/brec.c  		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              272 fs/hfs/brec.c  		if (++num_recs < node->num_recs)
node              275 fs/hfs/brec.c  		hfs_bnode_put(node);
node              288 fs/hfs/brec.c  		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              290 fs/hfs/brec.c  		hfs_bnode_put(node);
node              297 fs/hfs/brec.c  	new_node->num_recs = node->num_recs - num_recs;
node              298 fs/hfs/brec.c  	node->num_recs = num_recs;
node              309 fs/hfs/brec.c  		data_end = hfs_bnode_read_u16(node, old_rec_off);
node              314 fs/hfs/brec.c  	hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node              326 fs/hfs/brec.c  	node->next = new_node->this;
node              327 fs/hfs/brec.c  	hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node              328 fs/hfs/brec.c  	node_desc.next = cpu_to_be32(node->next);
node              329 fs/hfs/brec.c  	node_desc.num_recs = cpu_to_be16(node->num_recs);
node              330 fs/hfs/brec.c  	hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node              339 fs/hfs/brec.c  	} else if (node->this == tree->leaf_tail) {
node              345 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              347 fs/hfs/brec.c  	hfs_bnode_put(node);
node              355 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node, *parent;
node              361 fs/hfs/brec.c  	node = fd->bnode;
node              363 fs/hfs/brec.c  	if (!node->parent)
node              367 fs/hfs/brec.c  	parent = hfs_bnode_find(tree, node->parent);
node              378 fs/hfs/brec.c  		newkeylen = (hfs_bnode_read_u8(node, 14) | 1) + 1;
node              417 fs/hfs/brec.c  	hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node              422 fs/hfs/brec.c  	hfs_bnode_put(node);
node              423 fs/hfs/brec.c  	node = parent;
node              443 fs/hfs/brec.c  			if (new_node == node)
node              446 fs/hfs/brec.c  			hfs_bnode_read_key(node, fd->search_key, 14);
node              451 fs/hfs/brec.c  	if (!rec && node->parent)
node              454 fs/hfs/brec.c  	fd->bnode = node;
node              460 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node;
node              465 fs/hfs/brec.c  	node = NULL;
node              467 fs/hfs/brec.c  		node = hfs_bnode_find(tree, tree->root);
node              468 fs/hfs/brec.c  		if (IS_ERR(node))
node              469 fs/hfs/brec.c  			return PTR_ERR(node);
node              473 fs/hfs/brec.c  		hfs_bnode_put(node);
node              502 fs/hfs/brec.c  	if (node) {
node              504 fs/hfs/brec.c  		node->parent = tree->root;
node              505 fs/hfs/brec.c  		if (node->type == HFS_NODE_LEAF ||
node              507 fs/hfs/brec.c  			key_size = hfs_bnode_read_u8(node, 14) + 1;
node              510 fs/hfs/brec.c  		hfs_bnode_copy(new_node, 14, node, 14, key_size);
node              517 fs/hfs/brec.c  		cnid = cpu_to_be32(node->this);
node              523 fs/hfs/brec.c  		hfs_bnode_put(node);
node              139 fs/hfs/btree.c 	struct hfs_bnode *node;
node              146 fs/hfs/btree.c 		while ((node = tree->node_hash[i])) {
node              147 fs/hfs/btree.c 			tree->node_hash[i] = node->next_hash;
node              148 fs/hfs/btree.c 			if (atomic_read(&node->refcnt))
node              150 fs/hfs/btree.c 				       node->tree->cnid, node->this,
node              151 fs/hfs/btree.c 				       atomic_read(&node->refcnt));
node              152 fs/hfs/btree.c 			hfs_bnode_free(node);
node              163 fs/hfs/btree.c 	struct hfs_bnode *node;
node              166 fs/hfs/btree.c 	node = hfs_bnode_find(tree, 0);
node              167 fs/hfs/btree.c 	if (IS_ERR(node))
node              171 fs/hfs/btree.c 	page = node->page[0];
node              185 fs/hfs/btree.c 	hfs_bnode_put(node);
node              191 fs/hfs/btree.c 	struct hfs_bnode *node;
node              195 fs/hfs/btree.c 	node = hfs_bnode_create(tree, idx);
node              196 fs/hfs/btree.c 	if (IS_ERR(node))
node              197 fs/hfs/btree.c 		return node;
node              206 fs/hfs/btree.c 	node->type = HFS_NODE_MAP;
node              207 fs/hfs/btree.c 	node->num_recs = 1;
node              208 fs/hfs/btree.c 	hfs_bnode_clear(node, 0, tree->node_size);
node              215 fs/hfs/btree.c 	hfs_bnode_write(node, &desc, 0, sizeof(desc));
node              216 fs/hfs/btree.c 	hfs_bnode_write_u16(node, 14, 0x8000);
node              217 fs/hfs/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node              218 fs/hfs/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node              220 fs/hfs/btree.c 	return node;
node              249 fs/hfs/btree.c 	struct hfs_bnode *node, *next_node;
node              263 fs/hfs/btree.c 	node = hfs_bnode_find(tree, nidx);
node              264 fs/hfs/btree.c 	if (IS_ERR(node))
node              265 fs/hfs/btree.c 		return node;
node              266 fs/hfs/btree.c 	len = hfs_brec_lenoff(node, 2, &off16);
node              269 fs/hfs/btree.c 	off += node->page_offset;
node              270 fs/hfs/btree.c 	pagep = node->page + (off >> PAGE_SHIFT);
node              287 fs/hfs/btree.c 						hfs_bnode_put(node);
node              301 fs/hfs/btree.c 		nidx = node->next;
node              304 fs/hfs/btree.c 			next_node = hfs_bmap_new_bmap(node, idx);
node              307 fs/hfs/btree.c 		hfs_bnode_put(node);
node              310 fs/hfs/btree.c 		node = next_node;
node              312 fs/hfs/btree.c 		len = hfs_brec_lenoff(node, 0, &off16);
node              314 fs/hfs/btree.c 		off += node->page_offset;
node              315 fs/hfs/btree.c 		pagep = node->page + (off >> PAGE_SHIFT);
node              321 fs/hfs/btree.c void hfs_bmap_free(struct hfs_bnode *node)
node              329 fs/hfs/btree.c 	hfs_dbg(BNODE_MOD, "btree_free_node: %u\n", node->this);
node              330 fs/hfs/btree.c 	tree = node->tree;
node              331 fs/hfs/btree.c 	nidx = node->this;
node              332 fs/hfs/btree.c 	node = hfs_bnode_find(tree, 0);
node              333 fs/hfs/btree.c 	if (IS_ERR(node))
node              335 fs/hfs/btree.c 	len = hfs_brec_lenoff(node, 2, &off);
node              340 fs/hfs/btree.c 		i = node->next;
node              344 fs/hfs/btree.c 				node->this);
node              345 fs/hfs/btree.c 			hfs_bnode_put(node);
node              348 fs/hfs/btree.c 		hfs_bnode_put(node);
node              349 fs/hfs/btree.c 		node = hfs_bnode_find(tree, i);
node              350 fs/hfs/btree.c 		if (IS_ERR(node))
node              352 fs/hfs/btree.c 		if (node->type != HFS_NODE_MAP) {
node              355 fs/hfs/btree.c 				node->this, node->type);
node              356 fs/hfs/btree.c 			hfs_bnode_put(node);
node              359 fs/hfs/btree.c 		len = hfs_brec_lenoff(node, 0, &off);
node              361 fs/hfs/btree.c 	off += node->page_offset + nidx / 8;
node              362 fs/hfs/btree.c 	page = node->page[off >> PAGE_SHIFT];
node              369 fs/hfs/btree.c 			node->this, node->type);
node              371 fs/hfs/btree.c 		hfs_bnode_put(node);
node              377 fs/hfs/btree.c 	hfs_bnode_put(node);
node               87 fs/hfs/btree.h extern void hfs_bmap_free(struct hfs_bnode *node);
node               77 fs/hfs/inode.c 	struct hfs_bnode *node;
node               99 fs/hfs/inode.c 		node = hfs_bnode_findhash(tree, nidx);
node              100 fs/hfs/inode.c 		if (!node)
node              102 fs/hfs/inode.c 		else if (atomic_read(&node->refcnt))
node              104 fs/hfs/inode.c 		if (res && node) {
node              105 fs/hfs/inode.c 			hfs_bnode_unhash(node);
node              106 fs/hfs/inode.c 			hfs_bnode_free(node);
node              114 fs/hfs/inode.c 			node = hfs_bnode_findhash(tree, nidx++);
node              115 fs/hfs/inode.c 			if (!node)
node              117 fs/hfs/inode.c 			if (atomic_read(&node->refcnt)) {
node              121 fs/hfs/inode.c 			hfs_bnode_unhash(node);
node              122 fs/hfs/inode.c 			hfs_bnode_free(node);
node               22 fs/hfsplus/bnode.c void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len)
node               27 fs/hfsplus/bnode.c 	off += node->page_offset;
node               28 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_SHIFT);
node               43 fs/hfsplus/bnode.c u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off)
node               47 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &data, off, 2);
node               51 fs/hfsplus/bnode.c u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off)
node               55 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &data, off, 1);
node               59 fs/hfsplus/bnode.c void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off)
node               64 fs/hfsplus/bnode.c 	tree = node->tree;
node               65 fs/hfsplus/bnode.c 	if (node->type == HFS_NODE_LEAF ||
node               67 fs/hfsplus/bnode.c 	    node->tree->cnid == HFSPLUS_ATTR_CNID)
node               68 fs/hfsplus/bnode.c 		key_len = hfs_bnode_read_u16(node, off) + 2;
node               72 fs/hfsplus/bnode.c 	hfs_bnode_read(node, key, off, key_len);
node               75 fs/hfsplus/bnode.c void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len)
node               80 fs/hfsplus/bnode.c 	off += node->page_offset;
node               81 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_SHIFT);
node               98 fs/hfsplus/bnode.c void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data)
node              102 fs/hfsplus/bnode.c 	hfs_bnode_write(node, &v, off, 2);
node              105 fs/hfsplus/bnode.c void hfs_bnode_clear(struct hfs_bnode *node, int off, int len)
node              110 fs/hfsplus/bnode.c 	off += node->page_offset;
node              111 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_SHIFT);
node              185 fs/hfsplus/bnode.c void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
node              193 fs/hfsplus/bnode.c 	src += node->page_offset;
node              194 fs/hfsplus/bnode.c 	dst += node->page_offset;
node              197 fs/hfsplus/bnode.c 		src_page = node->page + (src >> PAGE_SHIFT);
node              200 fs/hfsplus/bnode.c 		dst_page = node->page + (dst >> PAGE_SHIFT);
node              247 fs/hfsplus/bnode.c 		src_page = node->page + (src >> PAGE_SHIFT);
node              249 fs/hfsplus/bnode.c 		dst_page = node->page + (dst >> PAGE_SHIFT);
node              298 fs/hfsplus/bnode.c void hfs_bnode_dump(struct hfs_bnode *node)
node              304 fs/hfsplus/bnode.c 	hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this);
node              305 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &desc, 0, sizeof(desc));
node              310 fs/hfsplus/bnode.c 	off = node->tree->node_size - 2;
node              312 fs/hfsplus/bnode.c 		key_off = hfs_bnode_read_u16(node, off);
node              314 fs/hfsplus/bnode.c 		if (i && node->type == HFS_NODE_INDEX) {
node              317 fs/hfsplus/bnode.c 			if (node->tree->attributes & HFS_TREE_VARIDXKEYS ||
node              318 fs/hfsplus/bnode.c 					node->tree->cnid == HFSPLUS_ATTR_CNID)
node              319 fs/hfsplus/bnode.c 				tmp = hfs_bnode_read_u16(node, key_off) + 2;
node              321 fs/hfsplus/bnode.c 				tmp = node->tree->max_key_len + 2;
node              323 fs/hfsplus/bnode.c 			hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node              325 fs/hfsplus/bnode.c 		} else if (i && node->type == HFS_NODE_LEAF) {
node              328 fs/hfsplus/bnode.c 			tmp = hfs_bnode_read_u16(node, key_off);
node              335 fs/hfsplus/bnode.c void hfs_bnode_unlink(struct hfs_bnode *node)
node              341 fs/hfsplus/bnode.c 	tree = node->tree;
node              342 fs/hfsplus/bnode.c 	if (node->prev) {
node              343 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
node              346 fs/hfsplus/bnode.c 		tmp->next = node->next;
node              351 fs/hfsplus/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              352 fs/hfsplus/bnode.c 		tree->leaf_head = node->next;
node              354 fs/hfsplus/bnode.c 	if (node->next) {
node              355 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
node              358 fs/hfsplus/bnode.c 		tmp->prev = node->prev;
node              363 fs/hfsplus/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              364 fs/hfsplus/bnode.c 		tree->leaf_tail = node->prev;
node              367 fs/hfsplus/bnode.c 	if (!node->prev && !node->next)
node              369 fs/hfsplus/bnode.c 	if (!node->parent) {
node              373 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_DELETED, &node->flags);
node              385 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              393 fs/hfsplus/bnode.c 	for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node              394 fs/hfsplus/bnode.c 			node; node = node->next_hash)
node              395 fs/hfsplus/bnode.c 		if (node->this == cnid)
node              396 fs/hfsplus/bnode.c 			return node;
node              402 fs/hfsplus/bnode.c 	struct hfs_bnode *node, *node2;
node              416 fs/hfsplus/bnode.c 	node = kzalloc(size, GFP_KERNEL);
node              417 fs/hfsplus/bnode.c 	if (!node)
node              419 fs/hfsplus/bnode.c 	node->tree = tree;
node              420 fs/hfsplus/bnode.c 	node->this = cnid;
node              421 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_NEW, &node->flags);
node              422 fs/hfsplus/bnode.c 	atomic_set(&node->refcnt, 1);
node              424 fs/hfsplus/bnode.c 		node->tree->cnid, node->this);
node              425 fs/hfsplus/bnode.c 	init_waitqueue_head(&node->lock_wq);
node              430 fs/hfsplus/bnode.c 		node->next_hash = tree->node_hash[hash];
node              431 fs/hfsplus/bnode.c 		tree->node_hash[hash] = node;
node              435 fs/hfsplus/bnode.c 		kfree(node);
node              445 fs/hfsplus/bnode.c 	node->page_offset = off & ~PAGE_MASK;
node              454 fs/hfsplus/bnode.c 		node->page[i] = page;
node              457 fs/hfsplus/bnode.c 	return node;
node              459 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              460 fs/hfsplus/bnode.c 	return node;
node              463 fs/hfsplus/bnode.c void hfs_bnode_unhash(struct hfs_bnode *node)
node              468 fs/hfsplus/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              469 fs/hfsplus/bnode.c 	for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node              470 fs/hfsplus/bnode.c 	     *p && *p != node; p = &(*p)->next_hash)
node              473 fs/hfsplus/bnode.c 	*p = node->next_hash;
node              474 fs/hfsplus/bnode.c 	node->tree->node_hash_cnt--;
node              480 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              486 fs/hfsplus/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              487 fs/hfsplus/bnode.c 	if (node) {
node              488 fs/hfsplus/bnode.c 		hfs_bnode_get(node);
node              490 fs/hfsplus/bnode.c 		wait_event(node->lock_wq,
node              491 fs/hfsplus/bnode.c 			!test_bit(HFS_BNODE_NEW, &node->flags));
node              492 fs/hfsplus/bnode.c 		if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              494 fs/hfsplus/bnode.c 		return node;
node              497 fs/hfsplus/bnode.c 	node = __hfs_bnode_create(tree, num);
node              498 fs/hfsplus/bnode.c 	if (!node)
node              500 fs/hfsplus/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              502 fs/hfsplus/bnode.c 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
node              503 fs/hfsplus/bnode.c 		return node;
node              505 fs/hfsplus/bnode.c 	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) +
node              506 fs/hfsplus/bnode.c 			node->page_offset);
node              507 fs/hfsplus/bnode.c 	node->prev = be32_to_cpu(desc->prev);
node              508 fs/hfsplus/bnode.c 	node->next = be32_to_cpu(desc->next);
node              509 fs/hfsplus/bnode.c 	node->num_recs = be16_to_cpu(desc->num_recs);
node              510 fs/hfsplus/bnode.c 	node->type = desc->type;
node              511 fs/hfsplus/bnode.c 	node->height = desc->height;
node              512 fs/hfsplus/bnode.c 	kunmap(node->page[0]);
node              514 fs/hfsplus/bnode.c 	switch (node->type) {
node              517 fs/hfsplus/bnode.c 		if (node->height != 0)
node              521 fs/hfsplus/bnode.c 		if (node->height != 1)
node              525 fs/hfsplus/bnode.c 		if (node->height <= 1 || node->height > tree->depth)
node              533 fs/hfsplus/bnode.c 	off = hfs_bnode_read_u16(node, rec_off);
node              536 fs/hfsplus/bnode.c 	for (i = 1; i <= node->num_recs; off = next_off, i++) {
node              538 fs/hfsplus/bnode.c 		next_off = hfs_bnode_read_u16(node, rec_off);
node              544 fs/hfsplus/bnode.c 		if (node->type != HFS_NODE_INDEX &&
node              545 fs/hfsplus/bnode.c 		    node->type != HFS_NODE_LEAF)
node              547 fs/hfsplus/bnode.c 		key_size = hfs_bnode_read_u16(node, off) + 2;
node              551 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              552 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              553 fs/hfsplus/bnode.c 	return node;
node              556 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              557 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              558 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              559 fs/hfsplus/bnode.c 	hfs_bnode_put(node);
node              563 fs/hfsplus/bnode.c void hfs_bnode_free(struct hfs_bnode *node)
node              567 fs/hfsplus/bnode.c 	for (i = 0; i < node->tree->pages_per_bnode; i++)
node              568 fs/hfsplus/bnode.c 		if (node->page[i])
node              569 fs/hfsplus/bnode.c 			put_page(node->page[i]);
node              570 fs/hfsplus/bnode.c 	kfree(node);
node              575 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              580 fs/hfsplus/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              582 fs/hfsplus/bnode.c 	if (node) {
node              585 fs/hfsplus/bnode.c 		return node;
node              587 fs/hfsplus/bnode.c 	node = __hfs_bnode_create(tree, num);
node              588 fs/hfsplus/bnode.c 	if (!node)
node              590 fs/hfsplus/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node              591 fs/hfsplus/bnode.c 		hfs_bnode_put(node);
node              595 fs/hfsplus/bnode.c 	pagep = node->page;
node              596 fs/hfsplus/bnode.c 	memset(kmap(*pagep) + node->page_offset, 0,
node              605 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              606 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              608 fs/hfsplus/bnode.c 	return node;
node              611 fs/hfsplus/bnode.c void hfs_bnode_get(struct hfs_bnode *node)
node              613 fs/hfsplus/bnode.c 	if (node) {
node              614 fs/hfsplus/bnode.c 		atomic_inc(&node->refcnt);
node              616 fs/hfsplus/bnode.c 			node->tree->cnid, node->this,
node              617 fs/hfsplus/bnode.c 			atomic_read(&node->refcnt));
node              622 fs/hfsplus/bnode.c void hfs_bnode_put(struct hfs_bnode *node)
node              624 fs/hfsplus/bnode.c 	if (node) {
node              625 fs/hfsplus/bnode.c 		struct hfs_btree *tree = node->tree;
node              629 fs/hfsplus/bnode.c 			node->tree->cnid, node->this,
node              630 fs/hfsplus/bnode.c 			atomic_read(&node->refcnt));
node              631 fs/hfsplus/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
node              632 fs/hfsplus/bnode.c 		if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node              635 fs/hfsplus/bnode.c 			if (!node->page[i])
node              637 fs/hfsplus/bnode.c 			mark_page_accessed(node->page[i]);
node              640 fs/hfsplus/bnode.c 		if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node              641 fs/hfsplus/bnode.c 			hfs_bnode_unhash(node);
node              644 fs/hfsplus/bnode.c 				hfs_bnode_clear(node, 0, tree->node_size);
node              645 fs/hfsplus/bnode.c 			hfs_bmap_free(node);
node              646 fs/hfsplus/bnode.c 			hfs_bnode_free(node);
node               20 fs/hfsplus/brec.c u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off)
node               25 fs/hfsplus/brec.c 	dataoff = node->tree->node_size - (rec + 2) * 2;
node               26 fs/hfsplus/brec.c 	hfs_bnode_read(node, retval, dataoff, 4);
node               32 fs/hfsplus/brec.c u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec)
node               36 fs/hfsplus/brec.c 	if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node               39 fs/hfsplus/brec.c 	if ((node->type == HFS_NODE_INDEX) &&
node               40 fs/hfsplus/brec.c 	   !(node->tree->attributes & HFS_TREE_VARIDXKEYS) &&
node               41 fs/hfsplus/brec.c 	   (node->tree->cnid != HFSPLUS_ATTR_CNID)) {
node               42 fs/hfsplus/brec.c 		retval = node->tree->max_key_len + 2;
node               44 fs/hfsplus/brec.c 		recoff = hfs_bnode_read_u16(node,
node               45 fs/hfsplus/brec.c 			node->tree->node_size - (rec + 1) * 2);
node               48 fs/hfsplus/brec.c 		if (recoff > node->tree->node_size - 2) {
node               53 fs/hfsplus/brec.c 		retval = hfs_bnode_read_u16(node, recoff) + 2;
node               54 fs/hfsplus/brec.c 		if (retval > node->tree->max_key_len + 2) {
node               66 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node;
node               76 fs/hfsplus/brec.c 		node = hfs_bnode_find(tree, tree->leaf_head);
node               77 fs/hfsplus/brec.c 		if (IS_ERR(node))
node               78 fs/hfsplus/brec.c 			return PTR_ERR(node);
node               79 fs/hfsplus/brec.c 		fd->bnode = node;
node               89 fs/hfsplus/brec.c 	node = fd->bnode;
node               90 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node               92 fs/hfsplus/brec.c 	end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node               93 fs/hfsplus/brec.c 	end_off = hfs_bnode_read_u16(node, end_rec_off);
node              105 fs/hfsplus/brec.c 	if (node->type == HFS_NODE_LEAF) {
node              109 fs/hfsplus/brec.c 	node->num_recs++;
node              111 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node,
node              113 fs/hfsplus/brec.c 		node->num_recs);
node              114 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node              122 fs/hfsplus/brec.c 		data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node              123 fs/hfsplus/brec.c 		hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node              128 fs/hfsplus/brec.c 	hfs_bnode_move(node, data_off + size, data_off,
node              132 fs/hfsplus/brec.c 	hfs_bnode_write(node, fd->search_key, data_off, key_len);
node              133 fs/hfsplus/brec.c 	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node              134 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              140 fs/hfsplus/brec.c 	if (!rec && new_node != node) {
node              141 fs/hfsplus/brec.c 		hfs_bnode_read_key(node, fd->search_key, data_off + size);
node              182 fs/hfsplus/brec.c 	struct hfs_bnode *node, *parent;
node              186 fs/hfsplus/brec.c 	node = fd->bnode;
node              189 fs/hfsplus/brec.c 	end_off = tree->node_size - (node->num_recs + 1) * 2;
node              191 fs/hfsplus/brec.c 	if (node->type == HFS_NODE_LEAF) {
node              195 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              198 fs/hfsplus/brec.c 	if (!--node->num_recs) {
node              199 fs/hfsplus/brec.c 		hfs_bnode_unlink(node);
node              200 fs/hfsplus/brec.c 		if (!node->parent)
node              202 fs/hfsplus/brec.c 		parent = hfs_bnode_find(tree, node->parent);
node              205 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              206 fs/hfsplus/brec.c 		node = fd->bnode = parent;
node              208 fs/hfsplus/brec.c 		__hfs_brec_find(node, fd, hfs_find_rec_by_key);
node              211 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node,
node              213 fs/hfsplus/brec.c 		node->num_recs);
node              220 fs/hfsplus/brec.c 		data_off = hfs_bnode_read_u16(node, rec_off);
node              221 fs/hfsplus/brec.c 		hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node              226 fs/hfsplus/brec.c 	hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node              229 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              238 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node, *next_node;
node              244 fs/hfsplus/brec.c 	node = fd->bnode;
node              248 fs/hfsplus/brec.c 	hfs_bnode_get(node);
node              250 fs/hfsplus/brec.c 		node->this, new_node->this, node->next);
node              251 fs/hfsplus/brec.c 	new_node->next = node->next;
node              252 fs/hfsplus/brec.c 	new_node->prev = node->this;
node              253 fs/hfsplus/brec.c 	new_node->parent = node->parent;
node              254 fs/hfsplus/brec.c 	new_node->type = node->type;
node              255 fs/hfsplus/brec.c 	new_node->height = node->height;
node              257 fs/hfsplus/brec.c 	if (node->next)
node              258 fs/hfsplus/brec.c 		next_node = hfs_bnode_find(tree, node->next);
node              263 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              268 fs/hfsplus/brec.c 	size = tree->node_size / 2 - node->num_recs * 2 - 14;
node              272 fs/hfsplus/brec.c 		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              276 fs/hfsplus/brec.c 		if (++num_recs < node->num_recs)
node              279 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              292 fs/hfsplus/brec.c 		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              294 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              301 fs/hfsplus/brec.c 	new_node->num_recs = node->num_recs - num_recs;
node              302 fs/hfsplus/brec.c 	node->num_recs = num_recs;
node              313 fs/hfsplus/brec.c 		data_end = hfs_bnode_read_u16(node, old_rec_off);
node              318 fs/hfsplus/brec.c 	hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node              330 fs/hfsplus/brec.c 	node->next = new_node->this;
node              331 fs/hfsplus/brec.c 	hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node              332 fs/hfsplus/brec.c 	node_desc.next = cpu_to_be32(node->next);
node              333 fs/hfsplus/brec.c 	node_desc.num_recs = cpu_to_be16(node->num_recs);
node              334 fs/hfsplus/brec.c 	hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node              343 fs/hfsplus/brec.c 	} else if (node->this == tree->leaf_tail) {
node              349 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              351 fs/hfsplus/brec.c 	hfs_bnode_put(node);
node              359 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node, *parent;
node              365 fs/hfsplus/brec.c 	node = fd->bnode;
node              367 fs/hfsplus/brec.c 	if (!node->parent)
node              371 fs/hfsplus/brec.c 	parent = hfs_bnode_find(tree, node->parent);
node              383 fs/hfsplus/brec.c 		newkeylen = hfs_bnode_read_u16(node, 14) + 2;
node              423 fs/hfsplus/brec.c 	hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node              426 fs/hfsplus/brec.c 	hfs_bnode_put(node);
node              427 fs/hfsplus/brec.c 	node = parent;
node              447 fs/hfsplus/brec.c 			if (new_node == node)
node              450 fs/hfsplus/brec.c 			hfs_bnode_read_key(node, fd->search_key, 14);
node              455 fs/hfsplus/brec.c 	if (!rec && node->parent)
node              458 fs/hfsplus/brec.c 	fd->bnode = node;
node              464 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node;
node              469 fs/hfsplus/brec.c 	node = NULL;
node              471 fs/hfsplus/brec.c 		node = hfs_bnode_find(tree, tree->root);
node              472 fs/hfsplus/brec.c 		if (IS_ERR(node))
node              473 fs/hfsplus/brec.c 			return PTR_ERR(node);
node              477 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              506 fs/hfsplus/brec.c 	if (node) {
node              508 fs/hfsplus/brec.c 		node->parent = tree->root;
node              509 fs/hfsplus/brec.c 		if (node->type == HFS_NODE_LEAF ||
node              512 fs/hfsplus/brec.c 			key_size = hfs_bnode_read_u16(node, 14) + 2;
node              515 fs/hfsplus/brec.c 		hfs_bnode_copy(new_node, 14, node, 14, key_size);
node              522 fs/hfsplus/brec.c 		cnid = cpu_to_be32(node->this);
node              528 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              260 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              267 fs/hfsplus/btree.c 		while ((node = tree->node_hash[i])) {
node              268 fs/hfsplus/btree.c 			tree->node_hash[i] = node->next_hash;
node              269 fs/hfsplus/btree.c 			if (atomic_read(&node->refcnt))
node              272 fs/hfsplus/btree.c 					node->tree->cnid, node->this,
node              273 fs/hfsplus/btree.c 					atomic_read(&node->refcnt));
node              274 fs/hfsplus/btree.c 			hfs_bnode_free(node);
node              285 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              288 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, 0);
node              289 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              293 fs/hfsplus/btree.c 	page = node->page[0];
node              308 fs/hfsplus/btree.c 	hfs_bnode_put(node);
node              315 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              319 fs/hfsplus/btree.c 	node = hfs_bnode_create(tree, idx);
node              320 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              321 fs/hfsplus/btree.c 		return node;
node              328 fs/hfsplus/btree.c 	node->type = HFS_NODE_MAP;
node              329 fs/hfsplus/btree.c 	node->num_recs = 1;
node              330 fs/hfsplus/btree.c 	hfs_bnode_clear(node, 0, tree->node_size);
node              337 fs/hfsplus/btree.c 	hfs_bnode_write(node, &desc, 0, sizeof(desc));
node              338 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, 14, 0x8000);
node              339 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node              340 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node              342 fs/hfsplus/btree.c 	return node;
node              375 fs/hfsplus/btree.c 	struct hfs_bnode *node, *next_node;
node              389 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, nidx);
node              390 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              391 fs/hfsplus/btree.c 		return node;
node              392 fs/hfsplus/btree.c 	len = hfs_brec_lenoff(node, 2, &off16);
node              395 fs/hfsplus/btree.c 	off += node->page_offset;
node              396 fs/hfsplus/btree.c 	pagep = node->page + (off >> PAGE_SHIFT);
node              413 fs/hfsplus/btree.c 						hfs_bnode_put(node);
node              428 fs/hfsplus/btree.c 		nidx = node->next;
node              431 fs/hfsplus/btree.c 			next_node = hfs_bmap_new_bmap(node, idx);
node              434 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              437 fs/hfsplus/btree.c 		node = next_node;
node              439 fs/hfsplus/btree.c 		len = hfs_brec_lenoff(node, 0, &off16);
node              441 fs/hfsplus/btree.c 		off += node->page_offset;
node              442 fs/hfsplus/btree.c 		pagep = node->page + (off >> PAGE_SHIFT);
node              448 fs/hfsplus/btree.c void hfs_bmap_free(struct hfs_bnode *node)
node              456 fs/hfsplus/btree.c 	hfs_dbg(BNODE_MOD, "btree_free_node: %u\n", node->this);
node              457 fs/hfsplus/btree.c 	BUG_ON(!node->this);
node              458 fs/hfsplus/btree.c 	tree = node->tree;
node              459 fs/hfsplus/btree.c 	nidx = node->this;
node              460 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, 0);
node              461 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              463 fs/hfsplus/btree.c 	len = hfs_brec_lenoff(node, 2, &off);
node              468 fs/hfsplus/btree.c 		i = node->next;
node              473 fs/hfsplus/btree.c 				node->this);
node              474 fs/hfsplus/btree.c 			hfs_bnode_put(node);
node              477 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              478 fs/hfsplus/btree.c 		node = hfs_bnode_find(tree, i);
node              479 fs/hfsplus/btree.c 		if (IS_ERR(node))
node              481 fs/hfsplus/btree.c 		if (node->type != HFS_NODE_MAP) {
node              485 fs/hfsplus/btree.c 				node->this, node->type);
node              486 fs/hfsplus/btree.c 			hfs_bnode_put(node);
node              489 fs/hfsplus/btree.c 		len = hfs_brec_lenoff(node, 0, &off);
node              491 fs/hfsplus/btree.c 	off += node->page_offset + nidx / 8;
node              492 fs/hfsplus/btree.c 	page = node->page[off >> PAGE_SHIFT];
node              500 fs/hfsplus/btree.c 			node->this, node->type);
node              502 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              508 fs/hfsplus/btree.c 	hfs_bnode_put(node);
node              401 fs/hfsplus/hfsplus_fs.h void hfs_bmap_free(struct hfs_bnode *node);
node              404 fs/hfsplus/hfsplus_fs.h void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len);
node              405 fs/hfsplus/hfsplus_fs.h u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off);
node              406 fs/hfsplus/hfsplus_fs.h u8 hfs_bnode_read_u8(struct hfs_bnode *node, int off);
node              407 fs/hfsplus/hfsplus_fs.h void hfs_bnode_read_key(struct hfs_bnode *node, void *key, int off);
node              408 fs/hfsplus/hfsplus_fs.h void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len);
node              409 fs/hfsplus/hfsplus_fs.h void hfs_bnode_write_u16(struct hfs_bnode *node, int off, u16 data);
node              410 fs/hfsplus/hfsplus_fs.h void hfs_bnode_clear(struct hfs_bnode *node, int off, int len);
node              413 fs/hfsplus/hfsplus_fs.h void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len);
node              414 fs/hfsplus/hfsplus_fs.h void hfs_bnode_dump(struct hfs_bnode *node);
node              415 fs/hfsplus/hfsplus_fs.h void hfs_bnode_unlink(struct hfs_bnode *node);
node              417 fs/hfsplus/hfsplus_fs.h void hfs_bnode_unhash(struct hfs_bnode *node);
node              419 fs/hfsplus/hfsplus_fs.h void hfs_bnode_free(struct hfs_bnode *node);
node              421 fs/hfsplus/hfsplus_fs.h void hfs_bnode_get(struct hfs_bnode *node);
node              422 fs/hfsplus/hfsplus_fs.h void hfs_bnode_put(struct hfs_bnode *node);
node              426 fs/hfsplus/hfsplus_fs.h u16 hfs_brec_lenoff(struct hfs_bnode *node, u16 rec, u16 *off);
node              427 fs/hfsplus/hfsplus_fs.h u16 hfs_brec_keylen(struct hfs_bnode *node, u16 rec);
node               71 fs/hfsplus/inode.c 	struct hfs_bnode *node;
node               95 fs/hfsplus/inode.c 		node = hfs_bnode_findhash(tree, nidx);
node               96 fs/hfsplus/inode.c 		if (!node)
node               98 fs/hfsplus/inode.c 		else if (atomic_read(&node->refcnt))
node              100 fs/hfsplus/inode.c 		if (res && node) {
node              101 fs/hfsplus/inode.c 			hfs_bnode_unhash(node);
node              102 fs/hfsplus/inode.c 			hfs_bnode_free(node);
node              111 fs/hfsplus/inode.c 			node = hfs_bnode_findhash(tree, nidx++);
node              112 fs/hfsplus/inode.c 			if (!node)
node              114 fs/hfsplus/inode.c 			if (atomic_read(&node->refcnt)) {
node              118 fs/hfsplus/inode.c 			hfs_bnode_unhash(node);
node              119 fs/hfsplus/inode.c 			hfs_bnode_free(node);
node               61 fs/hpfs/anode.c secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsigned fsecno)
node               73 fs/hpfs/anode.c 		if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1;
node               76 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) return -1;
node               79 fs/hpfs/anode.c 	a = node;
node              101 fs/hpfs/anode.c 				fnod?'f':'a', node);
node              113 fs/hpfs/anode.c 			hpfs_error(s, "empty file %08x, trying to add sector %08x", node, fsecno);
node              117 fs/hpfs/anode.c 		se = !fnod ? node : (node + 16384) & ~16383;
node              125 fs/hpfs/anode.c 		up = a != node ? le32_to_cpu(anode->up) : -1;
node              131 fs/hpfs/anode.c 		if (a == node && fnod) {
node              132 fs/hpfs/anode.c 			anode->up = cpu_to_le32(node);
node              163 fs/hpfs/anode.c 	if ((a == node && fnod) || na == -1) return se;
node              169 fs/hpfs/anode.c 		if (up != node || !fnod) {
node              188 fs/hpfs/anode.c 				if (up == node && fnod)
node              197 fs/hpfs/anode.c 		up = up != node ? le32_to_cpu(anode->up) : -1;
node              221 fs/hpfs/anode.c 		anode->up = cpu_to_le32(node);
node              228 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) {
node              234 fs/hpfs/anode.c 		if (!(fnode = hpfs_map_fnode(s, node, &bh))) {
node              240 fs/hpfs/anode.c 	ranode->up = cpu_to_le32(node);
node              405 fs/hpfs/anode.c 	anode_secno node = f;
node              432 fs/hpfs/anode.c 		hpfs_error(s, "internal btree %08x doesn't end with -1", node);
node              445 fs/hpfs/anode.c 		node = le32_to_cpu(btree->u.internal[i].down);
node              448 fs/hpfs/anode.c 			if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree"))
node              450 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) return;
node               81 fs/jffs2/debug.c 		struct jffs2_full_dnode *fn = frag->node;
node               99 fs/jffs2/debug.c 					&& frag_prev(frag)->size < PAGE_SIZE && frag_prev(frag)->node) {
node              106 fs/jffs2/debug.c 					&& frag_next(frag)->size < PAGE_SIZE && frag_next(frag)->node) {
node              711 fs/jffs2/debug.c 		if (this->node)
node              713 fs/jffs2/debug.c 				this->ofs, this->ofs+this->size, ref_offset(this->node->raw),
node              714 fs/jffs2/debug.c 				ref_flags(this->node->raw), this, frag_left(this), frag_right(this),
node              775 fs/jffs2/debug.c 	union jffs2_node_union node;
node              783 fs/jffs2/debug.c 	ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node);
node              790 fs/jffs2/debug.c 	printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic));
node              791 fs/jffs2/debug.c 	printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype));
node              792 fs/jffs2/debug.c 	printk(JFFS2_DBG "totlen:\t%#08x\n", je32_to_cpu(node.u.totlen));
node              793 fs/jffs2/debug.c 	printk(JFFS2_DBG "hdr_crc:\t%#08x\n", je32_to_cpu(node.u.hdr_crc));
node              795 fs/jffs2/debug.c 	crc = crc32(0, &node.u, sizeof(node.u) - 4);
node              796 fs/jffs2/debug.c 	if (crc != je32_to_cpu(node.u.hdr_crc)) {
node              801 fs/jffs2/debug.c 	if (je16_to_cpu(node.u.magic) != JFFS2_MAGIC_BITMASK &&
node              802 fs/jffs2/debug.c 		je16_to_cpu(node.u.magic) != JFFS2_OLD_MAGIC_BITMASK)
node              805 fs/jffs2/debug.c 			je16_to_cpu(node.u.magic), JFFS2_MAGIC_BITMASK);
node              809 fs/jffs2/debug.c 	switch(je16_to_cpu(node.u.nodetype)) {
node              814 fs/jffs2/debug.c 		printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.i.ino));
node              815 fs/jffs2/debug.c 		printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.i.version));
node              816 fs/jffs2/debug.c 		printk(JFFS2_DBG "mode:\t%#08x\n", node.i.mode.m);
node              817 fs/jffs2/debug.c 		printk(JFFS2_DBG "uid:\t%#04x\n", je16_to_cpu(node.i.uid));
node              818 fs/jffs2/debug.c 		printk(JFFS2_DBG "gid:\t%#04x\n", je16_to_cpu(node.i.gid));
node              819 fs/jffs2/debug.c 		printk(JFFS2_DBG "isize:\t%#08x\n", je32_to_cpu(node.i.isize));
node              820 fs/jffs2/debug.c 		printk(JFFS2_DBG "atime:\t%#08x\n", je32_to_cpu(node.i.atime));
node              821 fs/jffs2/debug.c 		printk(JFFS2_DBG "mtime:\t%#08x\n", je32_to_cpu(node.i.mtime));
node              822 fs/jffs2/debug.c 		printk(JFFS2_DBG "ctime:\t%#08x\n", je32_to_cpu(node.i.ctime));
node              823 fs/jffs2/debug.c 		printk(JFFS2_DBG "offset:\t%#08x\n", je32_to_cpu(node.i.offset));
node              824 fs/jffs2/debug.c 		printk(JFFS2_DBG "csize:\t%#08x\n", je32_to_cpu(node.i.csize));
node              825 fs/jffs2/debug.c 		printk(JFFS2_DBG "dsize:\t%#08x\n", je32_to_cpu(node.i.dsize));
node              826 fs/jffs2/debug.c 		printk(JFFS2_DBG "compr:\t%#02x\n", node.i.compr);
node              827 fs/jffs2/debug.c 		printk(JFFS2_DBG "usercompr:\t%#02x\n", node.i.usercompr);
node              828 fs/jffs2/debug.c 		printk(JFFS2_DBG "flags:\t%#04x\n", je16_to_cpu(node.i.flags));
node              829 fs/jffs2/debug.c 		printk(JFFS2_DBG "data_crc:\t%#08x\n", je32_to_cpu(node.i.data_crc));
node              830 fs/jffs2/debug.c 		printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.i.node_crc));
node              832 fs/jffs2/debug.c 		crc = crc32(0, &node.i, sizeof(node.i) - 8);
node              833 fs/jffs2/debug.c 		if (crc != je32_to_cpu(node.i.node_crc)) {
node              842 fs/jffs2/debug.c 		printk(JFFS2_DBG "pino:\t%#08x\n", je32_to_cpu(node.d.pino));
node              843 fs/jffs2/debug.c 		printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.d.version));
node              844 fs/jffs2/debug.c 		printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.d.ino));
node              845 fs/jffs2/debug.c 		printk(JFFS2_DBG "mctime:\t%#08x\n", je32_to_cpu(node.d.mctime));
node              846 fs/jffs2/debug.c 		printk(JFFS2_DBG "nsize:\t%#02x\n", node.d.nsize);
node              847 fs/jffs2/debug.c 		printk(JFFS2_DBG "type:\t%#02x\n", node.d.type);
node              848 fs/jffs2/debug.c 		printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.d.node_crc));
node              849 fs/jffs2/debug.c 		printk(JFFS2_DBG "name_crc:\t%#08x\n", je32_to_cpu(node.d.name_crc));
node              851 fs/jffs2/debug.c 		node.d.name[node.d.nsize] = '\0';
node              852 fs/jffs2/debug.c 		printk(JFFS2_DBG "name:\t\"%s\"\n", node.d.name);
node              854 fs/jffs2/debug.c 		crc = crc32(0, &node.d, sizeof(node.d) - 8);
node              855 fs/jffs2/debug.c 		if (crc != je32_to_cpu(node.d.node_crc)) {
node              535 fs/jffs2/gc.c  		if (frag->node && frag->node->raw == raw) {
node              536 fs/jffs2/gc.c  			fn = frag->node;
node              540 fs/jffs2/gc.c  			if (nrfrags == frag->node->frags)
node              549 fs/jffs2/gc.c  				frag->node->raw = f->inocache->nodes;
node              595 fs/jffs2/gc.c  	union jffs2_node_union *node;
node              624 fs/jffs2/gc.c  	node = kmalloc(rawlen, GFP_KERNEL);
node              625 fs/jffs2/gc.c  	if (!node)
node              628 fs/jffs2/gc.c  	ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node);
node              634 fs/jffs2/gc.c  	crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4);
node              635 fs/jffs2/gc.c  	if (je32_to_cpu(node->u.hdr_crc) != crc) {
node              637 fs/jffs2/gc.c  			ref_offset(raw), je32_to_cpu(node->u.hdr_crc), crc);
node              641 fs/jffs2/gc.c  	switch(je16_to_cpu(node->u.nodetype)) {
node              643 fs/jffs2/gc.c  		crc = crc32(0, node, sizeof(node->i)-8);
node              644 fs/jffs2/gc.c  		if (je32_to_cpu(node->i.node_crc) != crc) {
node              646 fs/jffs2/gc.c  				ref_offset(raw), je32_to_cpu(node->i.node_crc),
node              651 fs/jffs2/gc.c  		if (je32_to_cpu(node->i.dsize)) {
node              652 fs/jffs2/gc.c  			crc = crc32(0, node->i.data, je32_to_cpu(node->i.csize));
node              653 fs/jffs2/gc.c  			if (je32_to_cpu(node->i.data_crc) != crc) {
node              656 fs/jffs2/gc.c  					je32_to_cpu(node->i.data_crc), crc);
node              663 fs/jffs2/gc.c  		crc = crc32(0, node, sizeof(node->d)-8);
node              664 fs/jffs2/gc.c  		if (je32_to_cpu(node->d.node_crc) != crc) {
node              667 fs/jffs2/gc.c  				je32_to_cpu(node->d.node_crc), crc);
node              671 fs/jffs2/gc.c  		if (strnlen(node->d.name, node->d.nsize) != node->d.nsize) {
node              677 fs/jffs2/gc.c  		if (node->d.nsize) {
node              678 fs/jffs2/gc.c  			crc = crc32(0, node->d.name, node->d.nsize);
node              679 fs/jffs2/gc.c  			if (je32_to_cpu(node->d.name_crc) != crc) {
node              682 fs/jffs2/gc.c  					je32_to_cpu(node->d.name_crc), crc);
node              691 fs/jffs2/gc.c  				ref_offset(raw), je16_to_cpu(node->u.nodetype));
node              700 fs/jffs2/gc.c  	ret = jffs2_flash_write(c, phys_ofs, rawlen, &retlen, (char *)node);
node              751 fs/jffs2/gc.c  	kfree(node);
node             1143 fs/jffs2/gc.c  		if (frag->node == fn) {
node             1144 fs/jffs2/gc.c  			frag->node = new_fn;
node             1217 fs/jffs2/gc.c  			if (!frag->node || !frag->node->raw) {
node             1227 fs/jffs2/gc.c  				struct jffs2_raw_node_ref *raw = frag->node->raw;
node             1273 fs/jffs2/gc.c  			if (!frag->node || !frag->node->raw) {
node             1283 fs/jffs2/gc.c  				struct jffs2_raw_node_ref *raw = frag->node->raw;
node              286 fs/jffs2/malloc.c 	xd->node = (void *)xd;
node              306 fs/jffs2/malloc.c 	ref->node = (void *)ref;
node               93 fs/jffs2/nodelist.c 	if (frag->node && (frag->ofs & (PAGE_SIZE - 1)) == 0) {
node               96 fs/jffs2/nodelist.c 		frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE;
node              104 fs/jffs2/nodelist.c 	if (this->node) {
node              105 fs/jffs2/nodelist.c 		this->node->frags--;
node              106 fs/jffs2/nodelist.c 		if (!this->node->frags) {
node              109 fs/jffs2/nodelist.c 				ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size);
node              110 fs/jffs2/nodelist.c 			jffs2_mark_node_obsolete(c, this->node->raw);
node              111 fs/jffs2/nodelist.c 			jffs2_free_full_dnode(this->node);
node              114 fs/jffs2/nodelist.c 				ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags);
node              115 fs/jffs2/nodelist.c 			mark_ref_normal(this->node->raw);
node              157 fs/jffs2/nodelist.c 		newfrag->node = fn;
node              173 fs/jffs2/nodelist.c 	if (lastend < newfrag->node->ofs) {
node              177 fs/jffs2/nodelist.c 		holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend);
node              221 fs/jffs2/nodelist.c 	this = jffs2_lookup_node_frag(root, newfrag->node->ofs);
node              225 fs/jffs2/nodelist.c 			  this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this);
node              241 fs/jffs2/nodelist.c 			if (this->node)
node              242 fs/jffs2/nodelist.c 				mark_ref_normal(this->node->raw);
node              243 fs/jffs2/nodelist.c 			mark_ref_normal(newfrag->node->raw);
node              249 fs/jffs2/nodelist.c 	if (this->node)
node              252 fs/jffs2/nodelist.c 		ref_offset(this->node->raw), ref_flags(this->node->raw));
node              265 fs/jffs2/nodelist.c 		mark_ref_normal(newfrag->node->raw);
node              266 fs/jffs2/nodelist.c 		if (this->node)
node              267 fs/jffs2/nodelist.c 			mark_ref_normal(this->node->raw);
node              273 fs/jffs2/nodelist.c 			if (this->node)
node              275 fs/jffs2/nodelist.c 					this->ofs, this->ofs+this->size, ref_offset(this->node->raw));
node              281 fs/jffs2/nodelist.c 			newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size,
node              285 fs/jffs2/nodelist.c 			if (this->node)
node              286 fs/jffs2/nodelist.c 				this->node->frags++;
node              352 fs/jffs2/nodelist.c 	if (this->node)
node              353 fs/jffs2/nodelist.c 		mark_ref_normal(this->node->raw);
node              354 fs/jffs2/nodelist.c 	mark_ref_normal(newfrag->node->raw);
node              374 fs/jffs2/nodelist.c 	newfrag->node->frags = 1;
node              390 fs/jffs2/nodelist.c 		if (prev->node)
node              391 fs/jffs2/nodelist.c 			mark_ref_normal(prev->node->raw);
node              399 fs/jffs2/nodelist.c 			if (next->node)
node              400 fs/jffs2/nodelist.c 				mark_ref_normal(next->node->raw);
node              571 fs/jffs2/nodelist.c 		if (frag->node && !(--frag->node->frags)) {
node              575 fs/jffs2/nodelist.c 				jffs2_mark_node_obsolete(c, frag->node->raw);
node              577 fs/jffs2/nodelist.c 			jffs2_free_full_dnode(frag->node);
node              272 fs/jffs2/nodelist.h 	struct jffs2_full_dnode *node; /* NULL for holes */
node              329 fs/jffs2/nodelist.h 	struct rb_node *node = rb_first(root);
node              331 fs/jffs2/nodelist.h 	if (!node)
node              334 fs/jffs2/nodelist.h 	return rb_entry(node, struct jffs2_node_frag, rb);
node              339 fs/jffs2/nodelist.h 	struct rb_node *node = rb_last(root);
node              341 fs/jffs2/nodelist.h 	if (!node)
node              344 fs/jffs2/nodelist.h 	return rb_entry(node, struct jffs2_node_frag, rb);
node              191 fs/jffs2/read.c 		} else if (unlikely(!frag->node)) {
node              210 fs/jffs2/read.c 				  ref_offset(frag->node->raw),
node              211 fs/jffs2/read.c 				  ref_flags(frag->node->raw));
node              212 fs/jffs2/read.c 			ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen);
node              402 fs/jffs2/readinode.c static void eat_last(struct rb_root *root, struct rb_node *node)
node              404 fs/jffs2/readinode.c 	struct rb_node *parent = rb_parent(node);
node              408 fs/jffs2/readinode.c 	BUG_ON(node->rb_right);
node              412 fs/jffs2/readinode.c 	else if (node == parent->rb_left)
node              417 fs/jffs2/readinode.c 	*link = node->rb_left;
node              418 fs/jffs2/readinode.c 	if (node->rb_left)
node              419 fs/jffs2/readinode.c 		node->rb_left->__rb_parent_color = node->__rb_parent_color;
node              970 fs/jffs2/readinode.c 	union jffs2_node_union *node;
node             1041 fs/jffs2/readinode.c 		node = (union jffs2_node_union *)buf;
node             1044 fs/jffs2/readinode.c 		if (je32_to_cpu(node->u.hdr_crc) != crc32(0, node, sizeof(node->u)-4)) {
node             1046 fs/jffs2/readinode.c 				     ref_offset(ref), je16_to_cpu(node->u.magic),
node             1047 fs/jffs2/readinode.c 				     je16_to_cpu(node->u.nodetype),
node             1048 fs/jffs2/readinode.c 				     je32_to_cpu(node->u.totlen),
node             1049 fs/jffs2/readinode.c 				     je32_to_cpu(node->u.hdr_crc));
node             1054 fs/jffs2/readinode.c 		if (je16_to_cpu(node->u.magic) != JFFS2_MAGIC_BITMASK) {
node             1057 fs/jffs2/readinode.c 				     je16_to_cpu(node->u.magic), ref_offset(ref));
node             1062 fs/jffs2/readinode.c 		switch (je16_to_cpu(node->u.nodetype)) {
node             1073 fs/jffs2/readinode.c 			err = read_direntry(c, ref, &node->d, retlen, rii);
node             1088 fs/jffs2/readinode.c 			err = read_dnode(c, ref, &node->i, len, rii);
node             1102 fs/jffs2/readinode.c 			err = read_unknown(c, ref, &node->u);
node             1300 fs/jffs2/readinode.c 		f->metadata = frag_first(&f->fragtree)->node;
node              364 fs/jffs2/scan.c 		raw->next_in_ino = xd->node->next_in_ino;
node              365 fs/jffs2/scan.c 		xd->node->next_in_ino = raw;
node              444 fs/jffs2/scan.c 	struct jffs2_unknown_node *node;
node              638 fs/jffs2/scan.c 		if (jeb->offset + c->sector_size < ofs + sizeof(*node)) {
node              642 fs/jffs2/scan.c 				  sizeof(*node));
node              648 fs/jffs2/scan.c 		if (buf_ofs + buf_len < ofs + sizeof(*node)) {
node              659 fs/jffs2/scan.c 		node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs];
node              721 fs/jffs2/scan.c 		if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) {
node              729 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) {
node              736 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) {
node              744 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) != JFFS2_MAGIC_BITMASK) {
node              749 fs/jffs2/scan.c 				     je16_to_cpu(node->magic));
node              756 fs/jffs2/scan.c 		crcnode.magic = node->magic;
node              757 fs/jffs2/scan.c 		crcnode.nodetype = cpu_to_je16( je16_to_cpu(node->nodetype) | JFFS2_NODE_ACCURATE);
node              758 fs/jffs2/scan.c 		crcnode.totlen = node->totlen;
node              761 fs/jffs2/scan.c 		if (hdr_crc != je32_to_cpu(node->hdr_crc)) {
node              764 fs/jffs2/scan.c 				     ofs, je16_to_cpu(node->magic),
node              765 fs/jffs2/scan.c 				     je16_to_cpu(node->nodetype),
node              766 fs/jffs2/scan.c 				     je32_to_cpu(node->totlen),
node              767 fs/jffs2/scan.c 				     je32_to_cpu(node->hdr_crc),
node              775 fs/jffs2/scan.c 		if (ofs + je32_to_cpu(node->totlen) > jeb->offset + c->sector_size) {
node              778 fs/jffs2/scan.c 				ofs, je32_to_cpu(node->totlen));
node              786 fs/jffs2/scan.c 		if (!(je16_to_cpu(node->nodetype) & JFFS2_NODE_ACCURATE)) {
node              790 fs/jffs2/scan.c 			if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              792 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              796 fs/jffs2/scan.c 		switch(je16_to_cpu(node->nodetype)) {
node              807 fs/jffs2/scan.c 				node = (void *)buf;
node              809 fs/jffs2/scan.c 			err = jffs2_scan_inode_node(c, jeb, (void *)node, ofs, s);
node              811 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              815 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              818 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len,
node              824 fs/jffs2/scan.c 				node = (void *)buf;
node              826 fs/jffs2/scan.c 			err = jffs2_scan_dirent_node(c, jeb, (void *)node, ofs, s);
node              828 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              833 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              836 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len,
node              842 fs/jffs2/scan.c 				node = (void *)buf;
node              844 fs/jffs2/scan.c 			err = jffs2_scan_xattr_node(c, jeb, (void *)node, ofs, s);
node              847 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              850 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              853 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len,
node              859 fs/jffs2/scan.c 				node = (void *)buf;
node              861 fs/jffs2/scan.c 			err = jffs2_scan_xref_node(c, jeb, (void *)node, ofs, s);
node              864 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              870 fs/jffs2/scan.c 			if (je32_to_cpu(node->totlen) != c->cleanmarker_size) {
node              872 fs/jffs2/scan.c 					  ofs, je32_to_cpu(node->totlen),
node              892 fs/jffs2/scan.c 				jffs2_sum_add_padding_mem(s, je32_to_cpu(node->totlen));
node              893 fs/jffs2/scan.c 			if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              895 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              899 fs/jffs2/scan.c 			switch (je16_to_cpu(node->nodetype) & JFFS2_COMPAT_MASK) {
node              902 fs/jffs2/scan.c 					  je16_to_cpu(node->nodetype), ofs);
node              906 fs/jffs2/scan.c 				if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              908 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              913 fs/jffs2/scan.c 					  je16_to_cpu(node->nodetype), ofs);
node              918 fs/jffs2/scan.c 					  je16_to_cpu(node->nodetype), ofs);
node              919 fs/jffs2/scan.c 				if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              921 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              926 fs/jffs2/scan.c 					  je16_to_cpu(node->nodetype), ofs);
node              928 fs/jffs2/scan.c 				jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(node->totlen)), NULL);
node              932 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              251 fs/jffs2/summary.c 	union jffs2_node_union *node;
node              259 fs/jffs2/summary.c 	node = invecs[0].iov_base;
node              263 fs/jffs2/summary.c 	switch (je16_to_cpu(node->u.nodetype)) {
node              271 fs/jffs2/summary.c 			temp->nodetype = node->i.nodetype;
node              272 fs/jffs2/summary.c 			temp->inode = node->i.ino;
node              273 fs/jffs2/summary.c 			temp->version = node->i.version;
node              275 fs/jffs2/summary.c 			temp->totlen = node->i.totlen;
node              283 fs/jffs2/summary.c 				kmalloc(sizeof(struct jffs2_sum_dirent_mem) + node->d.nsize, GFP_KERNEL);
node              288 fs/jffs2/summary.c 			temp->nodetype = node->d.nodetype;
node              289 fs/jffs2/summary.c 			temp->totlen = node->d.totlen;
node              291 fs/jffs2/summary.c 			temp->pino = node->d.pino;
node              292 fs/jffs2/summary.c 			temp->version = node->d.version;
node              293 fs/jffs2/summary.c 			temp->ino = node->d.ino;
node              294 fs/jffs2/summary.c 			temp->nsize = node->d.nsize;
node              295 fs/jffs2/summary.c 			temp->type = node->d.type;
node              300 fs/jffs2/summary.c 					memcpy(temp->name,node->d.name,node->d.nsize);
node              304 fs/jffs2/summary.c 					memcpy(temp->name,invecs[1].iov_base,node->d.nsize);
node              321 fs/jffs2/summary.c 			temp->nodetype = node->x.nodetype;
node              322 fs/jffs2/summary.c 			temp->xid = node->x.xid;
node              323 fs/jffs2/summary.c 			temp->version = node->x.version;
node              324 fs/jffs2/summary.c 			temp->totlen = node->x.totlen;
node              335 fs/jffs2/summary.c 			temp->nodetype = node->r.nodetype;
node              344 fs/jffs2/summary.c 			c->summary->sum_padded += je32_to_cpu(node->u.totlen);
node              510 fs/jffs2/summary.c 					raw->next_in_ino = xd->node->next_in_ino;
node              511 fs/jffs2/summary.c 					xd->node->next_in_ino = raw;
node              543 fs/jffs2/summary.c 				*pseudo_random += ref->node->flash_offset;
node              185 fs/jffs2/wbuf.c 							    union jffs2_node_union *node)
node              191 fs/jffs2/wbuf.c 		    node, je16_to_cpu(node->u.magic), je16_to_cpu(node->u.nodetype));
node              193 fs/jffs2/wbuf.c 	BUG_ON(je16_to_cpu(node->u.magic) != 0x1985 &&
node              194 fs/jffs2/wbuf.c 	       je16_to_cpu(node->u.magic) != 0);
node              196 fs/jffs2/wbuf.c 	switch (je16_to_cpu(node->u.nodetype)) {
node              202 fs/jffs2/wbuf.c 		frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset));
node              205 fs/jffs2/wbuf.c 		while (!frag->node || frag->node->raw != raw) {
node              209 fs/jffs2/wbuf.c 		dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node);
node              210 fs/jffs2/wbuf.c 		return &frag->node->raw;
node              223 fs/jffs2/wbuf.c 			    je16_to_cpu(node->u.nodetype));
node              478 fs/jffs2/wbuf.c 			BUG_ON(xd->node != raw);
node              479 fs/jffs2/wbuf.c 			adjust_ref = &xd->node;
node              484 fs/jffs2/wbuf.c 			BUG_ON(xr->node != raw);
node              485 fs/jffs2/wbuf.c 			adjust_ref = &xr->node;
node               77 fs/jffs2/xattr.c 	for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node              143 fs/jffs2/xattr.c 	offset = ref_offset(xd->node);
node              144 fs/jffs2/xattr.c 	if (ref_flags(xd->node) == REF_PRISTINE)
node              184 fs/jffs2/xattr.c 	for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node              191 fs/jffs2/xattr.c 		raw->flash_offset = ref_offset(raw) | ((xd->node==raw) ? REF_PRISTINE : REF_NORMAL);
node              212 fs/jffs2/xattr.c 	BUG_ON(ref_flags(xd->node) != REF_PRISTINE);
node              220 fs/jffs2/xattr.c 	ret = jffs2_flash_read(c, ref_offset(xd->node)+sizeof(struct jffs2_raw_xattr),
node              225 fs/jffs2/xattr.c 			      ret, length, readlen, ref_offset(xd->node));
node              235 fs/jffs2/xattr.c 			      ref_offset(xd->node), xd->data_crc, crc);
node              410 fs/jffs2/xattr.c 		if (xd->node == (void *)xd) {
node              454 fs/jffs2/xattr.c 	if (ref_flags(ref->node) != REF_UNCHECKED)
node              456 fs/jffs2/xattr.c 	offset = ref_offset(ref->node);
node              490 fs/jffs2/xattr.c 	for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) {
node              497 fs/jffs2/xattr.c 		raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL);
node              502 fs/jffs2/xattr.c 		  ref->ino, ref->xid, ref_offset(ref->node));
node              794 fs/jffs2/xattr.c 		if (ref_flags(ref->node) != REF_PRISTINE) {
node              796 fs/jffs2/xattr.c 				BUG_ON(ref->node->next_in_ino != (void *)ref);
node              797 fs/jffs2/xattr.c 				ref->node->next_in_ino = NULL;
node              798 fs/jffs2/xattr.c 				jffs2_mark_node_obsolete(c, ref->node);
node              810 fs/jffs2/xattr.c 			raw = ref->node;
node              813 fs/jffs2/xattr.c 				raw->next_in_ino = tmp->node;
node              814 fs/jffs2/xattr.c 				tmp->node = raw;
node              816 fs/jffs2/xattr.c 				raw->next_in_ino = tmp->node->next_in_ino;
node              817 fs/jffs2/xattr.c 				tmp->node->next_in_ino = raw;
node             1226 fs/jffs2/xattr.c 	if (xd->node != raw)
node             1236 fs/jffs2/xattr.c 	old_ofs = ref_offset(xd->node);
node             1247 fs/jffs2/xattr.c 			  xd->xid, xd->version, old_ofs, ref_offset(xd->node));
node             1262 fs/jffs2/xattr.c 	BUG_ON(!ref->node);
node             1264 fs/jffs2/xattr.c 	if (ref->node != raw)
node             1269 fs/jffs2/xattr.c 	old_ofs = ref_offset(ref->node);
node             1270 fs/jffs2/xattr.c 	totlen = ref_totlen(c, c->gcblock, ref->node);
node             1281 fs/jffs2/xattr.c 			  ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node));
node             1304 fs/jffs2/xattr.c 		for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node             1312 fs/jffs2/xattr.c 				| ((xd->node == (void *)raw) ? REF_PRISTINE : REF_NORMAL);
node             1325 fs/jffs2/xattr.c 	if (atomic_read(&xd->refcnt) || xd->node != (void *)xd)
node             1337 fs/jffs2/xattr.c 	if (ref->node != (void *)ref)
node               26 fs/jffs2/xattr.h 	struct jffs2_raw_node_ref *node;
node               48 fs/jffs2/xattr.h 	struct jffs2_raw_node_ref *node;
node              350 fs/kernfs/dir.c 	struct rb_node **node = &kn->parent->dir.children.rb_node;
node              353 fs/kernfs/dir.c 	while (*node) {
node              357 fs/kernfs/dir.c 		pos = rb_to_kn(*node);
node              358 fs/kernfs/dir.c 		parent = *node;
node              361 fs/kernfs/dir.c 			node = &pos->rb.rb_left;
node              363 fs/kernfs/dir.c 			node = &pos->rb.rb_right;
node              369 fs/kernfs/dir.c 	rb_link_node(&kn->rb, parent, node);
node              835 fs/kernfs/dir.c 	struct rb_node *node = parent->dir.children.rb_node;
node              848 fs/kernfs/dir.c 	while (node) {
node              852 fs/kernfs/dir.c 		kn = rb_to_kn(node);
node              855 fs/kernfs/dir.c 			node = node->rb_left;
node              857 fs/kernfs/dir.c 			node = node->rb_right;
node             1621 fs/kernfs/dir.c 		struct rb_node *node = parent->dir.children.rb_node;
node             1622 fs/kernfs/dir.c 		while (node) {
node             1623 fs/kernfs/dir.c 			pos = rb_to_kn(node);
node             1626 fs/kernfs/dir.c 				node = node->rb_left;
node             1628 fs/kernfs/dir.c 				node = node->rb_right;
node             1635 fs/kernfs/dir.c 		struct rb_node *node = rb_next(&pos->rb);
node             1636 fs/kernfs/dir.c 		if (!node)
node             1639 fs/kernfs/dir.c 			pos = rb_to_kn(node);
node             1650 fs/kernfs/dir.c 			struct rb_node *node = rb_next(&pos->rb);
node             1651 fs/kernfs/dir.c 			if (!node)
node             1654 fs/kernfs/dir.c 				pos = rb_to_kn(node);
node              884 fs/kernfs/file.c 	list_for_each_entry(info, &kernfs_root(kn)->supers, node) {
node               71 fs/kernfs/kernfs-internal.h 	struct list_head	node;
node               36 fs/kernfs/mount.c 	struct kernfs_node *node = kernfs_dentry_node(dentry);
node               37 fs/kernfs/mount.c 	struct kernfs_root *root = kernfs_root(node);
node               41 fs/kernfs/mount.c 		return scops->show_path(sf, node, root);
node              305 fs/kernfs/mount.c 	INIT_LIST_HEAD(&info->node);
node              325 fs/kernfs/mount.c 		list_add(&info->node, &info->root->supers);
node              353 fs/kernfs/mount.c 	list_del(&info->node);
node              133 fs/mbcache.c   	struct hlist_bl_node *node;
node              139 fs/mbcache.c   		node = entry->e_hash_list.next;
node              141 fs/mbcache.c   		node = hlist_bl_first(head);
node              142 fs/mbcache.c   	while (node) {
node              143 fs/mbcache.c   		entry = hlist_bl_entry(node, struct mb_cache_entry,
node              149 fs/mbcache.c   		node = node->next;
node              201 fs/mbcache.c   	struct hlist_bl_node *node;
node              207 fs/mbcache.c   	hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
node              229 fs/mbcache.c   	struct hlist_bl_node *node;
node              235 fs/mbcache.c   	hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
node             1115 fs/namespace.c 	struct llist_node *node = llist_del_all(&delayed_mntput_list);
node             1118 fs/namespace.c 	llist_for_each_entry_safe(m, t, node, mnt_llist)
node              152 fs/nfs/blocklayout/blocklayout.c 		container_of(be->be_device, struct pnfs_block_dev, node);
node              590 fs/nfs/blocklayout/blocklayout.c 	struct nfs4_deviceid_node *node;
node              594 fs/nfs/blocklayout/blocklayout.c 	node = nfs4_find_get_deviceid(server, id, cred, gfp_mask);
node              595 fs/nfs/blocklayout/blocklayout.c 	if (!node)
node              598 fs/nfs/blocklayout/blocklayout.c 	if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags) == 0)
node              599 fs/nfs/blocklayout/blocklayout.c 		return node;
node              603 fs/nfs/blocklayout/blocklayout.c 	if (!time_in_range(node->timestamp_unavailable, start, end)) {
node              604 fs/nfs/blocklayout/blocklayout.c 		nfs4_delete_deviceid(node->ld, node->nfs_client, id);
node              102 fs/nfs/blocklayout/blocklayout.h 	struct nfs4_deviceid_node	node;
node               46 fs/nfs/blocklayout/dev.c 		container_of(d, struct pnfs_block_dev, node);
node               49 fs/nfs/blocklayout/dev.c 	kfree_rcu(dev, node.rcu);
node              499 fs/nfs/blocklayout/dev.c 	struct nfs4_deviceid_node *node = NULL;
node              537 fs/nfs/blocklayout/dev.c 	node = &top->node;
node              538 fs/nfs/blocklayout/dev.c 	nfs4_init_deviceid_node(node, server, &pdev->dev_id);
node              540 fs/nfs/blocklayout/dev.c 		nfs4_mark_deviceid_unavailable(node);
node              547 fs/nfs/blocklayout/dev.c 	return node;
node               13 fs/nfs/blocklayout/extent_tree.c ext_node(struct rb_node *node)
node               15 fs/nfs/blocklayout/extent_tree.c 	return rb_entry(node, struct pnfs_block_extent, be_node);
node               21 fs/nfs/blocklayout/extent_tree.c 	struct rb_node *node = rb_first(root);
node               22 fs/nfs/blocklayout/extent_tree.c 	return node ? ext_node(node) : NULL;
node               28 fs/nfs/blocklayout/extent_tree.c 	struct rb_node *node = rb_prev(&be->be_node);
node               29 fs/nfs/blocklayout/extent_tree.c 	return node ? ext_node(node) : NULL;
node               35 fs/nfs/blocklayout/extent_tree.c 	struct rb_node *node = rb_next(&be->be_node);
node               36 fs/nfs/blocklayout/extent_tree.c 	return node ? ext_node(node) : NULL;
node               48 fs/nfs/blocklayout/extent_tree.c 	struct rb_node *node = root->rb_node;
node               51 fs/nfs/blocklayout/extent_tree.c 	while (node) {
node               52 fs/nfs/blocklayout/extent_tree.c 		be = ext_node(node);
node               54 fs/nfs/blocklayout/extent_tree.c 			node = node->rb_left;
node               56 fs/nfs/blocklayout/extent_tree.c 			node = node->rb_right;
node              325 fs/nfs/blocklayout/extent_tree.c 	struct rb_node *node;
node              328 fs/nfs/blocklayout/extent_tree.c 	node = root->rb_node;
node              329 fs/nfs/blocklayout/extent_tree.c 	while (node) {
node              330 fs/nfs/blocklayout/extent_tree.c 		be = ext_node(node);
node              332 fs/nfs/blocklayout/extent_tree.c 			node = node->rb_left;
node              334 fs/nfs/blocklayout/extent_tree.c 			node = node->rb_right;
node              246 fs/nfs/filelayout/filelayout.c filelayout_test_devid_unavailable(struct nfs4_deviceid_node *node)
node              248 fs/nfs/filelayout/filelayout.c 	return filelayout_test_devid_invalid(node) ||
node              249 fs/nfs/filelayout/filelayout.c 		nfs4_test_deviceid_unavailable(node);
node              255 fs/nfs/filelayout/filelayout.c 	struct nfs4_deviceid_node *node = FILELAYOUT_DEVID_NODE(lseg);
node              257 fs/nfs/filelayout/filelayout.c 	return filelayout_test_devid_unavailable(node);
node               96 fs/nfs/filelayout/filelayout.h filelayout_test_devid_invalid(struct nfs4_deviceid_node *node)
node               98 fs/nfs/filelayout/filelayout.h 	return test_bit(NFS_DEVICEID_INVALID, &node->flags);
node              102 fs/nfs/filelayout/filelayout.h filelayout_test_devid_unavailable(struct nfs4_deviceid_node *node);
node              136 fs/nfs/flexfilelayout/flexfilelayout.h FF_LAYOUT_MIRROR_DS(struct nfs4_deviceid_node *node)
node              138 fs/nfs/flexfilelayout/flexfilelayout.h 	return container_of(node, struct nfs4_ff_layout_ds, id_node);
node              326 fs/nfs/flexfilelayout/flexfilelayoutdev.c 		struct nfs4_deviceid_node *node;
node              329 fs/nfs/flexfilelayout/flexfilelayoutdev.c 		node = nfs4_find_get_deviceid(NFS_SERVER(lo->plh_inode),
node              332 fs/nfs/flexfilelayout/flexfilelayoutdev.c 		if (node)
node              333 fs/nfs/flexfilelayout/flexfilelayoutdev.c 			mirror_ds = FF_LAYOUT_MIRROR_DS(node);
node              338 fs/nfs/flexfilelayout/flexfilelayoutdev.c 			nfs4_put_deviceid_node(node);
node              152 fs/nfs/fscache.c 		xkey = rb_entry(parent, struct nfs_fscache_key, node);
node              183 fs/nfs/fscache.c 	rb_link_node(&key->node, parent, p);
node              184 fs/nfs/fscache.c 	rb_insert_color(&key->node, &nfs_fscache_keys);
node              223 fs/nfs/fscache.c 		rb_erase(&nfss->fscache_key->node, &nfs_fscache_keys);
node               22 fs/nfs/fscache.h 	struct rb_node		node;
node              338 fs/nfs/pnfs.h  	struct hlist_node		node;
node              357 fs/nfs/pnfs.h  void nfs4_mark_deviceid_available(struct nfs4_deviceid_node *node);
node              358 fs/nfs/pnfs.h  void nfs4_mark_deviceid_unavailable(struct nfs4_deviceid_node *node);
node              359 fs/nfs/pnfs.h  bool nfs4_test_deviceid_unavailable(struct nfs4_deviceid_node *node);
node              457 fs/nfs/pnfs.h  pnfs_generic_mark_devid_invalid(struct nfs4_deviceid_node *node)
node              459 fs/nfs/pnfs.h  	set_bit(NFS_DEVICEID_INVALID, &node->flags);
node               83 fs/nfs/pnfs_dev.c 	hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node)
node              208 fs/nfs/pnfs_dev.c 	hlist_add_head_rcu(&new->node, &nfs4_deviceid_cache[hash]);
node              238 fs/nfs/pnfs_dev.c 	hlist_del_init_rcu(&d->node);
node              251 fs/nfs/pnfs_dev.c 	INIT_HLIST_NODE(&d->node);
node              287 fs/nfs/pnfs_dev.c nfs4_mark_deviceid_available(struct nfs4_deviceid_node *node)
node              289 fs/nfs/pnfs_dev.c 	if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags)) {
node              290 fs/nfs/pnfs_dev.c 		clear_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags);
node              297 fs/nfs/pnfs_dev.c nfs4_mark_deviceid_unavailable(struct nfs4_deviceid_node *node)
node              299 fs/nfs/pnfs_dev.c 	node->timestamp_unavailable = jiffies;
node              301 fs/nfs/pnfs_dev.c 	set_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags);
node              307 fs/nfs/pnfs_dev.c nfs4_test_deviceid_unavailable(struct nfs4_deviceid_node *node)
node              309 fs/nfs/pnfs_dev.c 	if (test_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags)) {
node              314 fs/nfs/pnfs_dev.c 		if (time_in_range(node->timestamp_unavailable, start, end))
node              316 fs/nfs/pnfs_dev.c 		clear_bit(NFS_DEVICEID_UNAVAILABLE, &node->flags);
node              331 fs/nfs/pnfs_dev.c 	hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[hash], node)
node              333 fs/nfs/pnfs_dev.c 			hlist_del_init_rcu(&d->node);
node              372 fs/nfs/pnfs_dev.c 		hlist_for_each_entry_rcu(d, &nfs4_deviceid_cache[i], node)
node             2657 fs/nfsd/nfs4state.c 	struct rb_node *node = root->rb_node;
node             2660 fs/nfsd/nfs4state.c 	while (node) {
node             2661 fs/nfsd/nfs4state.c 		clp = rb_entry(node, struct nfs4_client, cl_namenode);
node             2664 fs/nfsd/nfs4state.c 			node = node->rb_left;
node             2666 fs/nfsd/nfs4state.c 			node = node->rb_right;
node               73 fs/nilfs2/btree.c static int nilfs_btree_node_get_flags(const struct nilfs_btree_node *node)
node               75 fs/nilfs2/btree.c 	return node->bn_flags;
node               79 fs/nilfs2/btree.c nilfs_btree_node_set_flags(struct nilfs_btree_node *node, int flags)
node               81 fs/nilfs2/btree.c 	node->bn_flags = flags;
node               84 fs/nilfs2/btree.c static int nilfs_btree_node_root(const struct nilfs_btree_node *node)
node               86 fs/nilfs2/btree.c 	return nilfs_btree_node_get_flags(node) & NILFS_BTREE_NODE_ROOT;
node               89 fs/nilfs2/btree.c static int nilfs_btree_node_get_level(const struct nilfs_btree_node *node)
node               91 fs/nilfs2/btree.c 	return node->bn_level;
node               95 fs/nilfs2/btree.c nilfs_btree_node_set_level(struct nilfs_btree_node *node, int level)
node               97 fs/nilfs2/btree.c 	node->bn_level = level;
node              100 fs/nilfs2/btree.c static int nilfs_btree_node_get_nchildren(const struct nilfs_btree_node *node)
node              102 fs/nilfs2/btree.c 	return le16_to_cpu(node->bn_nchildren);
node              106 fs/nilfs2/btree.c nilfs_btree_node_set_nchildren(struct nilfs_btree_node *node, int nchildren)
node              108 fs/nilfs2/btree.c 	node->bn_nchildren = cpu_to_le16(nchildren);
node              122 fs/nilfs2/btree.c nilfs_btree_node_dkeys(const struct nilfs_btree_node *node)
node              124 fs/nilfs2/btree.c 	return (__le64 *)((char *)(node + 1) +
node              125 fs/nilfs2/btree.c 			  (nilfs_btree_node_root(node) ?
node              130 fs/nilfs2/btree.c nilfs_btree_node_dptrs(const struct nilfs_btree_node *node, int ncmax)
node              132 fs/nilfs2/btree.c 	return (__le64 *)(nilfs_btree_node_dkeys(node) + ncmax);
node              136 fs/nilfs2/btree.c nilfs_btree_node_get_key(const struct nilfs_btree_node *node, int index)
node              138 fs/nilfs2/btree.c 	return le64_to_cpu(*(nilfs_btree_node_dkeys(node) + index));
node              142 fs/nilfs2/btree.c nilfs_btree_node_set_key(struct nilfs_btree_node *node, int index, __u64 key)
node              144 fs/nilfs2/btree.c 	*(nilfs_btree_node_dkeys(node) + index) = cpu_to_le64(key);
node              148 fs/nilfs2/btree.c nilfs_btree_node_get_ptr(const struct nilfs_btree_node *node, int index,
node              151 fs/nilfs2/btree.c 	return le64_to_cpu(*(nilfs_btree_node_dptrs(node, ncmax) + index));
node              155 fs/nilfs2/btree.c nilfs_btree_node_set_ptr(struct nilfs_btree_node *node, int index, __u64 ptr,
node              158 fs/nilfs2/btree.c 	*(nilfs_btree_node_dptrs(node, ncmax) + index) = cpu_to_le64(ptr);
node              161 fs/nilfs2/btree.c static void nilfs_btree_node_init(struct nilfs_btree_node *node, int flags,
node              169 fs/nilfs2/btree.c 	nilfs_btree_node_set_flags(node, flags);
node              170 fs/nilfs2/btree.c 	nilfs_btree_node_set_level(node, level);
node              171 fs/nilfs2/btree.c 	nilfs_btree_node_set_nchildren(node, nchildren);
node              173 fs/nilfs2/btree.c 	dkeys = nilfs_btree_node_dkeys(node);
node              174 fs/nilfs2/btree.c 	dptrs = nilfs_btree_node_dptrs(node, ncmax);
node              238 fs/nilfs2/btree.c static void nilfs_btree_node_insert(struct nilfs_btree_node *node, int index,
node              245 fs/nilfs2/btree.c 	dkeys = nilfs_btree_node_dkeys(node);
node              246 fs/nilfs2/btree.c 	dptrs = nilfs_btree_node_dptrs(node, ncmax);
node              247 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              257 fs/nilfs2/btree.c 	nilfs_btree_node_set_nchildren(node, nchildren);
node              261 fs/nilfs2/btree.c static void nilfs_btree_node_delete(struct nilfs_btree_node *node, int index,
node              270 fs/nilfs2/btree.c 	dkeys = nilfs_btree_node_dkeys(node);
node              271 fs/nilfs2/btree.c 	dptrs = nilfs_btree_node_dptrs(node, ncmax);
node              274 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              287 fs/nilfs2/btree.c 	nilfs_btree_node_set_nchildren(node, nchildren);
node              290 fs/nilfs2/btree.c static int nilfs_btree_node_lookup(const struct nilfs_btree_node *node,
node              298 fs/nilfs2/btree.c 	high = nilfs_btree_node_get_nchildren(node) - 1;
node              303 fs/nilfs2/btree.c 		nkey = nilfs_btree_node_get_key(node, index);
node              317 fs/nilfs2/btree.c 	if (nilfs_btree_node_get_level(node) > NILFS_BTREE_LEVEL_NODE_MIN) {
node              338 fs/nilfs2/btree.c static int nilfs_btree_node_broken(const struct nilfs_btree_node *node,
node              345 fs/nilfs2/btree.c 	level = nilfs_btree_node_get_level(node);
node              346 fs/nilfs2/btree.c 	flags = nilfs_btree_node_get_flags(node);
node              347 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              370 fs/nilfs2/btree.c static int nilfs_btree_root_broken(const struct nilfs_btree_node *node,
node              376 fs/nilfs2/btree.c 	level = nilfs_btree_node_get_level(node);
node              377 fs/nilfs2/btree.c 	flags = nilfs_btree_node_get_flags(node);
node              378 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              436 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              439 fs/nilfs2/btree.c 		node = nilfs_btree_get_root(btree);
node              442 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node              445 fs/nilfs2/btree.c 	return node;
node              449 fs/nilfs2/btree.c 				struct nilfs_btree_node *node, int level)
node              451 fs/nilfs2/btree.c 	if (unlikely(nilfs_btree_node_get_level(node) != level)) {
node              456 fs/nilfs2/btree.c 			  nilfs_btree_node_get_level(node), level);
node              463 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;	/* parent node */
node              493 fs/nilfs2/btree.c 			ptr2 = nilfs_btree_node_get_ptr(ra->node, i, ra->ncmax);
node              540 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              545 fs/nilfs2/btree.c 	node = nilfs_btree_get_root(btree);
node              546 fs/nilfs2/btree.c 	level = nilfs_btree_node_get_level(node);
node              547 fs/nilfs2/btree.c 	if (level < minlevel || nilfs_btree_node_get_nchildren(node) <= 0)
node              550 fs/nilfs2/btree.c 	found = nilfs_btree_node_lookup(node, key, &index);
node              551 fs/nilfs2/btree.c 	ptr = nilfs_btree_node_get_ptr(node, index,
node              561 fs/nilfs2/btree.c 			p.node = nilfs_btree_get_node(btree, path, level + 1,
node              572 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node              573 fs/nilfs2/btree.c 		if (nilfs_btree_bad_node(btree, node, level))
node              576 fs/nilfs2/btree.c 			found = nilfs_btree_node_lookup(node, key, &index);
node              580 fs/nilfs2/btree.c 			ptr = nilfs_btree_node_get_ptr(node, index, ncmax);
node              601 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              605 fs/nilfs2/btree.c 	node = nilfs_btree_get_root(btree);
node              606 fs/nilfs2/btree.c 	index = nilfs_btree_node_get_nchildren(node) - 1;
node              609 fs/nilfs2/btree.c 	level = nilfs_btree_node_get_level(node);
node              610 fs/nilfs2/btree.c 	ptr = nilfs_btree_node_get_ptr(node, index,
node              620 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node              621 fs/nilfs2/btree.c 		if (nilfs_btree_bad_node(btree, node, level))
node              623 fs/nilfs2/btree.c 		index = nilfs_btree_node_get_nchildren(node) - 1;
node              624 fs/nilfs2/btree.c 		ptr = nilfs_btree_node_get_ptr(node, index, ncmax);
node              629 fs/nilfs2/btree.c 		*keyp = nilfs_btree_node_get_key(node, index);
node              650 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              658 fs/nilfs2/btree.c 			node = nilfs_btree_get_root(btree);
node              660 fs/nilfs2/btree.c 			node = nilfs_btree_get_nonroot_node(path, level);
node              663 fs/nilfs2/btree.c 		if (index < nilfs_btree_node_get_nchildren(node)) {
node              665 fs/nilfs2/btree.c 			*nextkey = nilfs_btree_node_get_key(node, index);
node              696 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              724 fs/nilfs2/btree.c 	node = nilfs_btree_get_node(btree, path, level, &ncmax);
node              727 fs/nilfs2/btree.c 		while (index < nilfs_btree_node_get_nchildren(node)) {
node              728 fs/nilfs2/btree.c 			if (nilfs_btree_node_get_key(node, index) !=
node              731 fs/nilfs2/btree.c 			ptr2 = nilfs_btree_node_get_ptr(node, index, ncmax);
node              747 fs/nilfs2/btree.c 		p.node = nilfs_btree_get_node(btree, path, level + 1, &p.ncmax);
node              750 fs/nilfs2/btree.c 		if (p.index >= nilfs_btree_node_get_nchildren(p.node) ||
node              751 fs/nilfs2/btree.c 		    nilfs_btree_node_get_key(p.node, p.index) != key + cnt)
node              753 fs/nilfs2/btree.c 		ptr2 = nilfs_btree_node_get_ptr(p.node, p.index, p.ncmax);
node              763 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node              802 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node              806 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node              808 fs/nilfs2/btree.c 		nilfs_btree_node_insert(node, path[level].bp_index,
node              815 fs/nilfs2/btree.c 						nilfs_btree_node_get_key(node,
node              818 fs/nilfs2/btree.c 		node = nilfs_btree_get_root(btree);
node              819 fs/nilfs2/btree.c 		nilfs_btree_node_insert(node, path[level].bp_index,
node              829 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *left;
node              832 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node              834 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              846 fs/nilfs2/btree.c 	nilfs_btree_node_move_left(left, node, n, ncblk, ncblk);
node              854 fs/nilfs2/btree.c 				nilfs_btree_node_get_key(node, 0));
node              875 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *right;
node              878 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node              880 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              892 fs/nilfs2/btree.c 	nilfs_btree_node_move_right(node, right, n, ncblk, ncblk);
node              908 fs/nilfs2/btree.c 		path[level].bp_index -= nilfs_btree_node_get_nchildren(node);
node              922 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *right;
node              925 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node              927 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node              937 fs/nilfs2/btree.c 	nilfs_btree_node_move_right(node, right, n, ncblk, ncblk);
node              945 fs/nilfs2/btree.c 		path[level].bp_index -= nilfs_btree_node_get_nchildren(node);
node             1000 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             1009 fs/nilfs2/btree.c 		node = nilfs_btree_get_node(btree, path, level, &ncmax);
node             1010 fs/nilfs2/btree.c 		return nilfs_btree_node_get_ptr(node,
node             1018 fs/nilfs2/btree.c 		node = nilfs_btree_get_node(btree, path, level, &ncmax);
node             1019 fs/nilfs2/btree.c 		return nilfs_btree_node_get_ptr(node, path[level].bp_index,
node             1052 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *parent, *sib;
node             1076 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node             1077 fs/nilfs2/btree.c 		if (nilfs_btree_node_get_nchildren(node) < ncblk) {
node             1144 fs/nilfs2/btree.c 	node = nilfs_btree_get_root(btree);
node             1145 fs/nilfs2/btree.c 	if (nilfs_btree_node_get_nchildren(node) <
node             1252 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             1256 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node             1258 fs/nilfs2/btree.c 		nilfs_btree_node_delete(node, path[level].bp_index,
node             1264 fs/nilfs2/btree.c 				nilfs_btree_node_get_key(node, 0));
node             1266 fs/nilfs2/btree.c 		node = nilfs_btree_get_root(btree);
node             1267 fs/nilfs2/btree.c 		nilfs_btree_node_delete(node, path[level].bp_index,
node             1277 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *left;
node             1282 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node             1284 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node             1290 fs/nilfs2/btree.c 	nilfs_btree_node_move_right(left, node, n, ncblk, ncblk);
node             1298 fs/nilfs2/btree.c 				nilfs_btree_node_get_key(node, 0));
node             1309 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *right;
node             1314 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node             1316 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node             1322 fs/nilfs2/btree.c 	nilfs_btree_node_move_left(node, right, n, ncblk, ncblk);
node             1342 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *left;
node             1347 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node             1351 fs/nilfs2/btree.c 	n = nilfs_btree_node_get_nchildren(node);
node             1353 fs/nilfs2/btree.c 	nilfs_btree_node_move_left(left, node, n, ncblk, ncblk);
node             1368 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *right;
node             1373 fs/nilfs2/btree.c 	node = nilfs_btree_get_nonroot_node(path, level);
node             1379 fs/nilfs2/btree.c 	nilfs_btree_node_move_left(node, right, n, ncblk, ncblk);
node             1426 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *parent, *sib;
node             1438 fs/nilfs2/btree.c 		node = nilfs_btree_get_nonroot_node(path, level);
node             1440 fs/nilfs2/btree.c 			nilfs_btree_node_get_ptr(node, dindex, ncblk);
node             1446 fs/nilfs2/btree.c 		if (nilfs_btree_node_get_nchildren(node) > ncmin) {
node             1507 fs/nilfs2/btree.c 			if (nilfs_btree_node_get_nchildren(node) - 1 <=
node             1527 fs/nilfs2/btree.c 	node = nilfs_btree_get_root(btree);
node             1529 fs/nilfs2/btree.c 		nilfs_btree_node_get_ptr(node, dindex,
node             1640 fs/nilfs2/btree.c 	struct nilfs_btree_node *root, *node;
node             1649 fs/nilfs2/btree.c 		node = root;
node             1660 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)bh->b_data;
node             1666 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node             1667 fs/nilfs2/btree.c 	maxkey = nilfs_btree_node_get_key(node, nchildren - 1);
node             1669 fs/nilfs2/btree.c 		nilfs_btree_node_get_key(node, nchildren - 2) : 0;
node             1680 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *root;
node             1690 fs/nilfs2/btree.c 		node = root;
node             1701 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)bh->b_data;
node             1705 fs/nilfs2/btree.c 		node = NULL;
node             1709 fs/nilfs2/btree.c 	nchildren = nilfs_btree_node_get_nchildren(node);
node             1712 fs/nilfs2/btree.c 	dkeys = nilfs_btree_node_dkeys(node);
node             1713 fs/nilfs2/btree.c 	dptrs = nilfs_btree_node_dptrs(node, ncmax);
node             1787 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             1807 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)bh->b_data;
node             1809 fs/nilfs2/btree.c 		nilfs_btree_node_init(node, 0, 1, n, ncblk, keys, ptrs);
node             1810 fs/nilfs2/btree.c 		nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk);
node             1819 fs/nilfs2/btree.c 		node = nilfs_btree_get_root(btree);
node             1821 fs/nilfs2/btree.c 		nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 2, 1,
node             1828 fs/nilfs2/btree.c 		node = nilfs_btree_get_root(btree);
node             1829 fs/nilfs2/btree.c 		nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 1, n,
node             1832 fs/nilfs2/btree.c 		nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr,
node             2055 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             2066 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)bh->b_data;
node             2067 fs/nilfs2/btree.c 		key = nilfs_btree_node_get_key(node, 0);
node             2068 fs/nilfs2/btree.c 		level = nilfs_btree_node_get_level(node);
node             2106 fs/nilfs2/btree.c 	struct nilfs_btree_node *node, *cnode;
node             2111 fs/nilfs2/btree.c 	node = (struct nilfs_btree_node *)bh->b_data;
node             2112 fs/nilfs2/btree.c 	key = nilfs_btree_node_get_key(node, 0);
node             2113 fs/nilfs2/btree.c 	level = nilfs_btree_node_get_level(node);
node             2250 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             2259 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)(*bh)->b_data;
node             2260 fs/nilfs2/btree.c 		key = nilfs_btree_node_get_key(node, 0);
node             2261 fs/nilfs2/btree.c 		level = nilfs_btree_node_get_level(node);
node             2288 fs/nilfs2/btree.c 	struct nilfs_btree_node *node;
node             2298 fs/nilfs2/btree.c 		node = (struct nilfs_btree_node *)(*bh)->b_data;
node             2299 fs/nilfs2/btree.c 		key = nilfs_btree_node_get_key(node, 0);
node              241 fs/notify/fsnotify.c 	struct hlist_node *node = NULL;
node              245 fs/notify/fsnotify.c 		node = srcu_dereference(conn->list.first, &fsnotify_mark_srcu);
node              247 fs/notify/fsnotify.c 	return hlist_entry_safe(node, struct fsnotify_mark, obj_list);
node              252 fs/notify/fsnotify.c 	struct hlist_node *node = NULL;
node              255 fs/notify/fsnotify.c 		node = srcu_dereference(mark->obj_list.next,
node              258 fs/notify/fsnotify.c 	return hlist_entry_safe(node, struct fsnotify_mark, obj_list);
node              578 fs/ocfs2/alloc.c 	struct ocfs2_path_item *node;
node              584 fs/ocfs2/alloc.c 		node = &path->p_node[i];
node              586 fs/ocfs2/alloc.c 		brelse(node->bh);
node              587 fs/ocfs2/alloc.c 		node->bh = NULL;
node              588 fs/ocfs2/alloc.c 		node->el = NULL;
node              741 fs/ocfs2/cluster/heartbeat.c 				struct o2nm_node *node,
node              748 fs/ocfs2/cluster/heartbeat.c 		(f->hc_func)(node, idx, f->hc_data);
node              794 fs/ocfs2/cluster/heartbeat.c 				  struct o2nm_node *node,
node              799 fs/ocfs2/cluster/heartbeat.c 	BUG_ON((!node) && (type != O2HB_NODE_DOWN_CB));
node              802 fs/ocfs2/cluster/heartbeat.c 	event->hn_node = node;
node              815 fs/ocfs2/cluster/heartbeat.c 	struct o2nm_node *node;
node              818 fs/ocfs2/cluster/heartbeat.c 	node = o2nm_get_node_by_num(slot->ds_node_num);
node              819 fs/ocfs2/cluster/heartbeat.c 	if (!node)
node              832 fs/ocfs2/cluster/heartbeat.c 			o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
node              842 fs/ocfs2/cluster/heartbeat.c 	o2nm_node_put(node);
node              894 fs/ocfs2/cluster/heartbeat.c 	struct o2nm_node *node;
node              908 fs/ocfs2/cluster/heartbeat.c 	node = o2nm_get_node_by_num(slot->ds_node_num);
node              909 fs/ocfs2/cluster/heartbeat.c 	if (!node) {
node              990 fs/ocfs2/cluster/heartbeat.c 			o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node,
node             1043 fs/ocfs2/cluster/heartbeat.c 					      node, slot->ds_node_num);
node             1065 fs/ocfs2/cluster/heartbeat.c 	if (node)
node             1066 fs/ocfs2/cluster/heartbeat.c 		o2nm_node_put(node);
node               34 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node               40 fs/ocfs2/cluster/nodemanager.c 	node = o2nm_single_cluster->cl_nodes[node_num];
node               41 fs/ocfs2/cluster/nodemanager.c 	if (node)
node               42 fs/ocfs2/cluster/nodemanager.c 		config_item_get(&node->nd_item);
node               45 fs/ocfs2/cluster/nodemanager.c 	return node;
node               73 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node, *ret = NULL;
node               79 fs/ocfs2/cluster/nodemanager.c 		node = rb_entry(parent, struct o2nm_node, nd_ip_node);
node               81 fs/ocfs2/cluster/nodemanager.c 		cmp = memcmp(&ip_needle, &node->nd_ipv4_address,
node               88 fs/ocfs2/cluster/nodemanager.c 			ret = node;
node              103 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node              110 fs/ocfs2/cluster/nodemanager.c 	node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL);
node              111 fs/ocfs2/cluster/nodemanager.c 	if (node)
node              112 fs/ocfs2/cluster/nodemanager.c 		config_item_get(&node->nd_item);
node              116 fs/ocfs2/cluster/nodemanager.c 	return node;
node              120 fs/ocfs2/cluster/nodemanager.c void o2nm_node_put(struct o2nm_node *node)
node              122 fs/ocfs2/cluster/nodemanager.c 	config_item_put(&node->nd_item);
node              126 fs/ocfs2/cluster/nodemanager.c void o2nm_node_get(struct o2nm_node *node)
node              128 fs/ocfs2/cluster/nodemanager.c 	config_item_get(&node->nd_item);
node              160 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              161 fs/ocfs2/cluster/nodemanager.c 	kfree(node);
node              169 fs/ocfs2/cluster/nodemanager.c static struct o2nm_cluster *to_o2nm_cluster_from_node(struct o2nm_node *node)
node              173 fs/ocfs2/cluster/nodemanager.c 	if (node->nd_item.ci_parent)
node              174 fs/ocfs2/cluster/nodemanager.c 		return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent);
node              188 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              205 fs/ocfs2/cluster/nodemanager.c 	if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node              206 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node              210 fs/ocfs2/cluster/nodemanager.c 	cluster = to_o2nm_cluster_from_node(node);
node              220 fs/ocfs2/cluster/nodemanager.c 			&node->nd_set_attributes))
node              223 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_nodes[tmp] = node;
node              224 fs/ocfs2/cluster/nodemanager.c 		node->nd_num = tmp;
node              243 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              256 fs/ocfs2/cluster/nodemanager.c 	if (test_and_set_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node              258 fs/ocfs2/cluster/nodemanager.c 	node->nd_ipv4_port = htons(tmp);
node              272 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              291 fs/ocfs2/cluster/nodemanager.c 	cluster = to_o2nm_cluster_from_node(node);
node              302 fs/ocfs2/cluster/nodemanager.c 			&node->nd_set_attributes))
node              305 fs/ocfs2/cluster/nodemanager.c 		rb_link_node(&node->nd_ip_node, parent, p);
node              306 fs/ocfs2/cluster/nodemanager.c 		rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node              314 fs/ocfs2/cluster/nodemanager.c 	memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr));
node              327 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              341 fs/ocfs2/cluster/nodemanager.c 	if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node              342 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) ||
node              343 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node              347 fs/ocfs2/cluster/nodemanager.c 	cluster = to_o2nm_cluster_from_node(node);
node              356 fs/ocfs2/cluster/nodemanager.c 	    cluster->cl_local_node != node->nd_num) {
node              363 fs/ocfs2/cluster/nodemanager.c 		ret = o2net_start_listening(node);
node              369 fs/ocfs2/cluster/nodemanager.c 	    cluster->cl_local_node == node->nd_num) {
node              370 fs/ocfs2/cluster/nodemanager.c 		o2net_stop_listening(node);
node              374 fs/ocfs2/cluster/nodemanager.c 	node->nd_local = tmp;
node              375 fs/ocfs2/cluster/nodemanager.c 	if (node->nd_local) {
node              377 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_local_node = node->nd_num;
node              586 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node              591 fs/ocfs2/cluster/nodemanager.c 	node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL);
node              592 fs/ocfs2/cluster/nodemanager.c 	if (node == NULL)
node              595 fs/ocfs2/cluster/nodemanager.c 	strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */
node              596 fs/ocfs2/cluster/nodemanager.c 	config_item_init_type_name(&node->nd_item, name, &o2nm_node_type);
node              597 fs/ocfs2/cluster/nodemanager.c 	spin_lock_init(&node->nd_lock);
node              601 fs/ocfs2/cluster/nodemanager.c 	return &node->nd_item;
node              607 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              610 fs/ocfs2/cluster/nodemanager.c 	if (cluster->cl_nodes[node->nd_num] == node) {
node              611 fs/ocfs2/cluster/nodemanager.c 		o2net_disconnect_node(node);
node              614 fs/ocfs2/cluster/nodemanager.c 		    (cluster->cl_local_node == node->nd_num)) {
node              617 fs/ocfs2/cluster/nodemanager.c 			o2net_stop_listening(node);
node              626 fs/ocfs2/cluster/nodemanager.c 	if (node->nd_ipv4_address)
node              627 fs/ocfs2/cluster/nodemanager.c 		rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node              630 fs/ocfs2/cluster/nodemanager.c 	if (cluster->cl_nodes[node->nd_num] == node) {
node              631 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_nodes[node->nd_num] = NULL;
node              632 fs/ocfs2/cluster/nodemanager.c 		clear_bit(node->nd_num, cluster->cl_nodes_bitmap);
node              637 fs/ocfs2/cluster/nodemanager.c 	     config_item_name(&node->nd_item));
node               65 fs/ocfs2/cluster/nodemanager.h void o2nm_node_get(struct o2nm_node *node);
node               66 fs/ocfs2/cluster/nodemanager.h void o2nm_node_put(struct o2nm_node *node);
node              164 fs/ocfs2/cluster/quorum.c static void o2quo_set_hold(struct o2quo_state *qs, u8 node)
node              168 fs/ocfs2/cluster/quorum.c 	if (!test_and_set_bit(node, qs->qs_hold_bm)) {
node              171 fs/ocfs2/cluster/quorum.c 			        "node %u\n", node);
node              172 fs/ocfs2/cluster/quorum.c 		mlog(0, "node %u, %d total\n", node, qs->qs_holds);
node              176 fs/ocfs2/cluster/quorum.c static void o2quo_clear_hold(struct o2quo_state *qs, u8 node)
node              180 fs/ocfs2/cluster/quorum.c 	if (test_and_clear_bit(node, qs->qs_hold_bm)) {
node              181 fs/ocfs2/cluster/quorum.c 		mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1);
node              189 fs/ocfs2/cluster/quorum.c 				node, qs->qs_holds);
node              197 fs/ocfs2/cluster/quorum.c void o2quo_hb_up(u8 node)
node              205 fs/ocfs2/cluster/quorum.c 		        "node %u\n", node);
node              206 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node              207 fs/ocfs2/cluster/quorum.c 	set_bit(node, qs->qs_hb_bm);
node              209 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node              211 fs/ocfs2/cluster/quorum.c 	if (!test_bit(node, qs->qs_conn_bm))
node              212 fs/ocfs2/cluster/quorum.c 		o2quo_set_hold(qs, node);
node              214 fs/ocfs2/cluster/quorum.c 		o2quo_clear_hold(qs, node);
node              221 fs/ocfs2/cluster/quorum.c void o2quo_hb_down(u8 node)
node              230 fs/ocfs2/cluster/quorum.c 			node, qs->qs_heartbeating);
node              231 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node              232 fs/ocfs2/cluster/quorum.c 	clear_bit(node, qs->qs_hb_bm);
node              234 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node              236 fs/ocfs2/cluster/quorum.c 	o2quo_clear_hold(qs, node);
node              246 fs/ocfs2/cluster/quorum.c void o2quo_hb_still_up(u8 node)
node              252 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u\n", node);
node              255 fs/ocfs2/cluster/quorum.c 	o2quo_clear_hold(qs, node);
node              265 fs/ocfs2/cluster/quorum.c void o2quo_conn_up(u8 node)
node              273 fs/ocfs2/cluster/quorum.c 		        "node %u\n", node);
node              274 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node);
node              275 fs/ocfs2/cluster/quorum.c 	set_bit(node, qs->qs_conn_bm);
node              277 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node              279 fs/ocfs2/cluster/quorum.c 	if (!test_bit(node, qs->qs_hb_bm))
node              280 fs/ocfs2/cluster/quorum.c 		o2quo_set_hold(qs, node);
node              282 fs/ocfs2/cluster/quorum.c 		o2quo_clear_hold(qs, node);
node              291 fs/ocfs2/cluster/quorum.c void o2quo_conn_err(u8 node)
node              297 fs/ocfs2/cluster/quorum.c 	if (test_bit(node, qs->qs_conn_bm)) {
node              301 fs/ocfs2/cluster/quorum.c 				node, qs->qs_connected);
node              303 fs/ocfs2/cluster/quorum.c 		clear_bit(node, qs->qs_conn_bm);
node              305 fs/ocfs2/cluster/quorum.c 		if (test_bit(node, qs->qs_hb_bm))
node              306 fs/ocfs2/cluster/quorum.c 			o2quo_set_hold(qs, node);
node              309 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node               14 fs/ocfs2/cluster/quorum.h void o2quo_hb_up(u8 node);
node               15 fs/ocfs2/cluster/quorum.h void o2quo_hb_down(u8 node);
node               16 fs/ocfs2/cluster/quorum.h void o2quo_hb_still_up(u8 node);
node               17 fs/ocfs2/cluster/quorum.h void o2quo_conn_up(u8 node);
node               18 fs/ocfs2/cluster/quorum.h void o2quo_conn_err(u8 node);
node              135 fs/ocfs2/cluster/tcp.c 			   u32 msgkey, struct task_struct *task, u8 node)
node              141 fs/ocfs2/cluster/tcp.c 	nst->st_node = node;
node              412 fs/ocfs2/cluster/tcp.c static struct o2net_sock_container *sc_alloc(struct o2nm_node *node)
node              424 fs/ocfs2/cluster/tcp.c 	o2nm_node_get(node);
node              425 fs/ocfs2/cluster/tcp.c 	sc->sc_node = node;
node              428 fs/ocfs2/cluster/tcp.c 	status = o2nm_depend_item(&node->nd_item);
node              431 fs/ocfs2/cluster/tcp.c 		o2nm_node_put(node);
node              998 fs/ocfs2/cluster/tcp.c 	int node, ret;
node             1003 fs/ocfs2/cluster/tcp.c 	for (node = 0; node < O2NM_MAX_NODES; ++node) {
node             1004 fs/ocfs2/cluster/tcp.c 		if (!o2net_tx_can_proceed(o2net_nn_from_num(node), &sc, &ret))
node             1007 fs/ocfs2/cluster/tcp.c 			set_bit(node, map);
node             1568 fs/ocfs2/cluster/tcp.c 	struct o2nm_node *node = NULL, *mynode = NULL;
node             1587 fs/ocfs2/cluster/tcp.c 	node = o2nm_get_node_by_num(o2net_num_from_nn(nn));
node             1588 fs/ocfs2/cluster/tcp.c 	if (node == NULL)
node             1613 fs/ocfs2/cluster/tcp.c 	sc = sc_alloc(node);
node             1661 fs/ocfs2/cluster/tcp.c 	remoteaddr.sin_addr.s_addr = node->nd_ipv4_address;
node             1662 fs/ocfs2/cluster/tcp.c 	remoteaddr.sin_port = node->nd_ipv4_port;
node             1681 fs/ocfs2/cluster/tcp.c 	if (node)
node             1682 fs/ocfs2/cluster/tcp.c 		o2nm_node_put(node);
node             1719 fs/ocfs2/cluster/tcp.c void o2net_disconnect_node(struct o2nm_node *node)
node             1721 fs/ocfs2/cluster/tcp.c 	struct o2net_node *nn = o2net_nn_from_num(node->nd_num);
node             1737 fs/ocfs2/cluster/tcp.c static void o2net_hb_node_down_cb(struct o2nm_node *node, int node_num,
node             1742 fs/ocfs2/cluster/tcp.c 	if (!node)
node             1746 fs/ocfs2/cluster/tcp.c 		o2net_disconnect_node(node);
node             1751 fs/ocfs2/cluster/tcp.c static void o2net_hb_node_up_cb(struct o2nm_node *node, int node_num,
node             1758 fs/ocfs2/cluster/tcp.c 	BUG_ON(!node);
node             1809 fs/ocfs2/cluster/tcp.c 	struct o2nm_node *node = NULL;
node             1855 fs/ocfs2/cluster/tcp.c 	node = o2nm_get_node_by_ip(sin.sin_addr.s_addr);
node             1856 fs/ocfs2/cluster/tcp.c 	if (node == NULL) {
node             1864 fs/ocfs2/cluster/tcp.c 	if (o2nm_this_node() >= node->nd_num) {
node             1873 fs/ocfs2/cluster/tcp.c 					node->nd_name,
node             1874 fs/ocfs2/cluster/tcp.c 					node->nd_num, &sin.sin_addr.s_addr,
node             1882 fs/ocfs2/cluster/tcp.c 	if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) {
node             1885 fs/ocfs2/cluster/tcp.c 		     node->nd_name, &sin.sin_addr.s_addr,
node             1891 fs/ocfs2/cluster/tcp.c 	nn = o2net_nn_from_num(node->nd_num);
node             1902 fs/ocfs2/cluster/tcp.c 		       node->nd_name, &sin.sin_addr.s_addr,
node             1907 fs/ocfs2/cluster/tcp.c 	sc = sc_alloc(node);
node             1930 fs/ocfs2/cluster/tcp.c 	if (node)
node             1931 fs/ocfs2/cluster/tcp.c 		o2nm_node_put(node);
node             2064 fs/ocfs2/cluster/tcp.c int o2net_start_listening(struct o2nm_node *node)
node             2078 fs/ocfs2/cluster/tcp.c 	ret = o2net_open_listening_sock(node->nd_ipv4_address,
node             2079 fs/ocfs2/cluster/tcp.c 					node->nd_ipv4_port);
node             2084 fs/ocfs2/cluster/tcp.c 		o2quo_conn_up(node->nd_num);
node             2091 fs/ocfs2/cluster/tcp.c void o2net_stop_listening(struct o2nm_node *node)
node             2106 fs/ocfs2/cluster/tcp.c 		struct o2nm_node *node = o2nm_get_node_by_num(i);
node             2107 fs/ocfs2/cluster/tcp.c 		if (node) {
node             2108 fs/ocfs2/cluster/tcp.c 			o2net_disconnect_node(node);
node             2109 fs/ocfs2/cluster/tcp.c 			o2nm_node_put(node);
node             2121 fs/ocfs2/cluster/tcp.c 	o2quo_conn_err(node->nd_num);
node              100 fs/ocfs2/cluster/tcp.h int o2net_start_listening(struct o2nm_node *node);
node              101 fs/ocfs2/cluster/tcp.h void o2net_stop_listening(struct o2nm_node *node);
node              102 fs/ocfs2/cluster/tcp.h void o2net_disconnect_node(struct o2nm_node *node);
node              190 fs/ocfs2/dlm/dlmast.c 				  lock->ml.node == dlm->node_num ? "master" :
node              219 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node != dlm->node_num);
node              239 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node == dlm->node_num);
node              255 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node != dlm->node_num);
node              281 fs/ocfs2/dlm/dlmast.c 	u8 node;
node              295 fs/ocfs2/dlm/dlmast.c 	node = past->node_idx;
node              323 fs/ocfs2/dlm/dlmast.c 		     locklen, name, node);
node              334 fs/ocfs2/dlm/dlmast.c 		     locklen, name, node);
node              383 fs/ocfs2/dlm/dlmast.c 	     locklen, name, node);
node              444 fs/ocfs2/dlm/dlmast.c 	     res->lockname.len, res->lockname.name, lock->ml.node, msg_type,
node              465 fs/ocfs2/dlm/dlmast.c 				     lock->ml.node, &status);
node              469 fs/ocfs2/dlm/dlmast.c 		     lock->ml.node);
node              473 fs/ocfs2/dlm/dlmast.c 			     "node is dead!\n", lock->ml.node);
node              477 fs/ocfs2/dlm/dlmast.c 			     "DLM_MIGRATING!\n", lock->ml.node);
node              481 fs/ocfs2/dlm/dlmast.c 			     lock->ml.node, status);
node              253 fs/ocfs2/dlm/dlmcommon.h 					  u8 node)
node              257 fs/ocfs2/dlm/dlmcommon.h 	dlm->joining_node = node;
node              337 fs/ocfs2/dlm/dlmcommon.h 	u8 node;
node              819 fs/ocfs2/dlm/dlmcommon.h struct dlm_lock * dlm_new_lock(int type, u8 node, u64 cookie,
node              852 fs/ocfs2/dlm/dlmcommon.h int dlm_is_node_dead(struct dlm_ctxt *dlm, u8 node);
node              853 fs/ocfs2/dlm/dlmcommon.h void dlm_wait_for_node_death(struct dlm_ctxt *dlm, u8 node, int timeout);
node              854 fs/ocfs2/dlm/dlmcommon.h void dlm_wait_for_node_recovery(struct dlm_ctxt *dlm, u8 node, int timeout);
node              952 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data);
node              953 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data);
node              194 fs/ocfs2/dlm/dlmconvert.c 	if (lock->ml.node == dlm->node_num)
node              486 fs/ocfs2/dlm/dlmconvert.c 		    tmp_lock->ml.node == cnv->node_idx) {
node               66 fs/ocfs2/dlm/dlmdebug.c 	       lock->ml.type, lock->ml.convert_type, lock->ml.node,
node              474 fs/ocfs2/dlm/dlmdebug.c 		       lock->ml.node,
node              665 fs/ocfs2/dlm/dlmdebug.c 	struct dlm_reco_node_data *node;
node              791 fs/ocfs2/dlm/dlmdebug.c 	list_for_each_entry(node, &dlm->reco.node_data, list) {
node              792 fs/ocfs2/dlm/dlmdebug.c 		switch (node->state) {
node              819 fs/ocfs2/dlm/dlmdebug.c 				node->node_num, state);
node              490 fs/ocfs2/dlm/dlmdomain.c 	unsigned int node;
node              496 fs/ocfs2/dlm/dlmdomain.c 	node = exit_msg->node_idx;
node              497 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "%s: Node %u sent a begin exit domain message\n", dlm->name, node);
node              500 fs/ocfs2/dlm/dlmdomain.c 	set_bit(node, dlm->exit_domain_map);
node              533 fs/ocfs2/dlm/dlmdomain.c 	int node = -1, num = 0;
node              538 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node              539 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node              540 fs/ocfs2/dlm/dlmdomain.c 		printk("%d ", node);
node              550 fs/ocfs2/dlm/dlmdomain.c 	unsigned int node;
node              558 fs/ocfs2/dlm/dlmdomain.c 	node = exit_msg->node_idx;
node              561 fs/ocfs2/dlm/dlmdomain.c 	clear_bit(node, dlm->domain_map);
node              562 fs/ocfs2/dlm/dlmdomain.c 	clear_bit(node, dlm->exit_domain_map);
node              563 fs/ocfs2/dlm/dlmdomain.c 	printk(KERN_NOTICE "o2dlm: Node %u leaves domain %s ", node, dlm->name);
node              567 fs/ocfs2/dlm/dlmdomain.c 	dlm_hb_event_notify_attached(dlm, node, 0);
node              577 fs/ocfs2/dlm/dlmdomain.c 				    unsigned int node)
node              583 fs/ocfs2/dlm/dlmdomain.c 	     msg_type, node);
node              589 fs/ocfs2/dlm/dlmdomain.c 				    sizeof(leave_msg), node, NULL);
node              592 fs/ocfs2/dlm/dlmdomain.c 		     "to node %u on domain %s\n", status, msg_type, node,
node              600 fs/ocfs2/dlm/dlmdomain.c 	int node = -1;
node              614 fs/ocfs2/dlm/dlmdomain.c 		node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node + 1);
node              615 fs/ocfs2/dlm/dlmdomain.c 		if (node >= O2NM_MAX_NODES)
node              617 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node              621 fs/ocfs2/dlm/dlmdomain.c 		dlm_send_one_domain_exit(dlm, DLM_BEGIN_EXIT_DOMAIN_MSG, node);
node              629 fs/ocfs2/dlm/dlmdomain.c 	int node, clear_node, status;
node              639 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node              652 fs/ocfs2/dlm/dlmdomain.c 						  node);
node              657 fs/ocfs2/dlm/dlmdomain.c 			     "to node %d\n", status, node);
node              670 fs/ocfs2/dlm/dlmdomain.c 			clear_bit(node, dlm->domain_map);
node              728 fs/ocfs2/dlm/dlmdomain.c static int dlm_query_join_proto_check(char *proto_type, int node,
node              739 fs/ocfs2/dlm/dlmdomain.c 		     node, proto_type,
node              749 fs/ocfs2/dlm/dlmdomain.c 		     node, proto_type,
node             1219 fs/ocfs2/dlm/dlmdomain.c 	struct o2nm_node *node;
node             1233 fs/ocfs2/dlm/dlmdomain.c 		node = o2nm_get_node_by_num(i);
node             1234 fs/ocfs2/dlm/dlmdomain.c 		if (!node)
node             1236 fs/ocfs2/dlm/dlmdomain.c 		qn->qn_nodes[count].ni_nodenum = node->nd_num;
node             1237 fs/ocfs2/dlm/dlmdomain.c 		qn->qn_nodes[count].ni_ipv4_port = node->nd_ipv4_port;
node             1238 fs/ocfs2/dlm/dlmdomain.c 		qn->qn_nodes[count].ni_ipv4_address = node->nd_ipv4_address;
node             1239 fs/ocfs2/dlm/dlmdomain.c 		mlog(0, "Node %3d, %pI4:%u\n", node->nd_num,
node             1240 fs/ocfs2/dlm/dlmdomain.c 		     &(node->nd_ipv4_address), ntohs(node->nd_ipv4_port));
node             1242 fs/ocfs2/dlm/dlmdomain.c 		o2nm_node_put(node);
node             1353 fs/ocfs2/dlm/dlmdomain.c 				    unsigned int node)
node             1364 fs/ocfs2/dlm/dlmdomain.c 				    &cancel_msg, sizeof(cancel_msg), node,
node             1369 fs/ocfs2/dlm/dlmdomain.c 		     node);
node             1383 fs/ocfs2/dlm/dlmdomain.c 	int node;
node             1394 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node             1395 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node             1396 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node             1397 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node             1400 fs/ocfs2/dlm/dlmdomain.c 		tmpstat = dlm_send_one_join_cancel(dlm, node);
node             1403 fs/ocfs2/dlm/dlmdomain.c 			     "node %d\n", tmpstat, node);
node             1415 fs/ocfs2/dlm/dlmdomain.c 			    int node,
node             1423 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "querying node %d\n", node);
node             1436 fs/ocfs2/dlm/dlmdomain.c 				    sizeof(join_msg), node, &join_resp);
node             1440 fs/ocfs2/dlm/dlmdomain.c 		     node);
node             1468 fs/ocfs2/dlm/dlmdomain.c 			     node);
node             1478 fs/ocfs2/dlm/dlmdomain.c 			     node,
node             1487 fs/ocfs2/dlm/dlmdomain.c 			     packet.code, node);
node             1494 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "status %d, node %d response is %d\n", status, node,
node             1502 fs/ocfs2/dlm/dlmdomain.c 				    unsigned int node)
node             1508 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "Sending join assert to node %u\n", node);
node             1516 fs/ocfs2/dlm/dlmdomain.c 				    &assert_msg, sizeof(assert_msg), node,
node             1521 fs/ocfs2/dlm/dlmdomain.c 		     node);
node             1531 fs/ocfs2/dlm/dlmdomain.c 	int status, node, live;
node             1534 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node             1535 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node             1536 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node             1537 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node             1544 fs/ocfs2/dlm/dlmdomain.c 			status = dlm_send_one_join_assert(dlm, node);
node             1547 fs/ocfs2/dlm/dlmdomain.c 			live = test_bit(node, dlm->live_nodes_map);
node             1552 fs/ocfs2/dlm/dlmdomain.c 				     "join on node %d\n", status, node);
node             1593 fs/ocfs2/dlm/dlmdomain.c 	int status = 0, tmpstat, node;
node             1618 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node             1619 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(ctxt->live_map, O2NM_MAX_NODES,
node             1620 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node             1621 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node             1624 fs/ocfs2/dlm/dlmdomain.c 		status = dlm_request_join(dlm, node, &response);
node             1633 fs/ocfs2/dlm/dlmdomain.c 			set_bit(node, ctxt->yes_resp_map);
node               49 fs/ocfs2/dlm/dlmlock.c 			  u8 node, u64 cookie);
node              117 fs/ocfs2/dlm/dlmlock.c 	    lock->ml.node != dlm->node_num) {
node              146 fs/ocfs2/dlm/dlmlock.c 			     lock->ml.node);
node              157 fs/ocfs2/dlm/dlmlock.c 				     lock->ml.node);
node              379 fs/ocfs2/dlm/dlmlock.c 			  u8 node, u64 cookie)
node              388 fs/ocfs2/dlm/dlmlock.c 	newlock->ml.node = node;
node              407 fs/ocfs2/dlm/dlmlock.c struct dlm_lock * dlm_new_lock(int type, u8 node, u64 cookie,
node              427 fs/ocfs2/dlm/dlmlock.c 	dlm_init_lock(lock, type, node, cookie);
node               42 fs/ocfs2/dlm/dlmmaster.c 			      struct o2nm_node *node,
node               46 fs/ocfs2/dlm/dlmmaster.c 			    struct o2nm_node *node,
node              362 fs/ocfs2/dlm/dlmmaster.c 			      struct o2nm_node *node, int idx)
node              376 fs/ocfs2/dlm/dlmmaster.c 			    struct o2nm_node *node, int idx)
node             1215 fs/ocfs2/dlm/dlmmaster.c 	int node;
node             1224 fs/ocfs2/dlm/dlmmaster.c 	node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node             1225 fs/ocfs2/dlm/dlmmaster.c 	while (node >= 0) {
node             1230 fs/ocfs2/dlm/dlmmaster.c 			mlog(ML_NOTICE, "node %d up while restarting\n", node);
node             1234 fs/ocfs2/dlm/dlmmaster.c 			clear_bit(node, mle->response_map);
node             1235 fs/ocfs2/dlm/dlmmaster.c 			set_bit(node, mle->vote_map);
node             1237 fs/ocfs2/dlm/dlmmaster.c 			mlog(ML_ERROR, "node down! %d\n", node);
node             1243 fs/ocfs2/dlm/dlmmaster.c 				clear_bit(node, mle->maybe_map);
node             1245 fs/ocfs2/dlm/dlmmaster.c 			       	if (node == lowest) {
node             1248 fs/ocfs2/dlm/dlmmaster.c 					    "waiting on it!\n", node);
node             1292 fs/ocfs2/dlm/dlmmaster.c 		node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node             2260 fs/ocfs2/dlm/dlmmaster.c 	u8 node;
node             2271 fs/ocfs2/dlm/dlmmaster.c 	node = deref->node_idx;
node             2278 fs/ocfs2/dlm/dlmmaster.c 		mlog(ML_ERROR, "Invalid node number: %u\n", node);
node             2299 fs/ocfs2/dlm/dlmmaster.c 		if (test_bit(node, res->refmap)) {
node             2300 fs/ocfs2/dlm/dlmmaster.c 			dlm_lockres_clear_refmap_bit(dlm, res, node);
node             2312 fs/ocfs2/dlm/dlmmaster.c 		     	res->lockname.len, res->lockname.name, node);
node             2328 fs/ocfs2/dlm/dlmmaster.c 	item->u.dl.deref_node = node;
node             2355 fs/ocfs2/dlm/dlmmaster.c 	u8 node;
node             2363 fs/ocfs2/dlm/dlmmaster.c 	node = deref->node_idx;
node             2370 fs/ocfs2/dlm/dlmmaster.c 		mlog(ML_ERROR, "Invalid node number: %u\n", node);
node             2391 fs/ocfs2/dlm/dlmmaster.c 			res->lockname.len, res->lockname.name, node);
node             2411 fs/ocfs2/dlm/dlmmaster.c 		struct dlm_lock_resource *res, u8 node)
node             2428 fs/ocfs2/dlm/dlmmaster.c 				 &deref, sizeof(deref), node, &r);
node             2432 fs/ocfs2/dlm/dlmmaster.c 				lockname, ret, node);
node             2436 fs/ocfs2/dlm/dlmmaster.c 		     dlm->name, namelen, lockname, node, r);
node             2445 fs/ocfs2/dlm/dlmmaster.c 	u8 node;
node             2450 fs/ocfs2/dlm/dlmmaster.c 	node = item->u.dl.deref_node;
node             2455 fs/ocfs2/dlm/dlmmaster.c 	if (test_bit(node, res->refmap)) {
node             2456 fs/ocfs2/dlm/dlmmaster.c 		dlm_lockres_clear_refmap_bit(dlm, res, node);
node             2461 fs/ocfs2/dlm/dlmmaster.c 	dlm_drop_lockres_ref_done(dlm, res, node);
node             2465 fs/ocfs2/dlm/dlmmaster.c 		     dlm->name, res->lockname.len, res->lockname.name, node);
node             2470 fs/ocfs2/dlm/dlmmaster.c 		     res->lockname.len, res->lockname.name, node);
node             2510 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node != dlm->node_num) {
node             2940 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node != dlm->node_num) {
node             2942 fs/ocfs2/dlm/dlmmaster.c 				     lock->ml.node);
node             2949 fs/ocfs2/dlm/dlmmaster.c 							     lock->ml.node);
node             2997 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node == dlm->node_num)
node             2999 fs/ocfs2/dlm/dlmmaster.c 			if (test_bit(lock->ml.node, dlm->exit_domain_map))
node             3001 fs/ocfs2/dlm/dlmmaster.c 			nodenum = lock->ml.node;
node              326 fs/ocfs2/dlm/dlmrecovery.c int dlm_is_node_dead(struct dlm_ctxt *dlm, u8 node)
node              330 fs/ocfs2/dlm/dlmrecovery.c 	dead = !test_bit(node, dlm->domain_map);
node              337 fs/ocfs2/dlm/dlmrecovery.c static int dlm_is_node_recovered(struct dlm_ctxt *dlm, u8 node)
node              341 fs/ocfs2/dlm/dlmrecovery.c 	recovered = !test_bit(node, dlm->recovery_map);
node              347 fs/ocfs2/dlm/dlmrecovery.c void dlm_wait_for_node_death(struct dlm_ctxt *dlm, u8 node, int timeout)
node              349 fs/ocfs2/dlm/dlmrecovery.c 	if (dlm_is_node_dead(dlm, node))
node              353 fs/ocfs2/dlm/dlmrecovery.c 	       "domain %s\n", node, dlm->name);
node              357 fs/ocfs2/dlm/dlmrecovery.c 				   dlm_is_node_dead(dlm, node),
node              361 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_dead(dlm, node));
node              364 fs/ocfs2/dlm/dlmrecovery.c void dlm_wait_for_node_recovery(struct dlm_ctxt *dlm, u8 node, int timeout)
node              366 fs/ocfs2/dlm/dlmrecovery.c 	if (dlm_is_node_recovered(dlm, node))
node              370 fs/ocfs2/dlm/dlmrecovery.c 	       "domain %s\n", node, dlm->name);
node              374 fs/ocfs2/dlm/dlmrecovery.c 				   dlm_is_node_recovered(dlm, node),
node              378 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_recovered(dlm, node));
node             1058 fs/ocfs2/dlm/dlmrecovery.c 				if (lock->ml.node == dead_node) {
node             1209 fs/ocfs2/dlm/dlmrecovery.c 	     lock->ml.node);
node             1232 fs/ocfs2/dlm/dlmrecovery.c 	ml->node = lock->ml.node;
node             1250 fs/ocfs2/dlm/dlmrecovery.c 	dummy.ml.node = dlm->node_num;
node             1263 fs/ocfs2/dlm/dlmrecovery.c 		*nodenum = ml->node;
node             1836 fs/ocfs2/dlm/dlmrecovery.c 		if (ml->node == dlm->node_num) {
node             1865 fs/ocfs2/dlm/dlmrecovery.c 				     ml->node, ml->list, ml->flags, ml->type,
node             1871 fs/ocfs2/dlm/dlmrecovery.c 			if (lock->ml.node != ml->node) {
node             1878 fs/ocfs2/dlm/dlmrecovery.c 				     lock->ml.node);
node             1885 fs/ocfs2/dlm/dlmrecovery.c 				     ml->node, ml->list, ml->flags, ml->type,
node             1917 fs/ocfs2/dlm/dlmrecovery.c 		newlock = dlm_new_lock(ml->type, ml->node,
node             2011 fs/ocfs2/dlm/dlmrecovery.c 	      			     ml->type, ml->convert_type, ml->node,
node             2034 fs/ocfs2/dlm/dlmrecovery.c 			     res->lockname.len, res->lockname.name, ml->node);
node             2035 fs/ocfs2/dlm/dlmrecovery.c 			dlm_lockres_set_refmap_bit(dlm, res, ml->node);
node             2244 fs/ocfs2/dlm/dlmrecovery.c 			if (lock->ml.node == search_node) {
node             2279 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2288 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2297 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2364 fs/ocfs2/dlm/dlmrecovery.c 					if (lock->ml.node == dead_node) {
node             2488 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data)
node             2509 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data)
node               42 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock)     ((lock)->ml.node != (dlm)->node_num)
node              419 fs/ocfs2/dlm/dlmthread.c 		     target->ml.convert_type, target->ml.node);
node              480 fs/ocfs2/dlm/dlmthread.c 		     target->ml.type, target->ml.node);
node              592 fs/ocfs2/dlm/dlmthread.c 		     lock->ml.type, lock->ml.node);
node              601 fs/ocfs2/dlm/dlmthread.c 		if (lock->ml.node != dlm->node_num) {
node              651 fs/ocfs2/dlm/dlmthread.c 		     hi, lock->ml.node);
node              653 fs/ocfs2/dlm/dlmthread.c 		if (lock->ml.node != dlm->node_num) {
node              468 fs/ocfs2/dlm/dlmunlock.c 		    	    lock->ml.node == unlock->node_idx) {
node              530 fs/ocfs2/refcounttree.c 	struct rb_node *node;
node              534 fs/ocfs2/refcounttree.c 	while ((node = rb_last(root)) != NULL) {
node              535 fs/ocfs2/refcounttree.c 		tree = rb_entry(node, struct ocfs2_refcount_tree, rf_node);
node               81 fs/ocfs2/reservations.c 	struct rb_node *node;
node               88 fs/ocfs2/reservations.c 	node = rb_first(&resmap->m_reservations);
node               89 fs/ocfs2/reservations.c 	while (node) {
node               90 fs/ocfs2/reservations.c 		resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
node               97 fs/ocfs2/reservations.c 		node = rb_next(node);
node              140 fs/ocfs2/reservations.c 	struct rb_node *node;
node              143 fs/ocfs2/reservations.c 	node = rb_first(&resmap->m_reservations);
node              144 fs/ocfs2/reservations.c 	while (node) {
node              145 fs/ocfs2/reservations.c 		resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
node              175 fs/ocfs2/reservations.c 		node = rb_next(node);
node              274 fs/ocfs2/reservations.c 	struct rb_node *node;
node              279 fs/ocfs2/reservations.c 	while ((node = rb_last(&resmap->m_reservations)) != NULL) {
node              280 fs/ocfs2/reservations.c 		resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
node              363 fs/ocfs2/reservations.c 	struct rb_node *node = resmap->m_reservations.rb_node;
node              367 fs/ocfs2/reservations.c 	if (!node)
node              370 fs/ocfs2/reservations.c 	node = rb_first(&resmap->m_reservations);
node              371 fs/ocfs2/reservations.c 	while (node) {
node              372 fs/ocfs2/reservations.c 		resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
node              384 fs/ocfs2/reservations.c 		node = rb_next(node);
node              394 fs/ocfs2/stack_o2cb.c 				  unsigned int *node)
node              405 fs/ocfs2/stack_o2cb.c 	*node = node_num;
node              463 fs/ocfs2/stackglue.c 			    unsigned int *node)
node              465 fs/ocfs2/stackglue.c 	return active_stack->sp_ops->this_node(conn, node);
node              153 fs/ocfs2/stackglue.h 			 unsigned int *node);
node              264 fs/ocfs2/stackglue.h 			    unsigned int *node);
node              134 fs/ocfs2/uptodate.c 	struct rb_node *node;
node              137 fs/ocfs2/uptodate.c 	while ((node = rb_last(root)) != NULL) {
node              138 fs/ocfs2/uptodate.c 		item = rb_entry(node, struct ocfs2_meta_cache_item, c_node);
node               33 fs/openpromfs/inode.c 	struct device_node	*node;
node              195 fs/openpromfs/inode.c 	dp = oi->u.node;
node              210 fs/openpromfs/inode.c 			ent_data.node = child;
node              269 fs/openpromfs/inode.c 	struct device_node *dp = oi->u.node;
node              399 fs/openpromfs/inode.c 	oi->u.node = of_find_node_by_path("/");
node              228 fs/orangefs/orangefs-kernel.h 	struct hlist_node node;
node              135 fs/orangefs/super.c 	hash_for_each_safe(orangefs_inode->xattr_cache, i, tmp, cx, node) {
node              136 fs/orangefs/super.c 		hlist_del(&cx->node);
node               72 fs/orangefs/xattr.c 	hlist_for_each_entry_safe(cx, tmp, h, node) {
node              178 fs/orangefs/xattr.c 				hash_add(orangefs_inode->xattr_cache, &cx->node,
node              232 fs/orangefs/xattr.c 			hash_add(orangefs_inode->xattr_cache, &cx->node,
node              295 fs/orangefs/xattr.c 	hlist_for_each_entry_safe(cx, tmp, h, node) {
node              297 fs/orangefs/xattr.c 			hlist_del(&cx->node);
node              386 fs/orangefs/xattr.c 	hlist_for_each_entry_safe(cx, tmp, h, node) {
node              388 fs/orangefs/xattr.c 			hlist_del(&cx->node);
node               24 fs/overlayfs/readdir.c 	struct rb_node node;
node               63 fs/overlayfs/readdir.c 	return rb_entry(n, struct ovl_cache_entry, node);
node               81 fs/overlayfs/readdir.c 			newp = &tmp->node.rb_right;
node               83 fs/overlayfs/readdir.c 			newp = &tmp->node.rb_left;
node               95 fs/overlayfs/readdir.c 	struct rb_node *node = root->rb_node;
node               98 fs/overlayfs/readdir.c 	while (node) {
node               99 fs/overlayfs/readdir.c 		struct ovl_cache_entry *p = ovl_cache_entry_from_node(node);
node              103 fs/overlayfs/readdir.c 			node = p->node.rb_right;
node              105 fs/overlayfs/readdir.c 			node = p->node.rb_left;
node              191 fs/overlayfs/readdir.c 	rb_link_node(&p->node, parent, newp);
node              192 fs/overlayfs/readdir.c 	rb_insert_color(&p->node, rdd->root);
node              589 fs/overlayfs/readdir.c 			rb_link_node(&p->node, parent, newp);
node              590 fs/overlayfs/readdir.c 			rb_insert_color(&p->node, root);
node               72 fs/proc/generic.c 	struct rb_node *node = dir->subdir.rb_node;
node               74 fs/proc/generic.c 	while (node) {
node               75 fs/proc/generic.c 		struct proc_dir_entry *de = rb_entry(node,
node               81 fs/proc/generic.c 			node = node->rb_left;
node               83 fs/proc/generic.c 			node = node->rb_right;
node              114 fs/proc/proc_sysctl.c 	struct rb_node *node = dir->root.rb_node;
node              116 fs/proc/proc_sysctl.c 	while (node)
node              122 fs/proc/proc_sysctl.c 		ctl_node = rb_entry(node, struct ctl_node, node);
node              124 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
node              129 fs/proc/proc_sysctl.c 			node = node->rb_left;
node              131 fs/proc/proc_sysctl.c 			node = node->rb_right;
node              142 fs/proc/proc_sysctl.c 	struct rb_node *node = &head->node[entry - head->ctl_table].node;
node              156 fs/proc/proc_sysctl.c 		parent_node = rb_entry(parent, struct ctl_node, node);
node              158 fs/proc/proc_sysctl.c 		parent_entry = &parent_head->ctl_table[parent_node - parent_head->node];
node              174 fs/proc/proc_sysctl.c 	rb_link_node(node, parent, p);
node              175 fs/proc/proc_sysctl.c 	rb_insert_color(node, &head->parent->root);
node              181 fs/proc/proc_sysctl.c 	struct rb_node *node = &head->node[entry - head->ctl_table].node;
node              183 fs/proc/proc_sysctl.c 	rb_erase(node, &head->parent->root);
node              188 fs/proc/proc_sysctl.c 	struct ctl_node *node, struct ctl_table *table)
node              199 fs/proc/proc_sysctl.c 	head->node = node;
node              201 fs/proc/proc_sysctl.c 	if (node) {
node              203 fs/proc/proc_sysctl.c 		for (entry = table; entry->procname; entry++, node++)
node              204 fs/proc/proc_sysctl.c 			node->header = head;
node              274 fs/proc/proc_sysctl.c 	struct hlist_node *node;
node              279 fs/proc/proc_sysctl.c 		node = hlist_first_rcu(&head->inodes);
node              280 fs/proc/proc_sysctl.c 		if (!node)
node              282 fs/proc/proc_sysctl.c 		ei = hlist_entry(node, struct proc_inode, sysctl_inodes);
node              384 fs/proc/proc_sysctl.c static struct ctl_node *first_usable_entry(struct rb_node *node)
node              388 fs/proc/proc_sysctl.c 	for (;node; node = rb_next(node)) {
node              389 fs/proc/proc_sysctl.c 		ctl_node = rb_entry(node, struct ctl_node, node);
node              408 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
node              418 fs/proc/proc_sysctl.c 	struct ctl_node *ctl_node = &head->node[entry - head->ctl_table];
node              423 fs/proc/proc_sysctl.c 	ctl_node = first_usable_entry(rb_next(&ctl_node->node));
node              428 fs/proc/proc_sysctl.c 		entry = &head->ctl_table[ctl_node - head->node];
node              969 fs/proc/proc_sysctl.c 	struct ctl_node *node;
node              978 fs/proc/proc_sysctl.c 	node = (struct ctl_node *)(new + 1);
node              979 fs/proc/proc_sysctl.c 	table = (struct ctl_table *)(node + 1);
node              985 fs/proc/proc_sysctl.c 	init_header(&new->header, set->dir.header.root, set, node, table);
node             1167 fs/proc/proc_sysctl.c 	struct ctl_node *node;
node             1187 fs/proc/proc_sysctl.c 	node = (struct ctl_node *)(links + 1);
node             1188 fs/proc/proc_sysctl.c 	link_table = (struct ctl_table *)(node + nr_entries);
node             1199 fs/proc/proc_sysctl.c 	init_header(links, dir->header.root, dir->header.set, node, link_table);
node             1325 fs/proc/proc_sysctl.c 	struct ctl_node *node;
node             1336 fs/proc/proc_sysctl.c 	node = (struct ctl_node *)(header + 1);
node             1337 fs/proc/proc_sysctl.c 	init_header(header, root, set, node, table);
node             1659 fs/proc/task_mmu.c 	unsigned long node[MAX_NUMNODES];
node             1691 fs/proc/task_mmu.c 	md->node[page_to_nid(page)] += nr_pages;
node             1884 fs/proc/task_mmu.c 		if (md->node[nid])
node             1885 fs/proc/task_mmu.c 			seq_printf(m, " N%d=%lu", nid, md->node[nid]);
node              290 fs/quota/dquot.c 	struct hlist_node *node;
node              293 fs/quota/dquot.c 	hlist_for_each (node, dquot_hash+hashent) {
node              294 fs/quota/dquot.c 		dquot = hlist_entry(node, struct dquot, dq_hash);
node              936 fs/seq_file.c  	struct hlist_node *node;
node              938 fs/seq_file.c  	hlist_for_each(node, head)
node              940 fs/seq_file.c  			return node;
node              973 fs/seq_file.c  	struct hlist_node *node = v;
node              979 fs/seq_file.c  		return node->next;
node              997 fs/seq_file.c  	struct hlist_node *node;
node              999 fs/seq_file.c  	__hlist_for_each_rcu(node, head)
node             1001 fs/seq_file.c  			return node;
node             1044 fs/seq_file.c  	struct hlist_node *node = v;
node             1050 fs/seq_file.c  		return rcu_dereference(node->next);
node             1065 fs/seq_file.c  	struct hlist_node *node;
node             1068 fs/seq_file.c  		hlist_for_each(node, per_cpu_ptr(head, *cpu)) {
node             1070 fs/seq_file.c  				return node;
node             1090 fs/seq_file.c  	struct hlist_node *node = v;
node             1094 fs/seq_file.c  	if (node->next)
node             1095 fs/seq_file.c  		return node->next;
node              292 fs/super.c     		WARN_ON(s->s_dentry_lru.node);
node              293 fs/super.c     		WARN_ON(s->s_inode_lru.node);
node               30 fs/ubifs/auth.c int __ubifs_node_calc_hash(const struct ubifs_info *c, const void *node,
node               33 fs/ubifs/auth.c 	const struct ubifs_ch *ch = node;
node               39 fs/ubifs/auth.c 	err = crypto_shash_digest(shash, node, le32_to_cpu(ch->len), hash);
node               78 fs/ubifs/auth.c int ubifs_prepare_auth_node(struct ubifs_info *c, void *node,
node               81 fs/ubifs/auth.c 	struct ubifs_auth_node *auth = node;
node              152 fs/ubifs/auth.c void ubifs_bad_hash(const struct ubifs_info *c, const void *node, const u8 *hash,
node              161 fs/ubifs/auth.c 	__ubifs_node_calc_hash(c, node, calc);
node              178 fs/ubifs/auth.c int __ubifs_node_check_hash(const struct ubifs_info *c, const void *node,
node              184 fs/ubifs/auth.c 	err = __ubifs_node_calc_hash(c, node, calc);
node              234 fs/ubifs/auth.c 	signode = snod->node;
node              401 fs/ubifs/auth.c static int ubifs_node_calc_hmac(const struct ubifs_info *c, const void *node,
node              418 fs/ubifs/auth.c 	err = crypto_shash_update(shash, node + 8, ofs_hmac - 8);
node              424 fs/ubifs/auth.c 		err = crypto_shash_update(shash, node + ofs_hmac + hmac_len,
node              445 fs/ubifs/auth.c int __ubifs_node_insert_hmac(const struct ubifs_info *c, void *node, int len,
node              448 fs/ubifs/auth.c 	return ubifs_node_calc_hmac(c, node, len, ofs_hmac, node + ofs_hmac);
node              461 fs/ubifs/auth.c int __ubifs_node_verify_hmac(const struct ubifs_info *c, const void *node,
node              472 fs/ubifs/auth.c 	err = ubifs_node_calc_hmac(c, node, len, ofs_hmac, hmac);
node              478 fs/ubifs/auth.c 	err = crypto_memneq(hmac, node + ofs_hmac, hmac_len);
node              294 fs/ubifs/debug.c void ubifs_dump_node(const struct ubifs_info *c, const void *node)
node              298 fs/ubifs/debug.c 	const struct ubifs_ch *ch = node;
node              305 fs/ubifs/debug.c 			       (void *)node, UBIFS_CH_SZ, 1);
node              310 fs/ubifs/debug.c 	dump_ch(node);
node              315 fs/ubifs/debug.c 		const struct ubifs_pad_node *pad = node;
node              322 fs/ubifs/debug.c 		const struct ubifs_sb_node *sup = node;
node              359 fs/ubifs/debug.c 		const struct ubifs_mst_node *mst = node;
node              401 fs/ubifs/debug.c 		const struct ubifs_ref_node *ref = node;
node              410 fs/ubifs/debug.c 		const struct ubifs_ino_node *ino = node;
node              444 fs/ubifs/debug.c 		const struct ubifs_dent_node *dent = node;
node              469 fs/ubifs/debug.c 		const struct ubifs_data_node *dn = node;
node              486 fs/ubifs/debug.c 		const struct ubifs_trun_node *trun = node;
node              497 fs/ubifs/debug.c 		const struct ubifs_idx_node *idx = node;
node              521 fs/ubifs/debug.c 		const struct ubifs_orph_node *orph = node;
node              806 fs/ubifs/debug.c 		ubifs_dump_node(c, snod->node);
node              837 fs/ubifs/debug.c 		ubifs_dump_node(c, snod->node);
node             1979 fs/ubifs/debug.c 	void *node;
node             1990 fs/ubifs/debug.c 	node = kmalloc(zbr->len, GFP_NOFS);
node             1991 fs/ubifs/debug.c 	if (!node)
node             1994 fs/ubifs/debug.c 	err = ubifs_tnc_read_node(c, zbr, node);
node             2003 fs/ubifs/debug.c 		fscki = add_inode(c, priv, node);
node             2020 fs/ubifs/debug.c 	ch = node;
node             2030 fs/ubifs/debug.c 		struct ubifs_data_node *dn = node;
node             2059 fs/ubifs/debug.c 		struct ubifs_dent_node *dent = node;
node             2107 fs/ubifs/debug.c 	kfree(node);
node             2112 fs/ubifs/debug.c 	ubifs_dump_node(c, node);
node             2114 fs/ubifs/debug.c 	kfree(node);
node             2316 fs/ubifs/debug.c 			ubifs_dump_node(c, sa->node);
node             2321 fs/ubifs/debug.c 			ubifs_dump_node(c, sb->node);
node             2352 fs/ubifs/debug.c 	ubifs_dump_node(c, sa->node);
node             2353 fs/ubifs/debug.c 	ubifs_dump_node(c, sb->node);
node             2384 fs/ubifs/debug.c 			ubifs_dump_node(c, sa->node);
node             2390 fs/ubifs/debug.c 			ubifs_dump_node(c, sb->node);
node             2440 fs/ubifs/debug.c 	ubifs_dump_node(c, sa->node);
node             2442 fs/ubifs/debug.c 	ubifs_dump_node(c, sb->node);
node              245 fs/ubifs/debug.h void ubifs_dump_node(const struct ubifs_info *c, const void *node);
node              312 fs/ubifs/gc.c  	err = ubifs_wbuf_write_nolock(wbuf, snod->node, snod->len);
node              371 fs/ubifs/gc.c  						 snod->node, snod->len);
node              403 fs/ubifs/gc.c  						 snod->node, snod->len);
node              554 fs/ubifs/gc.c  			struct ubifs_idx_node *idx = snod->node;
node              356 fs/ubifs/io.c  void ubifs_init_node(struct ubifs_info *c, void *node, int len, int pad)
node              358 fs/ubifs/io.c  	struct ubifs_ch *ch = node;
node              372 fs/ubifs/io.c  		ubifs_pad(c, node + len, pad);
node              376 fs/ubifs/io.c  void ubifs_crc_node(struct ubifs_info *c, void *node, int len)
node              378 fs/ubifs/io.c  	struct ubifs_ch *ch = node;
node              381 fs/ubifs/io.c  	crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node              400 fs/ubifs/io.c  int ubifs_prepare_node_hmac(struct ubifs_info *c, void *node, int len,
node              405 fs/ubifs/io.c  	ubifs_init_node(c, node, len, pad);
node              408 fs/ubifs/io.c  		err = ubifs_node_insert_hmac(c, node, len, hmac_offs);
node              413 fs/ubifs/io.c  	ubifs_crc_node(c, node, len);
node              429 fs/ubifs/io.c  void ubifs_prepare_node(struct ubifs_info *c, void *node, int len, int pad)
node              435 fs/ubifs/io.c  	ubifs_prepare_node_hmac(c, node, len, 0, pad);
node              448 fs/ubifs/io.c  void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last)
node              451 fs/ubifs/io.c  	struct ubifs_ch *ch = node;
node              464 fs/ubifs/io.c  	crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node              225 fs/ubifs/journal.c static int ubifs_hash_nodes(struct ubifs_info *c, void *node,
node              232 fs/ubifs/journal.c 		const struct ubifs_ch *ch = node;
node              243 fs/ubifs/journal.c 		err = ubifs_shash_update(c, hash, (void *)node, nodelen);
node              247 fs/ubifs/journal.c 		node += ALIGN(nodelen, 8);
node              251 fs/ubifs/journal.c 	return ubifs_prepare_auth_node(c, node, hash);
node              612 fs/ubifs/log.c 		    void *node)
node              614 fs/ubifs/log.c 	struct ubifs_ch *ch = node;
node              627 fs/ubifs/log.c 	memcpy(buf + *offs, node, len);
node              666 fs/ubifs/log.c 				struct ubifs_ref_node *ref = snod->node;
node              674 fs/ubifs/log.c 						       &offs, snod->node);
node              684 fs/ubifs/log.c 					       snod->node);
node             1128 fs/ubifs/lprops.c 			struct ubifs_idx_node *idx = snod->node;
node               69 fs/ubifs/master.c 	const void *node = mst;
node               75 fs/ubifs/master.c 	crypto_shash_digest(shash, node + sizeof(struct ubifs_ch),
node              110 fs/ubifs/master.c 		memcpy(c->mst_node, snod->node, snod->len);
node              129 fs/ubifs/master.c 	if (ubifs_compare_master_node(c, c->mst_node, snod->node))
node              230 fs/ubifs/misc.h 				   const union ubifs_key *key, void *node)
node              232 fs/ubifs/misc.h 	return ubifs_tnc_locate(c, key, node, NULL, NULL);
node              642 fs/ubifs/orphan.c 			ubifs_dump_node(c, snod->node);
node              647 fs/ubifs/orphan.c 		orph = snod->node;
node              670 fs/ubifs/orphan.c 				ubifs_dump_node(c, snod->node);
node              841 fs/ubifs/orphan.c 	struct ubifs_ino_node *node;
node              925 fs/ubifs/orphan.c 		err = ubifs_tnc_read_node(c, zbr, ci->node);
node              930 fs/ubifs/orphan.c 		if (ci->node->nlink == 0)
node              954 fs/ubifs/orphan.c 		orph = snod->node;
node             1013 fs/ubifs/orphan.c 	ci.node = kmalloc(UBIFS_MAX_INO_NODE_SZ, GFP_NOFS);
node             1014 fs/ubifs/orphan.c 	if (!ci.node) {
node             1041 fs/ubifs/orphan.c 	kfree(ci.node);
node              570 fs/ubifs/recovery.c 		ch = snod->node;
node              615 fs/ubifs/replay.c 			struct ubifs_auth_node *auth = snod->node;
node              632 fs/ubifs/replay.c 			err = crypto_shash_update(log_hash, snod->node,
node              737 fs/ubifs/replay.c 		ubifs_node_calc_hash(c, snod->node, hash);
node              745 fs/ubifs/replay.c 			struct ubifs_ino_node *ino = snod->node;
node              757 fs/ubifs/replay.c 			struct ubifs_data_node *dn = snod->node;
node              770 fs/ubifs/replay.c 			struct ubifs_dent_node *dent = snod->node;
node              784 fs/ubifs/replay.c 			struct ubifs_trun_node *trun = snod->node;
node              838 fs/ubifs/replay.c 	ubifs_dump_node(c, snod->node);
node              994 fs/ubifs/replay.c 	const struct ubifs_cs_node *node;
node             1016 fs/ubifs/replay.c 	node = sleb->buf;
node             1031 fs/ubifs/replay.c 		if (le64_to_cpu(node->cmt_no) != c->cmt_no) {
node             1034 fs/ubifs/replay.c 				  (unsigned long long)le64_to_cpu(node->cmt_no),
node             1039 fs/ubifs/replay.c 		c->cs_sqnum = le64_to_cpu(node->ch.sqnum);
node             1046 fs/ubifs/replay.c 		err = ubifs_shash_update(c, c->log_hash, node, UBIFS_CS_NODE_SZ);
node             1088 fs/ubifs/replay.c 			const struct ubifs_ref_node *ref = snod->node;
node             1136 fs/ubifs/replay.c 	ubifs_dump_node(c, snod->node);
node              196 fs/ubifs/scan.c 	snod->node = buf;
node              356 fs/ubifs/scan.c 	struct ubifs_scan_node *node;
node              361 fs/ubifs/scan.c 		node = list_entry(head->next, struct ubifs_scan_node, list);
node              362 fs/ubifs/scan.c 		list_del(&node->list);
node              363 fs/ubifs/scan.c 		kfree(node);
node               28 fs/ubifs/tnc.c 			      struct ubifs_zbranch *zbr, void *node);
node              306 fs/ubifs/tnc.c 		   const void *node)
node              310 fs/ubifs/tnc.c 	const struct ubifs_dent_node *dent = node;
node              323 fs/ubifs/tnc.c 	lnc_node = kmemdup(node, zbr->len, GFP_NOFS);
node              342 fs/ubifs/tnc.c 			    void *node)
node              349 fs/ubifs/tnc.c 	err = ubifs_validate_entry(c, node);
node              352 fs/ubifs/tnc.c 		ubifs_dump_node(c, node);
node              356 fs/ubifs/tnc.c 	zbr->leaf = node;
node              385 fs/ubifs/tnc.c 				void *node)
node              394 fs/ubifs/tnc.c 		memcpy(node, zbr->leaf, zbr->len);
node              399 fs/ubifs/tnc.c 		err = fallible_read_node(c, &zbr->key, zbr, node);
node              409 fs/ubifs/tnc.c 		err = ubifs_tnc_read_node(c, zbr, node);
node              415 fs/ubifs/tnc.c 	err = lnc_add(c, zbr, node);
node              498 fs/ubifs/tnc.c 			      struct ubifs_zbranch *zbr, void *node)
node              504 fs/ubifs/tnc.c 	ret = try_read_node(c, node, key_type(c, key), zbr);
node              507 fs/ubifs/tnc.c 		struct ubifs_dent_node *dent = node;
node             1442 fs/ubifs/tnc.c 		     void *node, int *lnum, int *offs)
node             1468 fs/ubifs/tnc.c 		err = tnc_read_hashed_node(c, zt, node);
node             1472 fs/ubifs/tnc.c 		err = ubifs_tnc_read_node(c, zt, node);
node             1482 fs/ubifs/tnc.c 		err = ubifs_tnc_read_node(c, &zbr, node);
node             1486 fs/ubifs/tnc.c 	err = fallible_read_node(c, key, &zbr, node);
node             1810 fs/ubifs/tnc.c 			void *node, const struct fscrypt_name *nm)
node             1837 fs/ubifs/tnc.c 	err = tnc_read_hashed_node(c, &znode->zbranch[n], node);
node             1858 fs/ubifs/tnc.c 			void *node, const struct fscrypt_name *nm)
node             1861 fs/ubifs/tnc.c 	const struct ubifs_dent_node *dent = node;
node             1867 fs/ubifs/tnc.c 	err = ubifs_tnc_lookup(c, key, node);
node             1880 fs/ubifs/tnc.c 	return do_lookup_nm(c, key, node, nm);
node             1961 fs/ubifs/tnc.c 			void *node, uint32_t cookie)
node             1964 fs/ubifs/tnc.c 	const struct ubifs_dent_node *dent = node;
node             1973 fs/ubifs/tnc.c 	err = ubifs_tnc_lookup(c, key, node);
node             1984 fs/ubifs/tnc.c 	return do_lookup_dh(c, key, node, cookie);
node              256 fs/ubifs/tnc_commit.c 		idx = snod->node;
node              462 fs/ubifs/tnc_misc.c 			void *node)
node              474 fs/ubifs/tnc_misc.c 		err = ubifs_read_node_wbuf(wbuf, node, type, zbr->len,
node              477 fs/ubifs/tnc_misc.c 		err = ubifs_read_node(c, node, type, zbr->len, zbr->lnum,
node              486 fs/ubifs/tnc_misc.c 	key_read(c, node + UBIFS_KEY_OFFSET, &key1);
node              492 fs/ubifs/tnc_misc.c 		ubifs_dump_node(c, node);
node              496 fs/ubifs/tnc_misc.c 	err = ubifs_node_check_hash(c, node, zbr->hash);
node              498 fs/ubifs/tnc_misc.c 		ubifs_bad_hash(c, node, zbr->hash, zbr->lnum, zbr->offs);
node              308 fs/ubifs/ubifs.h 	void *node;
node             1561 fs/ubifs/ubifs.h int ubifs_prepare_auth_node(struct ubifs_info *c, void *node,
node             1594 fs/ubifs/ubifs.h void ubifs_bad_hash(const struct ubifs_info *c, const void *node,
node             1717 fs/ubifs/ubifs.h int ubifs_write_node(struct ubifs_info *c, void *node, int len, int lnum,
node             1726 fs/ubifs/ubifs.h int ubifs_prepare_node_hmac(struct ubifs_info *c, void *node, int len,
node             1728 fs/ubifs/ubifs.h void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last);
node             1821 fs/ubifs/ubifs.h 			void *node, const struct fscrypt_name *nm);
node             1823 fs/ubifs/ubifs.h 			void *node, uint32_t secondary_hash);
node             1825 fs/ubifs/ubifs.h 		     void *node, int *lnum, int *offs);
node             1873 fs/ubifs/ubifs.h 			void *node);
node              366 fs/unicode/mkutf8data.c 	struct node *parent;
node              382 fs/unicode/mkutf8data.c 	struct node *node;
node              385 fs/unicode/mkutf8data.c 	node = tree->root;
node              386 fs/unicode/mkutf8data.c 	while (!leaf && node) {
node              387 fs/unicode/mkutf8data.c 		if (node->nextbyte)
node              389 fs/unicode/mkutf8data.c 		if (*key & (1 << (node->bitnum & 7))) {
node              391 fs/unicode/mkutf8data.c 			if (node->rightnode == NODE) {
node              392 fs/unicode/mkutf8data.c 				node = node->right;
node              393 fs/unicode/mkutf8data.c 			} else if (node->rightnode == LEAF) {
node              394 fs/unicode/mkutf8data.c 				leaf = node->right;
node              396 fs/unicode/mkutf8data.c 				node = NULL;
node              400 fs/unicode/mkutf8data.c 			if (node->leftnode == NODE) {
node              401 fs/unicode/mkutf8data.c 				node = node->left;
node              402 fs/unicode/mkutf8data.c 			} else if (node->leftnode == LEAF) {
node              403 fs/unicode/mkutf8data.c 				leaf = node->left;
node              405 fs/unicode/mkutf8data.c 				node = NULL;
node              419 fs/unicode/mkutf8data.c 	struct node *node;
node              435 fs/unicode/mkutf8data.c 		node = tree->root;
node              437 fs/unicode/mkutf8data.c 		while (node) {
node              440 fs/unicode/mkutf8data.c 				indent, "", node,
node              441 fs/unicode/mkutf8data.c 				node->bitnum, node->nextbyte,
node              442 fs/unicode/mkutf8data.c 				node->left, node->right,
node              443 fs/unicode/mkutf8data.c 				node->keymask, node->keybits);
node              445 fs/unicode/mkutf8data.c 			if (!(node->left && node->right))
node              448 fs/unicode/mkutf8data.c 			while (node) {
node              449 fs/unicode/mkutf8data.c 				bitmask = 1 << node->bitnum;
node              452 fs/unicode/mkutf8data.c 					if (node->leftnode == LEAF) {
node              453 fs/unicode/mkutf8data.c 						assert(node->left);
node              454 fs/unicode/mkutf8data.c 						tree->leaf_print(node->left,
node              457 fs/unicode/mkutf8data.c 					} else if (node->left) {
node              458 fs/unicode/mkutf8data.c 						assert(node->leftnode == NODE);
node              460 fs/unicode/mkutf8data.c 						node = node->left;
node              466 fs/unicode/mkutf8data.c 					if (node->rightnode == LEAF) {
node              467 fs/unicode/mkutf8data.c 						assert(node->right);
node              468 fs/unicode/mkutf8data.c 						tree->leaf_print(node->right,
node              471 fs/unicode/mkutf8data.c 					} else if (node->right) {
node              472 fs/unicode/mkutf8data.c 						assert(node->rightnode == NODE);
node              474 fs/unicode/mkutf8data.c 						node = node->right;
node              480 fs/unicode/mkutf8data.c 				node = node->parent;
node              492 fs/unicode/mkutf8data.c static struct node *alloc_node(struct node *parent)
node              494 fs/unicode/mkutf8data.c 	struct node *node;
node              497 fs/unicode/mkutf8data.c 	node = malloc(sizeof(*node));
node              498 fs/unicode/mkutf8data.c 	node->left = node->right = NULL;
node              499 fs/unicode/mkutf8data.c 	node->parent = parent;
node              500 fs/unicode/mkutf8data.c 	node->leftnode = NODE;
node              501 fs/unicode/mkutf8data.c 	node->rightnode = NODE;
node              502 fs/unicode/mkutf8data.c 	node->keybits = 0;
node              503 fs/unicode/mkutf8data.c 	node->keymask = 0;
node              504 fs/unicode/mkutf8data.c 	node->mark = 0;
node              505 fs/unicode/mkutf8data.c 	node->index = 0;
node              506 fs/unicode/mkutf8data.c 	node->offset = -1;
node              507 fs/unicode/mkutf8data.c 	node->size = 4;
node              509 fs/unicode/mkutf8data.c 	if (node->parent) {
node              512 fs/unicode/mkutf8data.c 			node->bitnum = bitnum + 7 + 8;
node              513 fs/unicode/mkutf8data.c 			node->nextbyte = 1;
node              515 fs/unicode/mkutf8data.c 			node->bitnum = bitnum - 1;
node              516 fs/unicode/mkutf8data.c 			node->nextbyte = 0;
node              519 fs/unicode/mkutf8data.c 		node->bitnum = 7;
node              520 fs/unicode/mkutf8data.c 		node->nextbyte = 0;
node              523 fs/unicode/mkutf8data.c 	return node;
node              535 fs/unicode/mkutf8data.c 	struct node *node;
node              536 fs/unicode/mkutf8data.c 	struct node *parent;
node              542 fs/unicode/mkutf8data.c 	node = NULL;
node              549 fs/unicode/mkutf8data.c 			*cursor = alloc_node(node);
node              550 fs/unicode/mkutf8data.c 		node = *cursor;
node              551 fs/unicode/mkutf8data.c 		if (node->nextbyte)
node              553 fs/unicode/mkutf8data.c 		if (*key & (1 << (node->bitnum & 7)))
node              554 fs/unicode/mkutf8data.c 			cursor = &node->right;
node              556 fs/unicode/mkutf8data.c 			cursor = &node->left;
node              562 fs/unicode/mkutf8data.c 	while (node) {
node              563 fs/unicode/mkutf8data.c 		if (*key & (1 << (node->bitnum & 7)))
node              564 fs/unicode/mkutf8data.c 			node->rightnode = LEAF;
node              566 fs/unicode/mkutf8data.c 			node->leftnode = LEAF;
node              567 fs/unicode/mkutf8data.c 		if (node->nextbyte)
node              569 fs/unicode/mkutf8data.c 		if (node->leftnode == NODE || node->rightnode == NODE)
node              571 fs/unicode/mkutf8data.c 		assert(node->left);
node              572 fs/unicode/mkutf8data.c 		assert(node->right);
node              574 fs/unicode/mkutf8data.c 		if (! tree->leaf_equal(node->left, node->right))
node              577 fs/unicode/mkutf8data.c 		leaf = node->left;
node              579 fs/unicode/mkutf8data.c 		parent = node->parent;
node              584 fs/unicode/mkutf8data.c 		} else if (parent->left == node) {
node              591 fs/unicode/mkutf8data.c 				parent->keymask |= (1 << node->bitnum);
node              593 fs/unicode/mkutf8data.c 		} else if (parent->right == node) {
node              600 fs/unicode/mkutf8data.c 				parent->keymask |= (1 << node->bitnum);
node              601 fs/unicode/mkutf8data.c 				parent->keybits |= (1 << node->bitnum);
node              607 fs/unicode/mkutf8data.c 		free(node);
node              608 fs/unicode/mkutf8data.c 		node = parent;
node              612 fs/unicode/mkutf8data.c 	while (node) {
node              613 fs/unicode/mkutf8data.c 		parent = node->parent;
node              617 fs/unicode/mkutf8data.c 		if (node->keymask == 0) {
node              624 fs/unicode/mkutf8data.c 			assert((parent->keymask & node->keymask) == 0);
node              625 fs/unicode/mkutf8data.c 			parent->keymask |= node->keymask;
node              627 fs/unicode/mkutf8data.c 			parent->keybits |= node->keybits;
node              631 fs/unicode/mkutf8data.c 		node = parent;
node              656 fs/unicode/mkutf8data.c 	struct node *node;
node              657 fs/unicode/mkutf8data.c 	struct node *left;
node              658 fs/unicode/mkutf8data.c 	struct node *right;
node              659 fs/unicode/mkutf8data.c 	struct node *parent;
node              677 fs/unicode/mkutf8data.c 	node = tree->root;
node              678 fs/unicode/mkutf8data.c 	while (node) {
node              679 fs/unicode/mkutf8data.c 		if (node->nextbyte)
node              681 fs/unicode/mkutf8data.c 		if (node->leftnode == LEAF)
node              683 fs/unicode/mkutf8data.c 		if (node->rightnode == LEAF)
node              685 fs/unicode/mkutf8data.c 		if (!node->left)
node              687 fs/unicode/mkutf8data.c 		if (!node->right)
node              689 fs/unicode/mkutf8data.c 		left = node->left;
node              690 fs/unicode/mkutf8data.c 		right = node->right;
node              733 fs/unicode/mkutf8data.c 		parent = node->parent;
node              734 fs/unicode/mkutf8data.c 		left = node->left;
node              735 fs/unicode/mkutf8data.c 		right = node->right;
node              736 fs/unicode/mkutf8data.c 		if (parent->left == node)
node              738 fs/unicode/mkutf8data.c 		else if (parent->right == node)
node              743 fs/unicode/mkutf8data.c 		left->keymask |= (1 << node->bitnum);
node              744 fs/unicode/mkutf8data.c 		node->left = NULL;
node              745 fs/unicode/mkutf8data.c 		while (node) {
node              746 fs/unicode/mkutf8data.c 			bitmask = 1 << node->bitnum;
node              749 fs/unicode/mkutf8data.c 			if (node->leftnode == NODE && node->left) {
node              750 fs/unicode/mkutf8data.c 				left = node->left;
node              751 fs/unicode/mkutf8data.c 				free(node);
node              753 fs/unicode/mkutf8data.c 				node = left;
node              754 fs/unicode/mkutf8data.c 			} else if (node->rightnode == NODE && node->right) {
node              755 fs/unicode/mkutf8data.c 				right = node->right;
node              756 fs/unicode/mkutf8data.c 				free(node);
node              758 fs/unicode/mkutf8data.c 				node = right;
node              760 fs/unicode/mkutf8data.c 				node = NULL;
node              764 fs/unicode/mkutf8data.c 		node = parent;
node              766 fs/unicode/mkutf8data.c 		bitmask = 1 << node->bitnum;
node              770 fs/unicode/mkutf8data.c 			if (node->left && node->right)
node              772 fs/unicode/mkutf8data.c 			if (node->left) {
node              773 fs/unicode/mkutf8data.c 				left = node->left;
node              774 fs/unicode/mkutf8data.c 				node->keymask |= left->keymask;
node              775 fs/unicode/mkutf8data.c 				node->keybits |= left->keybits;
node              777 fs/unicode/mkutf8data.c 			if (node->right) {
node              778 fs/unicode/mkutf8data.c 				right = node->right;
node              779 fs/unicode/mkutf8data.c 				node->keymask |= right->keymask;
node              780 fs/unicode/mkutf8data.c 				node->keybits |= right->keybits;
node              782 fs/unicode/mkutf8data.c 			node->keymask |= (1 << node->bitnum);
node              783 fs/unicode/mkutf8data.c 			node = node->parent;
node              785 fs/unicode/mkutf8data.c 			bitmask = 1 << node->bitnum;
node              790 fs/unicode/mkutf8data.c 		bitmask = 1 << node->bitnum;
node              792 fs/unicode/mkutf8data.c 		    node->leftnode == NODE &&
node              793 fs/unicode/mkutf8data.c 		    node->left) {
node              795 fs/unicode/mkutf8data.c 			node = node->left;
node              797 fs/unicode/mkutf8data.c 			   node->rightnode == NODE &&
node              798 fs/unicode/mkutf8data.c 			   node->right) {
node              800 fs/unicode/mkutf8data.c 			node = node->right;
node              804 fs/unicode/mkutf8data.c 			node = node->parent;
node              817 fs/unicode/mkutf8data.c 	struct node *node;
node              818 fs/unicode/mkutf8data.c 	struct node *n;
node              831 fs/unicode/mkutf8data.c 	node = tree->root;
node              833 fs/unicode/mkutf8data.c 	while (node) {
node              834 fs/unicode/mkutf8data.c 		bitmask = 1 << node->bitnum;
node              837 fs/unicode/mkutf8data.c 			if (node->leftnode == LEAF) {
node              838 fs/unicode/mkutf8data.c 				assert(node->left);
node              839 fs/unicode/mkutf8data.c 				if (tree->leaf_mark(node->left)) {
node              840 fs/unicode/mkutf8data.c 					n = node;
node              847 fs/unicode/mkutf8data.c 			} else if (node->left) {
node              848 fs/unicode/mkutf8data.c 				assert(node->leftnode == NODE);
node              849 fs/unicode/mkutf8data.c 				node = node->left;
node              855 fs/unicode/mkutf8data.c 			if (node->rightnode == LEAF) {
node              856 fs/unicode/mkutf8data.c 				assert(node->right);
node              857 fs/unicode/mkutf8data.c 				if (tree->leaf_mark(node->right)) {
node              858 fs/unicode/mkutf8data.c 					n = node;
node              865 fs/unicode/mkutf8data.c 			} else if (node->right) {
node              866 fs/unicode/mkutf8data.c 				assert(node->rightnode == NODE);
node              867 fs/unicode/mkutf8data.c 				node = node->right;
node              873 fs/unicode/mkutf8data.c 		node = node->parent;
node              879 fs/unicode/mkutf8data.c 	node = tree->root;
node              881 fs/unicode/mkutf8data.c 	while (node) {
node              882 fs/unicode/mkutf8data.c 		bitmask = 1 << node->bitnum;
node              885 fs/unicode/mkutf8data.c 			if (node->leftnode == LEAF) {
node              886 fs/unicode/mkutf8data.c 				assert(node->left);
node              887 fs/unicode/mkutf8data.c 				if (tree->leaf_mark(node->left)) {
node              888 fs/unicode/mkutf8data.c 					n = node;
node              895 fs/unicode/mkutf8data.c 			} else if (node->left) {
node              896 fs/unicode/mkutf8data.c 				assert(node->leftnode == NODE);
node              897 fs/unicode/mkutf8data.c 				node = node->left;
node              898 fs/unicode/mkutf8data.c 				if (!node->mark && node->parent->mark) {
node              900 fs/unicode/mkutf8data.c 					node->mark = 1;
node              907 fs/unicode/mkutf8data.c 			if (node->rightnode == LEAF) {
node              908 fs/unicode/mkutf8data.c 				assert(node->right);
node              909 fs/unicode/mkutf8data.c 				if (tree->leaf_mark(node->right)) {
node              910 fs/unicode/mkutf8data.c 					n = node;
node              917 fs/unicode/mkutf8data.c 			} else if (node->right) {
node              918 fs/unicode/mkutf8data.c 				assert(node->rightnode == NODE);
node              919 fs/unicode/mkutf8data.c 				node = node->right;
node              920 fs/unicode/mkutf8data.c 				if (!node->mark && node->parent->mark &&
node              921 fs/unicode/mkutf8data.c 				    !node->parent->left) {
node              923 fs/unicode/mkutf8data.c 					node->mark = 1;
node              930 fs/unicode/mkutf8data.c 		node = node->parent;
node              944 fs/unicode/mkutf8data.c 	struct node *node;
node              966 fs/unicode/mkutf8data.c 	node = tree->root;
node              968 fs/unicode/mkutf8data.c 	while (node) {
node              969 fs/unicode/mkutf8data.c 		if (!node->mark)
node              972 fs/unicode/mkutf8data.c 		if (node->index != index)
node              973 fs/unicode/mkutf8data.c 			node->index = index;
node              974 fs/unicode/mkutf8data.c 		index += node->size;
node              976 fs/unicode/mkutf8data.c 		while (node) {
node              977 fs/unicode/mkutf8data.c 			bitmask = 1 << node->bitnum;
node              978 fs/unicode/mkutf8data.c 			if (node->mark && (leftmask & bitmask) == 0) {
node              980 fs/unicode/mkutf8data.c 				if (node->leftnode == LEAF) {
node              981 fs/unicode/mkutf8data.c 					assert(node->left);
node              982 fs/unicode/mkutf8data.c 					*tree->leaf_index(tree, node->left) =
node              984 fs/unicode/mkutf8data.c 					index += tree->leaf_size(node->left);
node              986 fs/unicode/mkutf8data.c 				} else if (node->left) {
node              987 fs/unicode/mkutf8data.c 					assert(node->leftnode == NODE);
node              989 fs/unicode/mkutf8data.c 					node = node->left;
node              993 fs/unicode/mkutf8data.c 			if (node->mark && (rightmask & bitmask) == 0) {
node              995 fs/unicode/mkutf8data.c 				if (node->rightnode == LEAF) {
node              996 fs/unicode/mkutf8data.c 					assert(node->right);
node              997 fs/unicode/mkutf8data.c 					*tree->leaf_index(tree, node->right) = index;
node              998 fs/unicode/mkutf8data.c 					index += tree->leaf_size(node->right);
node             1000 fs/unicode/mkutf8data.c 				} else if (node->right) {
node             1001 fs/unicode/mkutf8data.c 					assert(node->rightnode == NODE);
node             1003 fs/unicode/mkutf8data.c 					node = node->right;
node             1009 fs/unicode/mkutf8data.c 			node = node->parent;
node             1025 fs/unicode/mkutf8data.c static int mark_subtree(struct node *node)
node             1029 fs/unicode/mkutf8data.c 	if (!node || node->mark)
node             1031 fs/unicode/mkutf8data.c 	node->mark = 1;
node             1032 fs/unicode/mkutf8data.c 	node->index = node->parent->index;
node             1034 fs/unicode/mkutf8data.c 	if (node->leftnode == NODE)
node             1035 fs/unicode/mkutf8data.c 		changed += mark_subtree(node->left);
node             1036 fs/unicode/mkutf8data.c 	if (node->rightnode == NODE)
node             1037 fs/unicode/mkutf8data.c 		changed += mark_subtree(node->right);
node             1051 fs/unicode/mkutf8data.c 	struct node *node;
node             1052 fs/unicode/mkutf8data.c 	struct node *right;
node             1053 fs/unicode/mkutf8data.c 	struct node *n;
node             1077 fs/unicode/mkutf8data.c 	node = tree->root;
node             1079 fs/unicode/mkutf8data.c 	while (node) {
node             1080 fs/unicode/mkutf8data.c 		if (!node->mark)
node             1083 fs/unicode/mkutf8data.c 		if (!node->left || !node->right) {
node             1086 fs/unicode/mkutf8data.c 			if (node->rightnode == NODE) {
node             1093 fs/unicode/mkutf8data.c 				right = node->right;
node             1098 fs/unicode/mkutf8data.c 					while (n->bitnum != node->bitnum) {
node             1112 fs/unicode/mkutf8data.c 					if (n->bitnum != node->bitnum)
node             1121 fs/unicode/mkutf8data.c 				offset = right->index - node->index;
node             1123 fs/unicode/mkutf8data.c 				offset = *tree->leaf_index(tree, node->right);
node             1124 fs/unicode/mkutf8data.c 				offset -= node->index;
node             1136 fs/unicode/mkutf8data.c 		if (node->size != size || node->offset != offset) {
node             1137 fs/unicode/mkutf8data.c 			node->size = size;
node             1138 fs/unicode/mkutf8data.c 			node->offset = offset;
node             1142 fs/unicode/mkutf8data.c 		while (node) {
node             1143 fs/unicode/mkutf8data.c 			bitmask = 1 << node->bitnum;
node             1145 fs/unicode/mkutf8data.c 			if (node->mark && (leftmask & bitmask) == 0) {
node             1147 fs/unicode/mkutf8data.c 				if (node->leftnode == LEAF) {
node             1148 fs/unicode/mkutf8data.c 					assert(node->left);
node             1149 fs/unicode/mkutf8data.c 				} else if (node->left) {
node             1150 fs/unicode/mkutf8data.c 					assert(node->leftnode == NODE);
node             1152 fs/unicode/mkutf8data.c 					node = node->left;
node             1156 fs/unicode/mkutf8data.c 			if (node->mark && (rightmask & bitmask) == 0) {
node             1159 fs/unicode/mkutf8data.c 				if (node->rightnode == LEAF) {
node             1160 fs/unicode/mkutf8data.c 					assert(node->right);
node             1161 fs/unicode/mkutf8data.c 				} else if (node->right) {
node             1162 fs/unicode/mkutf8data.c 					assert(node->rightnode == NODE);
node             1164 fs/unicode/mkutf8data.c 					node = node->right;
node             1172 fs/unicode/mkutf8data.c 			node = node->parent;
node             1187 fs/unicode/mkutf8data.c 	struct node *node;
node             1219 fs/unicode/mkutf8data.c 	node = tree->root;
node             1221 fs/unicode/mkutf8data.c 	while (node) {
node             1222 fs/unicode/mkutf8data.c 		if (!node->mark)
node             1224 fs/unicode/mkutf8data.c 		assert(node->offset != -1);
node             1225 fs/unicode/mkutf8data.c 		assert(node->index == index);
node             1228 fs/unicode/mkutf8data.c 		if (node->nextbyte)
node             1230 fs/unicode/mkutf8data.c 		byte |= (node->bitnum & BITNUM);
node             1231 fs/unicode/mkutf8data.c 		if (node->left && node->right) {
node             1232 fs/unicode/mkutf8data.c 			if (node->leftnode == NODE)
node             1234 fs/unicode/mkutf8data.c 			if (node->rightnode == NODE)
node             1236 fs/unicode/mkutf8data.c 			if (node->offset <= 0xff)
node             1238 fs/unicode/mkutf8data.c 			else if (node->offset <= 0xffff)
node             1243 fs/unicode/mkutf8data.c 			offset = node->offset;
node             1252 fs/unicode/mkutf8data.c 		} else if (node->left) {
node             1253 fs/unicode/mkutf8data.c 			if (node->leftnode == NODE)
node             1258 fs/unicode/mkutf8data.c 		} else if (node->right) {
node             1260 fs/unicode/mkutf8data.c 			if (node->rightnode == NODE)
node             1269 fs/unicode/mkutf8data.c 		while (node) {
node             1270 fs/unicode/mkutf8data.c 			bitmask = 1 << node->bitnum;
node             1271 fs/unicode/mkutf8data.c 			if (node->mark && (leftmask & bitmask) == 0) {
node             1273 fs/unicode/mkutf8data.c 				if (node->leftnode == LEAF) {
node             1274 fs/unicode/mkutf8data.c 					assert(node->left);
node             1275 fs/unicode/mkutf8data.c 					data = tree->leaf_emit(node->left,
node             1277 fs/unicode/mkutf8data.c 					size = tree->leaf_size(node->left);
node             1281 fs/unicode/mkutf8data.c 				} else if (node->left) {
node             1282 fs/unicode/mkutf8data.c 					assert(node->leftnode == NODE);
node             1284 fs/unicode/mkutf8data.c 					node = node->left;
node             1288 fs/unicode/mkutf8data.c 			if (node->mark && (rightmask & bitmask) == 0) {
node             1290 fs/unicode/mkutf8data.c 				if (node->rightnode == LEAF) {
node             1291 fs/unicode/mkutf8data.c 					assert(node->right);
node             1292 fs/unicode/mkutf8data.c 					data = tree->leaf_emit(node->right,
node             1294 fs/unicode/mkutf8data.c 					size = tree->leaf_size(node->right);
node             1298 fs/unicode/mkutf8data.c 				} else if (node->right) {
node             1299 fs/unicode/mkutf8data.c 					assert(node->rightnode == NODE);
node             1301 fs/unicode/mkutf8data.c 					node = node->right;
node             1307 fs/unicode/mkutf8data.c 			node = node->parent;
node             2713 fs/unicode/mkutf8data.c 	int		node;
node             2719 fs/unicode/mkutf8data.c 	node = 1;
node             2721 fs/unicode/mkutf8data.c 	while (node) {
node             2733 fs/unicode/mkutf8data.c 				node = (*trie & RIGHTNODE);
node             2742 fs/unicode/mkutf8data.c 				node = (*trie & TRIENODE);
node             2752 fs/unicode/mkutf8data.c 				node = (*trie & LEFTNODE);
node             2759 fs/unicode/mkutf8data.c 				node = (*trie & TRIENODE);
node              326 fs/unicode/utf8-norm.c 	int		node;
node              334 fs/unicode/utf8-norm.c 	node = 1;
node              335 fs/unicode/utf8-norm.c 	while (node) {
node              347 fs/unicode/utf8-norm.c 				node = (*trie & RIGHTNODE);
node              356 fs/unicode/utf8-norm.c 				node = (*trie & TRIENODE);
node              366 fs/unicode/utf8-norm.c 				node = (*trie & LEFTNODE);
node              373 fs/unicode/utf8-norm.c 				node = (*trie & TRIENODE);
node             1085 fs/xfs/libxfs/xfs_attr_leaf.c 	struct xfs_da_intnode	*node;
node             1122 fs/xfs/libxfs/xfs_attr_leaf.c 	node = bp1->b_addr;
node             1123 fs/xfs/libxfs/xfs_attr_leaf.c 	dp->d_ops->node_hdr_from_disk(&icnodehdr, node);
node             1124 fs/xfs/libxfs/xfs_attr_leaf.c 	btree = dp->d_ops->node_tree_p(node);
node             1134 fs/xfs/libxfs/xfs_attr_leaf.c 	dp->d_ops->node_hdr_to_disk(node, &icnodehdr);
node              335 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node              351 fs/xfs/libxfs/xfs_da_btree.c 	node = bp->b_addr;
node              366 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_to_disk(node, &ichdr);
node              368 fs/xfs/libxfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size));
node              385 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node              504 fs/xfs/libxfs/xfs_da_btree.c 	node = oldblk->bp->b_addr;
node              505 fs/xfs/libxfs/xfs_da_btree.c 	if (node->hdr.info.forw) {
node              506 fs/xfs/libxfs/xfs_da_btree.c 		if (be32_to_cpu(node->hdr.info.forw) != addblk->blkno) {
node              510 fs/xfs/libxfs/xfs_da_btree.c 		node = addblk->bp->b_addr;
node              511 fs/xfs/libxfs/xfs_da_btree.c 		node->hdr.info.back = cpu_to_be32(oldblk->blkno);
node              513 fs/xfs/libxfs/xfs_da_btree.c 				  XFS_DA_LOGRANGE(node, &node->hdr.info,
node              514 fs/xfs/libxfs/xfs_da_btree.c 				  sizeof(node->hdr.info)));
node              516 fs/xfs/libxfs/xfs_da_btree.c 	node = oldblk->bp->b_addr;
node              517 fs/xfs/libxfs/xfs_da_btree.c 	if (node->hdr.info.back) {
node              518 fs/xfs/libxfs/xfs_da_btree.c 		if (be32_to_cpu(node->hdr.info.back) != addblk->blkno) {
node              522 fs/xfs/libxfs/xfs_da_btree.c 		node = addblk->bp->b_addr;
node              523 fs/xfs/libxfs/xfs_da_btree.c 		node->hdr.info.forw = cpu_to_be32(oldblk->blkno);
node              525 fs/xfs/libxfs/xfs_da_btree.c 				  XFS_DA_LOGRANGE(node, &node->hdr.info,
node              526 fs/xfs/libxfs/xfs_da_btree.c 				  sizeof(node->hdr.info)));
node              544 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node              574 fs/xfs/libxfs/xfs_da_btree.c 	node = bp->b_addr;
node              616 fs/xfs/libxfs/xfs_da_btree.c 	memcpy(node, oldroot, size);
node              619 fs/xfs/libxfs/xfs_da_btree.c 		struct xfs_da3_intnode *node3 = (struct xfs_da3_intnode *)node;
node              639 fs/xfs/libxfs/xfs_da_btree.c 	node = bp->b_addr;
node              640 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node              641 fs/xfs/libxfs/xfs_da_btree.c 	btree = dp->d_ops->node_tree_p(node);
node              647 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_to_disk(node, &nodehdr);
node              661 fs/xfs/libxfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, btree, sizeof(xfs_da_node_entry_t) * 2));
node              678 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node              688 fs/xfs/libxfs/xfs_da_btree.c 	node = oldblk->bp->b_addr;
node              689 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node              735 fs/xfs/libxfs/xfs_da_btree.c 	node = oldblk->bp->b_addr;
node              736 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node              915 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node              923 fs/xfs/libxfs/xfs_da_btree.c 	node = oldblk->bp->b_addr;
node              924 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node              925 fs/xfs/libxfs/xfs_da_btree.c 	btree = dp->d_ops->node_tree_p(node);
node              944 fs/xfs/libxfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, &btree[oldblk->index],
node              948 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_to_disk(node, &nodehdr);
node              950 fs/xfs/libxfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size));
node             1152 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node             1174 fs/xfs/libxfs/xfs_da_btree.c 	node = (xfs_da_intnode_t *)info;
node             1175 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1232 fs/xfs/libxfs/xfs_da_btree.c 		node = bp->b_addr;
node             1233 fs/xfs/libxfs/xfs_da_btree.c 		dp->d_ops->node_hdr_from_disk(&thdr, node);
node             1275 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	 *node;
node             1279 fs/xfs/libxfs/xfs_da_btree.c 	node = bp->b_addr;
node             1280 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1285 fs/xfs/libxfs/xfs_da_btree.c 	btree = dp->d_ops->node_tree_p(node);
node             1299 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node             1330 fs/xfs/libxfs/xfs_da_btree.c 		node = blk->bp->b_addr;
node             1331 fs/xfs/libxfs/xfs_da_btree.c 		dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1332 fs/xfs/libxfs/xfs_da_btree.c 		btree = dp->d_ops->node_tree_p(node);
node             1338 fs/xfs/libxfs/xfs_da_btree.c 				  XFS_DA_LOGRANGE(node, &btree[blk->index],
node             1353 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node             1362 fs/xfs/libxfs/xfs_da_btree.c 	node = drop_blk->bp->b_addr;
node             1363 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1371 fs/xfs/libxfs/xfs_da_btree.c 	btree = dp->d_ops->node_tree_p(node);
node             1377 fs/xfs/libxfs/xfs_da_btree.c 		    XFS_DA_LOGRANGE(node, &btree[index], tmp));
node             1382 fs/xfs/libxfs/xfs_da_btree.c 	    XFS_DA_LOGRANGE(node, &btree[index], sizeof(btree[index])));
node             1384 fs/xfs/libxfs/xfs_da_btree.c 	dp->d_ops->node_hdr_to_disk(node, &nodehdr);
node             1386 fs/xfs/libxfs/xfs_da_btree.c 	    XFS_DA_LOGRANGE(node, &node->hdr, dp->d_ops->node_hdr_size));
node             1489 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node             1552 fs/xfs/libxfs/xfs_da_btree.c 		node = blk->bp->b_addr;
node             1553 fs/xfs/libxfs/xfs_da_btree.c 		dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1554 fs/xfs/libxfs/xfs_da_btree.c 		btree = dp->d_ops->node_tree_p(node);
node             1881 fs/xfs/libxfs/xfs_da_btree.c 	struct xfs_da_intnode	*node;
node             1904 fs/xfs/libxfs/xfs_da_btree.c 		node = blk->bp->b_addr;
node             1905 fs/xfs/libxfs/xfs_da_btree.c 		dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1906 fs/xfs/libxfs/xfs_da_btree.c 		btree = dp->d_ops->node_tree_p(node);
node             1965 fs/xfs/libxfs/xfs_da_btree.c 			node = (xfs_da_intnode_t *)info;
node             1966 fs/xfs/libxfs/xfs_da_btree.c 			dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node             1967 fs/xfs/libxfs/xfs_da_btree.c 			btree = dp->d_ops->node_tree_p(node);
node              161 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = ifp->if_u1.if_root;
node              168 fs/xfs/libxfs/xfs_iext_tree.c 		node = node->ptrs[0];
node              169 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(node);
node              172 fs/xfs/libxfs/xfs_iext_tree.c 	return node;
node              179 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = ifp->if_u1.if_root;
node              187 fs/xfs/libxfs/xfs_iext_tree.c 			if (!node->ptrs[i])
node              189 fs/xfs/libxfs/xfs_iext_tree.c 		node = node->ptrs[i - 1];
node              190 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(node);
node              193 fs/xfs/libxfs/xfs_iext_tree.c 	return node;
node              277 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node              281 fs/xfs/libxfs/xfs_iext_tree.c 	if (node->keys[n] > offset)
node              283 fs/xfs/libxfs/xfs_iext_tree.c 	if (node->keys[n] < offset)
node              309 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = ifp->if_u1.if_root;
node              317 fs/xfs/libxfs/xfs_iext_tree.c 			if (xfs_iext_key_cmp(node, i, offset) > 0)
node              320 fs/xfs/libxfs/xfs_iext_tree.c 		node = node->ptrs[i - 1];
node              321 fs/xfs/libxfs/xfs_iext_tree.c 		if (!node)
node              325 fs/xfs/libxfs/xfs_iext_tree.c 	return node;
node              330 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node              336 fs/xfs/libxfs/xfs_iext_tree.c 		if (xfs_iext_key_cmp(node, i, offset) > 0)
node              345 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node              351 fs/xfs/libxfs/xfs_iext_tree.c 		if (xfs_iext_key_cmp(node, i, offset) > 0)
node              360 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node              366 fs/xfs/libxfs/xfs_iext_tree.c 		if (node->keys[i] == XFS_IEXT_KEY_INVALID)
node              401 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = kmem_zalloc(NODE_SIZE, KM_NOFS);
node              407 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[0] = xfs_iext_leaf_key(prev, 0);
node              408 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[0] = prev;
node              414 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[0] = prev->keys[0];
node              415 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[0] = prev;
node              419 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[i] = XFS_IEXT_KEY_INVALID;
node              421 fs/xfs/libxfs/xfs_iext_tree.c 	ifp->if_u1.if_root = node;
node              433 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = ifp->if_u1.if_root;
node              438 fs/xfs/libxfs/xfs_iext_tree.c 			if (i > 0 && xfs_iext_key_cmp(node, i, old_offset) > 0)
node              440 fs/xfs/libxfs/xfs_iext_tree.c 			if (node->keys[i] == old_offset)
node              441 fs/xfs/libxfs/xfs_iext_tree.c 				node->keys[i] = new_offset;
node              443 fs/xfs/libxfs/xfs_iext_tree.c 		node = node->ptrs[i - 1];
node              444 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(node);
node              447 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(node == ptr);
node              456 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node = *nodep;
node              472 fs/xfs/libxfs/xfs_iext_tree.c 		new->keys[i] = node->keys[nr_keep + i];
node              473 fs/xfs/libxfs/xfs_iext_tree.c 		new->ptrs[i] = node->ptrs[nr_keep + i];
node              475 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[nr_keep + i] = XFS_IEXT_KEY_INVALID;
node              476 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[nr_keep + i] = NULL;
node              499 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node, *new;
node              507 fs/xfs/libxfs/xfs_iext_tree.c 	node = xfs_iext_find_level(ifp, offset, level);
node              508 fs/xfs/libxfs/xfs_iext_tree.c 	pos = xfs_iext_node_insert_pos(node, offset);
node              509 fs/xfs/libxfs/xfs_iext_tree.c 	nr_entries = xfs_iext_node_nr_entries(node, pos);
node              511 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(pos >= nr_entries || xfs_iext_key_cmp(node, pos, offset) != 0);
node              515 fs/xfs/libxfs/xfs_iext_tree.c 		new = xfs_iext_split_node(&node, &pos, &nr_entries);
node              521 fs/xfs/libxfs/xfs_iext_tree.c 	if (node != new && pos == 0 && nr_entries > 0)
node              522 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_update_node(ifp, node->keys[0], offset, level, node);
node              525 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[i] = node->keys[i - 1];
node              526 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[i] = node->ptrs[i - 1];
node              528 fs/xfs/libxfs/xfs_iext_tree.c 	node->keys[pos] = offset;
node              529 fs/xfs/libxfs/xfs_iext_tree.c 	node->ptrs[pos] = ptr;
node              675 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node              684 fs/xfs/libxfs/xfs_iext_tree.c 		return node;
node              692 fs/xfs/libxfs/xfs_iext_tree.c 				prev->keys[nr_prev + i] = node->keys[i];
node              693 fs/xfs/libxfs/xfs_iext_tree.c 				prev->ptrs[nr_prev + i] = node->ptrs[i];
node              695 fs/xfs/libxfs/xfs_iext_tree.c 			return node;
node              710 fs/xfs/libxfs/xfs_iext_tree.c 				node->keys[nr_entries + i] = next->keys[i];
node              711 fs/xfs/libxfs/xfs_iext_tree.c 				node->ptrs[nr_entries + i] = next->ptrs[i];
node              728 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node, *parent;
node              732 fs/xfs/libxfs/xfs_iext_tree.c 	node = xfs_iext_find_level(ifp, offset, level);
node              733 fs/xfs/libxfs/xfs_iext_tree.c 	pos = xfs_iext_node_pos(node, offset);
node              735 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(node->ptrs[pos]);
node              736 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(node->ptrs[pos] == victim);
node              739 fs/xfs/libxfs/xfs_iext_tree.c 	nr_entries = xfs_iext_node_nr_entries(node, pos) - 1;
node              740 fs/xfs/libxfs/xfs_iext_tree.c 	offset = node->keys[0];
node              742 fs/xfs/libxfs/xfs_iext_tree.c 		node->keys[i] = node->keys[i + 1];
node              743 fs/xfs/libxfs/xfs_iext_tree.c 		node->ptrs[i] = node->ptrs[i + 1];
node              745 fs/xfs/libxfs/xfs_iext_tree.c 	node->keys[nr_entries] = XFS_IEXT_KEY_INVALID;
node              746 fs/xfs/libxfs/xfs_iext_tree.c 	node->ptrs[nr_entries] = NULL;
node              749 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_update_node(ifp, offset, node->keys[0], level, node);
node              750 fs/xfs/libxfs/xfs_iext_tree.c 		offset = node->keys[0];
node              767 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(parent->ptrs[pos] == node);
node              769 fs/xfs/libxfs/xfs_iext_tree.c 		node = xfs_iext_rebalance_node(parent, &pos, node, nr_entries);
node              770 fs/xfs/libxfs/xfs_iext_tree.c 		if (node) {
node              771 fs/xfs/libxfs/xfs_iext_tree.c 			victim = node;
node              772 fs/xfs/libxfs/xfs_iext_tree.c 			node = parent;
node              780 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(node == ifp->if_u1.if_root);
node              781 fs/xfs/libxfs/xfs_iext_tree.c 		ifp->if_u1.if_root = node->ptrs[0];
node              783 fs/xfs/libxfs/xfs_iext_tree.c 		kmem_free(node);
node             1025 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_node	*node,
node             1032 fs/xfs/libxfs/xfs_iext_tree.c 			if (node->keys[i] == XFS_IEXT_KEY_INVALID)
node             1034 fs/xfs/libxfs/xfs_iext_tree.c 			xfs_iext_destroy_node(node->ptrs[i], level - 1);
node             1038 fs/xfs/libxfs/xfs_iext_tree.c 	kmem_free(node);
node              333 fs/xfs/scrub/dabtree.c 	struct xfs_da_intnode		*node;
node              435 fs/xfs/scrub/dabtree.c 		node = blk->bp->b_addr;
node              436 fs/xfs/scrub/dabtree.c 		ip->d_ops->node_hdr_from_disk(&nodehdr, node);
node              437 fs/xfs/scrub/dabtree.c 		btree = ip->d_ops->node_tree_p(node);
node              199 fs/xfs/xfs_attr_inactive.c 	xfs_da_intnode_t *node;
node              215 fs/xfs/xfs_attr_inactive.c 	node = bp->b_addr;
node              216 fs/xfs/xfs_attr_inactive.c 	dp->d_ops->node_hdr_from_disk(&ichdr, node);
node              222 fs/xfs/xfs_attr_inactive.c 	btree = dp->d_ops->node_tree_p(node);
node              286 fs/xfs/xfs_attr_inactive.c 			node = bp->b_addr;
node              287 fs/xfs/xfs_attr_inactive.c 			btree = dp->d_ops->node_tree_p(node);
node              202 fs/xfs/xfs_attr_list.c 	struct xfs_da_intnode		*node;
node              220 fs/xfs/xfs_attr_list.c 		node = bp->b_addr;
node              221 fs/xfs/xfs_attr_list.c 		magic = be16_to_cpu(node->hdr.info.magic);
node              228 fs/xfs/xfs_attr_list.c 					node, sizeof(*node));
node              232 fs/xfs/xfs_attr_list.c 		dp->d_ops->node_hdr_from_disk(&nodehdr, node);
node              246 fs/xfs/xfs_attr_list.c 		btree = dp->d_ops->node_tree_p(node);
node              283 fs/xfs/xfs_attr_list.c 	struct xfs_da_intnode		*node;
node              308 fs/xfs/xfs_attr_list.c 			node = bp->b_addr;
node              309 fs/xfs/xfs_attr_list.c 			switch (be16_to_cpu(node->hdr.info.magic)) {
node              328 include/acpi/acpi_bus.h 	struct list_head node;
node              356 include/acpi/acpi_bus.h 	struct list_head node;
node              476 include/acpi/acpi_bus.h 	struct list_head node;
node             1147 include/acpi/actbl2.h #define ACPI_NFIT_BUILD_DEVICE_HANDLE(dimm, channel, memory, socket, node) \
node             1152 include/acpi/actbl2.h 	((node)    << ACPI_NFIT_NODE_ID_OFFSET))
node               38 include/asm-generic/topology.h #define set_numa_node(node)
node               41 include/asm-generic/topology.h #define set_cpu_numa_node(cpu, node)
node               49 include/asm-generic/topology.h     #define cpumask_of_node(node)	((node) == 0 ? cpu_online_mask : cpu_none_mask)
node               51 include/asm-generic/topology.h     #define cpumask_of_node(node)	((void)(node), cpu_online_mask)
node               69 include/asm-generic/topology.h #define set_numa_mem(node)
node               72 include/asm-generic/topology.h #define set_cpu_numa_mem(cpu, node)
node              118 include/clocksource/timer-ti-dm.h 	struct list_head node;
node               60 include/drm/drm_flip_work.h 	struct list_head node;
node              106 include/drm/drm_mipi_dsi.h struct mipi_dsi_host *of_find_mipi_dsi_host_by_node(struct device_node *node);
node              156 include/drm/drm_mipi_dsi.h 	struct device_node *node;
node              196 include/drm/drm_mm.h 	void (*color_adjust)(const struct drm_mm_node *node,
node              254 include/drm/drm_mm.h static inline bool drm_mm_node_allocated(const struct drm_mm_node *node)
node              256 include/drm/drm_mm.h 	return node->allocated;
node              289 include/drm/drm_mm.h static inline bool drm_mm_hole_follows(const struct drm_mm_node *node)
node              291 include/drm/drm_mm.h 	return node->hole_size;
node              402 include/drm/drm_mm.h int drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node);
node              404 include/drm/drm_mm.h 				struct drm_mm_node *node,
node              430 include/drm/drm_mm.h drm_mm_insert_node_generic(struct drm_mm *mm, struct drm_mm_node *node,
node              435 include/drm/drm_mm.h 	return drm_mm_insert_node_in_range(mm, node,
node              455 include/drm/drm_mm.h 				     struct drm_mm_node *node,
node              458 include/drm/drm_mm.h 	return drm_mm_insert_node_generic(mm, node, size, 0, 0, 0);
node              461 include/drm/drm_mm.h void drm_mm_remove_node(struct drm_mm_node *node);
node              542 include/drm/drm_mm.h 			   struct drm_mm_node *node);
node              544 include/drm/drm_mm.h 			      struct drm_mm_node *node);
node               27 include/drm/drm_of.h 				struct device_node *node);
node               31 include/drm/drm_of.h int drm_of_encoder_active_endpoint(struct device_node *node,
node               55 include/drm/drm_of.h 			   struct device_node *node)
node               67 include/drm/drm_of.h static inline int drm_of_encoder_active_endpoint(struct device_node *node,
node              110 include/drm/drm_of.h static inline int drm_of_encoder_active_endpoint_id(struct device_node *node,
node              114 include/drm/drm_of.h 	int ret = drm_of_encoder_active_endpoint(node, encoder,
node              120 include/drm/drm_of.h static inline int drm_of_encoder_active_port_id(struct device_node *node,
node              124 include/drm/drm_of.h 	int ret = drm_of_encoder_active_endpoint(node, encoder,
node               72 include/drm/drm_vma_manager.h 		       struct drm_vma_offset_node *node, unsigned long pages);
node               74 include/drm/drm_vma_manager.h 			   struct drm_vma_offset_node *node);
node               76 include/drm/drm_vma_manager.h int drm_vma_node_allow(struct drm_vma_offset_node *node, struct drm_file *tag);
node               77 include/drm/drm_vma_manager.h void drm_vma_node_revoke(struct drm_vma_offset_node *node,
node               79 include/drm/drm_vma_manager.h bool drm_vma_node_is_allowed(struct drm_vma_offset_node *node,
node               99 include/drm/drm_vma_manager.h 	struct drm_vma_offset_node *node;
node              101 include/drm/drm_vma_manager.h 	node = drm_vma_offset_lookup_locked(mgr, start, pages);
node              102 include/drm/drm_vma_manager.h 	return (node && node->vm_node.start == start) ? node : NULL;
node              147 include/drm/drm_vma_manager.h static inline void drm_vma_node_reset(struct drm_vma_offset_node *node)
node              149 include/drm/drm_vma_manager.h 	memset(node, 0, sizeof(*node));
node              150 include/drm/drm_vma_manager.h 	node->vm_files = RB_ROOT;
node              151 include/drm/drm_vma_manager.h 	rwlock_init(&node->vm_lock);
node              168 include/drm/drm_vma_manager.h static inline unsigned long drm_vma_node_start(const struct drm_vma_offset_node *node)
node              170 include/drm/drm_vma_manager.h 	return node->vm_node.start;
node              185 include/drm/drm_vma_manager.h static inline unsigned long drm_vma_node_size(struct drm_vma_offset_node *node)
node              187 include/drm/drm_vma_manager.h 	return node->vm_node.size;
node              202 include/drm/drm_vma_manager.h static inline __u64 drm_vma_node_offset_addr(struct drm_vma_offset_node *node)
node              204 include/drm/drm_vma_manager.h 	return ((__u64)node->vm_node.start) << PAGE_SHIFT;
node              219 include/drm/drm_vma_manager.h static inline void drm_vma_node_unmap(struct drm_vma_offset_node *node,
node              222 include/drm/drm_vma_manager.h 	if (drm_mm_node_allocated(&node->vm_node))
node              224 include/drm/drm_vma_manager.h 				    drm_vma_node_offset_addr(node),
node              225 include/drm/drm_vma_manager.h 				    drm_vma_node_size(node) << PAGE_SHIFT, 1);
node              240 include/drm/drm_vma_manager.h static inline int drm_vma_node_verify_access(struct drm_vma_offset_node *node,
node              243 include/drm/drm_vma_manager.h 	return drm_vma_node_is_allowed(node, tag) ? 0 : -EACCES;
node              188 include/drm/gpu_scheduler.h 	struct list_head		node;
node               65 include/drm/spsc_queue.h static inline bool spsc_queue_push(struct spsc_queue *queue, struct spsc_node *node)
node               69 include/drm/spsc_queue.h 	node->next = NULL;
node               73 include/drm/spsc_queue.h 	tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next);
node               74 include/drm/spsc_queue.h 	WRITE_ONCE(*tail, node);
node               91 include/drm/spsc_queue.h 	struct spsc_node *next, *node;
node               96 include/drm/spsc_queue.h 	node = READ_ONCE(queue->head);
node               98 include/drm/spsc_queue.h 	if (!node)
node              101 include/drm/spsc_queue.h 	next = READ_ONCE(node->next);
node              108 include/drm/spsc_queue.h 				(long)&node->next, (long) &queue->head) != (long)&node->next) {
node              112 include/drm/spsc_queue.h 			} while (unlikely(!(queue->head = READ_ONCE(node->next))));
node              117 include/drm/spsc_queue.h 	return node;
node               43 include/linux/alarmtimer.h 	struct timerqueue_node	node;
node               39 include/linux/async.h 				   int node);
node               41 include/linux/async.h 					  int node,
node               31 include/linux/async_tx.h 	struct list_head node;
node               17 include/linux/attribute_container.h 	struct list_head	node;
node              212 include/linux/backlight.h struct backlight_device *of_find_backlight_by_node(struct device_node *node);
node              215 include/linux/backlight.h of_find_backlight_by_node(struct device_node *node)
node              532 include/linux/blkdev.h 	int			node;
node               50 include/linux/bpf-cgroup.h 	struct rb_node node;
node               55 include/linux/bpf-cgroup.h 	struct list_head node;
node               36 include/linux/btree.h 	unsigned long *node;
node               54 include/linux/ceph/mon_client.h 	struct rb_node node;
node              267 include/linux/ceph/osd_client.h 	struct rb_node node;            /* osd */
node              296 include/linux/ceph/osd_client.h 	struct rb_node node;
node               45 include/linux/ceph/osdmap.h 	struct rb_node node;
node              141 include/linux/ceph/osdmap.h 	struct rb_node node;
node               13 include/linux/ceph/string_table.h 		struct rb_node node;
node              571 include/linux/cgroup-defs.h 	struct list_head node;		/* anchored at ss->cfts */
node              587 include/linux/clk-provider.h void of_fixed_factor_clk_setup(struct device_node *node);
node              798 include/linux/clk-provider.h int __must_check of_clk_hw_register(struct device_node *node, struct clk_hw *hw);
node               57 include/linux/clk.h 	struct list_head		node;
node               15 include/linux/clk/clk-conf.h int of_clk_set_defaults(struct device_node *node, bool clk_supplier);
node               17 include/linux/clk/clk-conf.h static inline int of_clk_set_defaults(struct device_node *node,
node              151 include/linux/clk/ti.h 	struct list_head	node;
node               19 include/linux/clkdev.h 	struct list_head	node;
node              241 include/linux/compaction.h struct node;
node              243 include/linux/compaction.h extern int compaction_register_node(struct node *node);
node              244 include/linux/compaction.h extern void compaction_unregister_node(struct node *node);
node              248 include/linux/compaction.h static inline int compaction_register_node(struct node *node)
node              253 include/linux/compaction.h static inline void compaction_unregister_node(struct node *node)
node              432 include/linux/cper.h 	u16	node;
node              452 include/linux/cper.h 	u16	node;
node              472 include/linux/cper.h 	u16	node;
node              269 include/linux/cpuhotplug.h 							 struct hlist_node *node),
node              271 include/linux/cpuhotplug.h 							  struct hlist_node *node))
node              278 include/linux/cpuhotplug.h int __cpuhp_state_add_instance(enum cpuhp_state state, struct hlist_node *node,
node              281 include/linux/cpuhotplug.h 					  struct hlist_node *node, bool invoke);
node              294 include/linux/cpuhotplug.h 					   struct hlist_node *node)
node              296 include/linux/cpuhotplug.h 	return __cpuhp_state_add_instance(state, node, true);
node              309 include/linux/cpuhotplug.h 						   struct hlist_node *node)
node              311 include/linux/cpuhotplug.h 	return __cpuhp_state_add_instance(state, node, false);
node              316 include/linux/cpuhotplug.h 					    struct hlist_node *node)
node              318 include/linux/cpuhotplug.h 	return __cpuhp_state_add_instance_cpuslocked(state, node, false);
node              365 include/linux/cpuhotplug.h 				  struct hlist_node *node, bool invoke);
node              377 include/linux/cpuhotplug.h 					      struct hlist_node *node)
node              379 include/linux/cpuhotplug.h 	return __cpuhp_state_remove_instance(state, node, true);
node              391 include/linux/cpuhotplug.h 						      struct hlist_node *node)
node              393 include/linux/cpuhotplug.h 	return __cpuhp_state_remove_instance(state, node, false);
node              192 include/linux/cpumask.h static inline unsigned int cpumask_local_spread(unsigned int i, int node)
node              247 include/linux/cpumask.h unsigned int cpumask_local_spread(unsigned int i, int node);
node              737 include/linux/cpumask.h bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node);
node              739 include/linux/cpumask.h bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node);
node              762 include/linux/cpumask.h 					  int node)
node              774 include/linux/cpumask.h 					  int node)
node               67 include/linux/cpuset.h extern bool __cpuset_node_allowed(int node, gfp_t gfp_mask);
node               69 include/linux/cpuset.h static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask)
node               72 include/linux/cpuset.h 		return __cpuset_node_allowed(node, gfp_mask);
node              207 include/linux/cpuset.h static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask)
node              196 include/linux/crush/crush.h 	struct rb_node node;
node               20 include/linux/dca.h 	struct list_head	node;
node               27 include/linux/dca.h 	struct list_head	node;
node               29 include/linux/debugobjects.h 	struct hlist_node	node;
node               27 include/linux/devfreq-event.h 	struct list_head node;
node              149 include/linux/devfreq.h 	struct list_head node;
node              531 include/linux/device.h 	struct list_head node;
node              761 include/linux/device.h 	struct list_head	node;
node              951 include/linux/device.h 			    struct device_node *node, int index,
node             1356 include/linux/device.h static inline void set_dev_node(struct device *dev, int node)
node             1358 include/linux/device.h 	dev->numa_node = node;
node             1365 include/linux/device.h static inline void set_dev_node(struct device *dev, int node)
node               87 include/linux/dio.h 	struct list_head node;
node              343 include/linux/dma-buf.h 	struct list_head node;
node              117 include/linux/dma-fence.h 	struct list_head node;
node              280 include/linux/dmar.h extern int dmar_alloc_hwirq(int id, int node, void *arg);
node             1324 include/linux/efi.h struct device *efi_get_device_by_path(struct efi_dev_path **node, size_t *len);
node              164 include/linux/elevator.h #define rb_entry_rq(node)	rb_entry((node), struct request, rb_node)
node              100 include/linux/enclosure.h 	struct list_head node;
node              225 include/linux/extcon.h extern struct extcon_dev *extcon_find_edev_by_node(struct device_node *node);
node              279 include/linux/extcon.h static inline struct extcon_dev *extcon_find_edev_by_node(struct device_node *node)
node              441 include/linux/fb.h 	int node;
node              809 include/linux/fb.h 	pr_err("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__)
node              811 include/linux/fb.h 	pr_notice("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__)
node              813 include/linux/fb.h 	pr_warn("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__)
node              815 include/linux/fb.h 	pr_info("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__)
node              817 include/linux/fb.h 	pr_debug("fb%d: " fmt, (fb_info)->node, ##__VA_ARGS__)
node              185 include/linux/firewire.h 	struct fw_node *node;
node              288 include/linux/firmware/xlnx-zynqmp.h 	int (*request_node)(const u32 node,
node              292 include/linux/firmware/xlnx-zynqmp.h 	int (*release_node)(const u32 node);
node              293 include/linux/firmware/xlnx-zynqmp.h 	int (*set_requirement)(const u32 node,
node               41 include/linux/fpga/fpga-bridge.h 	struct list_head node;
node               47 include/linux/fpga/fpga-bridge.h struct fpga_bridge *of_fpga_bridge_get(struct device_node *node,
node              188 include/linux/fpga/fpga-mgr.h struct fpga_manager *of_fpga_mgr_get(struct device_node *node);
node              743 include/linux/fs.h static inline unsigned int i_blocksize(const struct inode *node)
node              745 include/linux/fs.h 	return (1 << node->i_blkbits);
node               87 include/linux/fsl/mc.h 	struct list_head node;
node               47 include/linux/gameport.h 	struct list_head node;
node              536 include/linux/gfp.h 			int node, bool hugepage);
node              542 include/linux/gfp.h #define alloc_pages_vma(gfp_mask, order, vma, addr, node, false)\
node              550 include/linux/gfp.h #define alloc_page_vma_node(gfp_mask, vma, addr, node)		\
node              551 include/linux/gfp.h 	alloc_pages_vma(gfp_mask, 0, vma, addr, node, false)
node              567 include/linux/gpio/consumer.h struct gpio_desc *gpiod_get_from_of_node(struct device_node *node,
node              577 include/linux/gpio/consumer.h struct gpio_desc *gpiod_get_from_of_node(struct device_node *node,
node              591 include/linux/gpio/consumer.h 					      struct device_node *node,
node              602 include/linux/gpio/consumer.h 					      struct device_node *node,
node              671 include/linux/gpio/driver.h 	struct list_head node;
node               60 include/linux/hashtable.h #define hash_add(hashtable, node, key)						\
node               61 include/linux/hashtable.h 	hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))])
node               69 include/linux/hashtable.h #define hash_add_rcu(hashtable, node, key)					\
node               70 include/linux/hashtable.h 	hlist_add_head_rcu(node, &hashtable[hash_min(key, HASH_BITS(hashtable))])
node               76 include/linux/hashtable.h static inline bool hash_hashed(struct hlist_node *node)
node               78 include/linux/hashtable.h 	return !hlist_unhashed(node);
node              105 include/linux/hashtable.h static inline void hash_del(struct hlist_node *node)
node              107 include/linux/hashtable.h 	hlist_del_init(node);
node              114 include/linux/hashtable.h static inline void hash_del_rcu(struct hlist_node *node)
node              116 include/linux/hashtable.h 	hlist_del_init_rcu(node);
node               34 include/linux/hid-debug.h 	struct list_head node;
node               33 include/linux/hidraw.h 	struct list_head node;
node              118 include/linux/hrtimer.h 	struct timerqueue_node		node;
node              241 include/linux/hrtimer.h 	timer->node.expires = time;
node              248 include/linux/hrtimer.h 	timer->node.expires = ktime_add_safe(time, delta);
node              254 include/linux/hrtimer.h 	timer->node.expires = ktime_add_safe(time, ns_to_ktime(delta));
node              259 include/linux/hrtimer.h 	timer->node.expires = tv64;
node              265 include/linux/hrtimer.h 	timer->node.expires = ktime_add_safe(timer->node.expires, time);
node              271 include/linux/hrtimer.h 	timer->node.expires = ktime_add_ns(timer->node.expires, ns);
node              277 include/linux/hrtimer.h 	return timer->node.expires;
node              287 include/linux/hrtimer.h 	return timer->node.expires;
node              296 include/linux/hrtimer.h 	return ktime_to_ns(timer->node.expires);
node              301 include/linux/hrtimer.h 	return ktime_sub(timer->node.expires, timer->base->get_time());
node              335 include/linux/hrtimer.h 	ktime_t rem = ktime_sub(timer->node.expires, now);
node              929 include/linux/i2c.h extern struct i2c_client *of_find_i2c_device_by_node(struct device_node *node);
node              932 include/linux/i2c.h extern struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node);
node              935 include/linux/i2c.h struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node);
node              941 include/linux/i2c.h int of_i2c_get_board_info(struct device *dev, struct device_node *node,
node              946 include/linux/i2c.h static inline struct i2c_client *of_find_i2c_device_by_node(struct device_node *node)
node              951 include/linux/i2c.h static inline struct i2c_adapter *of_find_i2c_adapter_by_node(struct device_node *node)
node              956 include/linux/i2c.h static inline struct i2c_adapter *of_get_i2c_adapter_by_node(struct device_node *node)
node              969 include/linux/i2c.h 					struct device_node *node,
node               42 include/linux/i3c/master.h 	struct list_head node;
node               64 include/linux/i3c/master.h 	struct list_head node;
node              180 include/linux/i3c/master.h 	struct list_head node;
node              507 include/linux/i3c/master.h 	list_for_each_entry(dev, &(bus)->devs.i2c, common.node)
node              518 include/linux/i3c/master.h 	list_for_each_entry(dev, &(bus)->devs.i3c, common.node)
node              195 include/linux/input.h 	struct list_head	node;
node              322 include/linux/input.h 	struct list_head	node;
node              557 include/linux/intel-iommu.h 	int		node;
node              654 include/linux/intel-iommu.h void *alloc_pgtable_page(int node);
node               50 include/linux/interconnect-provider.h 	int (*aggregate)(struct icc_node *node, u32 tag, u32 avg_bw,
node               52 include/linux/interconnect-provider.h 	void (*pre_aggregate)(struct icc_node *node);
node               97 include/linux/interconnect-provider.h int icc_link_create(struct icc_node *node, const int dst_id);
node               99 include/linux/interconnect-provider.h void icc_node_add(struct icc_node *node, struct icc_provider *provider);
node              100 include/linux/interconnect-provider.h void icc_node_del(struct icc_node *node);
node              115 include/linux/interconnect-provider.h static inline int icc_link_create(struct icc_node *node, const int dst_id)
node              125 include/linux/interconnect-provider.h void icc_node_add(struct icc_node *node, struct icc_provider *provider)
node              129 include/linux/interconnect-provider.h void icc_node_del(struct icc_node *node)
node               15 include/linux/interval_tree.h interval_tree_insert(struct interval_tree_node *node,
node               19 include/linux/interval_tree.h interval_tree_remove(struct interval_tree_node *node,
node               27 include/linux/interval_tree.h interval_tree_iter_next(struct interval_tree_node *node,
node               38 include/linux/interval_tree_generic.h ITSTATIC void ITPREFIX ## _insert(ITSTRUCT *node,			      \
node               42 include/linux/interval_tree_generic.h 	ITTYPE start = ITSTART(node), last = ITLAST(node);		      \
node               59 include/linux/interval_tree_generic.h 	node->ITSUBTREE = last;						      \
node               60 include/linux/interval_tree_generic.h 	rb_link_node(&node->ITRB, rb_parent, link);			      \
node               61 include/linux/interval_tree_generic.h 	rb_insert_augmented_cached(&node->ITRB, root,			      \
node               65 include/linux/interval_tree_generic.h ITSTATIC void ITPREFIX ## _remove(ITSTRUCT *node,			      \
node               68 include/linux/interval_tree_generic.h 	rb_erase_augmented_cached(&node->ITRB, root, &ITPREFIX ## _augment);  \
node               81 include/linux/interval_tree_generic.h ITPREFIX ## _subtree_search(ITSTRUCT *node, ITTYPE start, ITTYPE last)	      \
node               88 include/linux/interval_tree_generic.h 		if (node->ITRB.rb_left) {				      \
node               89 include/linux/interval_tree_generic.h 			ITSTRUCT *left = rb_entry(node->ITRB.rb_left,	      \
node              100 include/linux/interval_tree_generic.h 				node = left;				      \
node              104 include/linux/interval_tree_generic.h 		if (ITSTART(node) <= last) {		/* Cond1 */	      \
node              105 include/linux/interval_tree_generic.h 			if (start <= ITLAST(node))	/* Cond2 */	      \
node              106 include/linux/interval_tree_generic.h 				return node;	/* node is leftmost match */  \
node              107 include/linux/interval_tree_generic.h 			if (node->ITRB.rb_right) {			      \
node              108 include/linux/interval_tree_generic.h 				node = rb_entry(node->ITRB.rb_right,	      \
node              110 include/linux/interval_tree_generic.h 				if (start <= node->ITSUBTREE)		      \
node              122 include/linux/interval_tree_generic.h 	ITSTRUCT *node, *leftmost;					      \
node              140 include/linux/interval_tree_generic.h 	node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB);		      \
node              141 include/linux/interval_tree_generic.h 	if (node->ITSUBTREE < start)					      \
node              148 include/linux/interval_tree_generic.h 	return ITPREFIX ## _subtree_search(node, start, last);		      \
node              152 include/linux/interval_tree_generic.h ITPREFIX ## _iter_next(ITSTRUCT *node, ITTYPE start, ITTYPE last)	      \
node              154 include/linux/interval_tree_generic.h 	struct rb_node *rb = node->ITRB.rb_right, *prev;		      \
node              173 include/linux/interval_tree_generic.h 			rb = rb_parent(&node->ITRB);			      \
node              176 include/linux/interval_tree_generic.h 			prev = &node->ITRB;				      \
node              177 include/linux/interval_tree_generic.h 			node = rb_entry(rb, ITSTRUCT, ITRB);		      \
node              178 include/linux/interval_tree_generic.h 			rb = node->ITRB.rb_right;			      \
node              182 include/linux/interval_tree_generic.h 		if (last < ITSTART(node))		/* !Cond1 */	      \
node              184 include/linux/interval_tree_generic.h 		else if (start <= ITLAST(node))		/* Cond2 */	      \
node              185 include/linux/interval_tree_generic.h 			return node;					      \
node              152 include/linux/iocontext.h 				       gfp_t gfp_flags, int node);
node               20 include/linux/iova.h 	struct rb_node	node;
node              146 include/linux/irq.h 	unsigned int		node;
node              820 include/linux/irq.h 	return d->node;
node              868 include/linux/irq.h int __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node,
node              873 include/linux/irq.h 			   unsigned int cnt, int node, struct module *owner,
node              877 include/linux/irq.h #define irq_alloc_descs(irq, from, cnt, node)	\
node              878 include/linux/irq.h 	__irq_alloc_descs(irq, from, cnt, node, THIS_MODULE, NULL)
node              880 include/linux/irq.h #define irq_alloc_desc(node)			\
node              881 include/linux/irq.h 	irq_alloc_descs(-1, 0, 1, node)
node              883 include/linux/irq.h #define irq_alloc_desc_at(at, node)		\
node              884 include/linux/irq.h 	irq_alloc_descs(at, at, 1, node)
node              886 include/linux/irq.h #define irq_alloc_desc_from(from, node)		\
node              887 include/linux/irq.h 	irq_alloc_descs(-1, from, 1, node)
node              889 include/linux/irq.h #define irq_alloc_descs_from(from, cnt, node)	\
node              890 include/linux/irq.h 	irq_alloc_descs(-1, from, cnt, node)
node              892 include/linux/irq.h #define devm_irq_alloc_descs(dev, irq, from, cnt, node)		\
node              893 include/linux/irq.h 	__devm_irq_alloc_descs(dev, irq, from, cnt, node, THIS_MODULE, NULL)
node              895 include/linux/irq.h #define devm_irq_alloc_desc(dev, node)				\
node              896 include/linux/irq.h 	devm_irq_alloc_descs(dev, -1, 0, 1, node)
node              898 include/linux/irq.h #define devm_irq_alloc_desc_at(dev, at, node)			\
node              899 include/linux/irq.h 	devm_irq_alloc_descs(dev, at, at, 1, node)
node              901 include/linux/irq.h #define devm_irq_alloc_desc_from(dev, from, node)		\
node              902 include/linux/irq.h 	devm_irq_alloc_descs(dev, -1, from, 1, node)
node              904 include/linux/irq.h #define devm_irq_alloc_descs_from(dev, from, cnt, node)		\
node              905 include/linux/irq.h 	devm_irq_alloc_descs(dev, -1, from, cnt, node)
node              914 include/linux/irq.h unsigned int irq_alloc_hwirqs(int cnt, int node);
node              915 include/linux/irq.h static inline unsigned int irq_alloc_hwirq(int node)
node              917 include/linux/irq.h 	return irq_alloc_hwirqs(1, node);
node              924 include/linux/irq.h int arch_setup_hwirq(unsigned int irq, int node);
node               46 include/linux/irqbypass.h 	struct list_head node;
node               72 include/linux/irqbypass.h 	struct list_head node;
node              146 include/linux/irqchip/arm-gic.h int gic_of_init(struct device_node *node, struct device_node *parent);
node               20 include/linux/irqchip/arm-vic.h 		u32 vic_sources, u32 resume_sources, struct device_node *node);
node               10 include/linux/irqchip/versatile-fpga.h 		struct device_node *node);
node               11 include/linux/irqchip/versatile-fpga.h int fpga_irq_of_init(struct device_node *node,
node              104 include/linux/irqdomain.h 	int (*match)(struct irq_domain *d, struct device_node *node,
node              110 include/linux/irqdomain.h 	int (*xlate)(struct irq_domain *d, struct device_node *node,
node              278 include/linux/irqdomain.h 				  irq_hw_number_t hwirq, int node,
node              281 include/linux/irqdomain.h static inline struct fwnode_handle *of_node_to_fwnode(struct device_node *node)
node              283 include/linux/irqdomain.h 	return node ? &node->fwnode : NULL;
node              307 include/linux/irqdomain.h static inline struct irq_domain *irq_find_matching_host(struct device_node *node,
node              310 include/linux/irqdomain.h 	return irq_find_matching_fwnode(of_node_to_fwnode(node), bus_token);
node              313 include/linux/irqdomain.h static inline struct irq_domain *irq_find_host(struct device_node *node)
node              317 include/linux/irqdomain.h 	d = irq_find_matching_host(node, DOMAIN_BUS_WIRED);
node              319 include/linux/irqdomain.h 		d = irq_find_matching_host(node, DOMAIN_BUS_ANY);
node              456 include/linux/irqdomain.h 					    struct device_node *node,
node              461 include/linux/irqdomain.h 					   of_node_to_fwnode(node),
node              466 include/linux/irqdomain.h 				   unsigned int nr_irqs, int node, void *arg,
node              474 include/linux/irqdomain.h 			unsigned int nr_irqs, int node, void *arg)
node              476 include/linux/irqdomain.h 	return __irq_domain_alloc_irqs(domain, -1, nr_irqs, node, arg, false,
node              541 include/linux/irqdomain.h 			unsigned int nr_irqs, int node, void *arg)
node               11 include/linux/kthread.h 					   int node,
node               96 include/linux/kthread.h 	struct list_head	node;
node              115 include/linux/kthread.h 	.node = LIST_HEAD_INIT((work).node),				\
node              160 include/linux/kthread.h 		INIT_LIST_HEAD(&(work)->node);				\
node              121 include/linux/leds.h 	struct list_head	 node;			/* LED Device list */
node              690 include/linux/lightnvm.h static inline struct nvm_dev *nvm_alloc_dev(int node)
node               53 include/linux/list_lru.h 	struct list_lru_node	*node;
node               73 include/linux/livepatch.h 	struct list_head node;
node              127 include/linux/livepatch.h 	struct list_head node;
node              166 include/linux/livepatch.h 	list_for_each_entry_safe(obj, tmp_obj, &patch->obj_list, node)
node              169 include/linux/livepatch.h 	list_for_each_entry(obj, &patch->obj_list, node)
node              177 include/linux/livepatch.h 	list_for_each_entry_safe(func, tmp_func, &obj->func_list, node)
node              180 include/linux/livepatch.h 	list_for_each_entry(func, &obj->func_list, node)
node              114 include/linux/llist.h #define llist_for_each(pos, node)			\
node              115 include/linux/llist.h 	for ((pos) = (node); pos; (pos) = (pos)->next)
node              133 include/linux/llist.h #define llist_for_each_safe(pos, n, node)			\
node              134 include/linux/llist.h 	for ((pos) = (node); (pos) && ((n) = (pos)->next, true); (pos) = (n))
node              151 include/linux/llist.h #define llist_for_each_entry(pos, node, member)				\
node              152 include/linux/llist.h 	for ((pos) = llist_entry((node), typeof(*(pos)), member);	\
node              173 include/linux/llist.h #define llist_for_each_entry_safe(pos, n, node, member)			       \
node              174 include/linux/llist.h 	for (pos = llist_entry((node), typeof(*pos), member);		       \
node              192 include/linux/llist.h static inline struct llist_node *llist_next(struct llist_node *node)
node              194 include/linux/llist.h 	return node->next;
node               86 include/linux/mailbox_controller.h 	struct list_head node;
node               90 include/linux/mfd/ipaq-micro.h 	struct list_head node;
node               84 include/linux/mfd/mc13xxx.h 	struct device_node *node;
node              136 include/linux/mfd/ucb1x00.h 	struct list_head	node;
node              152 include/linux/mfd/ucb1x00.h 	struct list_head	node;
node              127 include/linux/migrate.h 				  struct vm_area_struct *vma, int node);
node              134 include/linux/migrate.h 					 struct vm_area_struct *vma, int node)
node              145 include/linux/migrate.h 			struct page *page, int node);
node              151 include/linux/migrate.h 			struct page *page, int node)
node             1109 include/linux/mlx4/device.h int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf, int node);
node              935 include/linux/mlx5/driver.h 			struct mlx5_frag_buf *buf, int node);
node              940 include/linux/mlx5/driver.h 			     struct mlx5_frag_buf *buf, int node);
node              989 include/linux/mlx5/driver.h 		       int node);
node             8433 include/linux/mlx5/mlx5_ifc.h 	u8         node[0x8];
node              667 include/linux/mm.h extern void *kvmalloc_node(size_t size, gfp_t flags, int node);
node              672 include/linux/mm.h static inline void *kvzalloc_node(size_t size, gfp_t flags, int node)
node              674 include/linux/mm.h 	return kvmalloc_node(size, flags | __GFP_ZERO, node);
node             1283 include/linux/mm.h static inline void set_page_node(struct page *page, unsigned long node)
node             1286 include/linux/mm.h 	page->flags |= (node & NODES_MASK) << NODES_PGSHIFT;
node             1290 include/linux/mm.h 	unsigned long node, unsigned long pfn)
node             1293 include/linux/mm.h 	set_page_node(page, node);
node             2242 include/linux/mm.h void vma_interval_tree_insert(struct vm_area_struct *node,
node             2244 include/linux/mm.h void vma_interval_tree_insert_after(struct vm_area_struct *node,
node             2247 include/linux/mm.h void vma_interval_tree_remove(struct vm_area_struct *node,
node             2251 include/linux/mm.h struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node,
node             2258 include/linux/mm.h void anon_vma_interval_tree_insert(struct anon_vma_chain *node,
node             2260 include/linux/mm.h void anon_vma_interval_tree_remove(struct anon_vma_chain *node,
node             2266 include/linux/mm.h 	struct anon_vma_chain *node, unsigned long start, unsigned long last);
node             2268 include/linux/mm.h void anon_vma_interval_tree_verify(struct anon_vma_chain *node);
node             2775 include/linux/mm.h pgd_t *vmemmap_pgd_populate(unsigned long addr, int node);
node             2776 include/linux/mm.h p4d_t *vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node);
node             2777 include/linux/mm.h pud_t *vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node);
node             2778 include/linux/mm.h pmd_t *vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node);
node             2779 include/linux/mm.h pte_t *vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node);
node             2780 include/linux/mm.h void *vmemmap_alloc_block(unsigned long size, int node);
node             2782 include/linux/mm.h void *vmemmap_alloc_block_buf(unsigned long size, int node);
node             2786 include/linux/mm.h 			       int node);
node             2787 include/linux/mm.h int vmemmap_populate(unsigned long start, unsigned long end, int node,
node              438 include/linux/mmzone.h 	int node;
node              886 include/linux/mmzone.h 	return zone->node;
node              891 include/linux/mmzone.h 	zone->node = nid;
node              304 include/linux/module.h 	struct latch_tree_node node;
node              652 include/linux/netdevice.h static inline void netdev_queue_numa_node_write(struct netdev_queue *q, int node)
node              655 include/linux/netdevice.h 	q->numa_node = node;
node               98 include/linux/node.h extern struct node *node_devices[];
node               99 include/linux/node.h typedef  void (*node_registration_func_t)(struct node *);
node              112 include/linux/node.h extern void unregister_node(struct node *node);
node              181 include/linux/node.h #define to_node(device) container_of(device, struct node, dev)
node              127 include/linux/nodemask.h #define node_set(node, dst) __node_set((node), &(dst))
node              128 include/linux/nodemask.h static __always_inline void __node_set(int node, volatile nodemask_t *dstp)
node              130 include/linux/nodemask.h 	set_bit(node, dstp->bits);
node              133 include/linux/nodemask.h #define node_clear(node, dst) __node_clear((node), &(dst))
node              134 include/linux/nodemask.h static inline void __node_clear(int node, volatile nodemask_t *dstp)
node              136 include/linux/nodemask.h 	clear_bit(node, dstp->bits);
node              152 include/linux/nodemask.h #define node_isset(node, nodemask) test_bit((node), (nodemask).bits)
node              154 include/linux/nodemask.h #define node_test_and_set(node, nodemask) \
node              155 include/linux/nodemask.h 			__node_test_and_set((node), &(nodemask))
node              156 include/linux/nodemask.h static inline int __node_test_and_set(int node, nodemask_t *addr)
node              158 include/linux/nodemask.h 	return test_and_set_bit(node, addr->bits);
node              279 include/linux/nodemask.h int __next_node_in(int node, const nodemask_t *srcp);
node              281 include/linux/nodemask.h static inline void init_nodemask_of_node(nodemask_t *mask, int node)
node              284 include/linux/nodemask.h 	node_set(node, *mask);
node              287 include/linux/nodemask.h #define nodemask_of_node(node)						\
node              291 include/linux/nodemask.h 		m.bits[0] = 1UL << (node);				\
node              293 include/linux/nodemask.h 		init_nodemask_of_node(&m, (node));			\
node              378 include/linux/nodemask.h #define for_each_node_mask(node, mask)			\
node              379 include/linux/nodemask.h 	for ((node) = first_node(mask);			\
node              380 include/linux/nodemask.h 		(node) < MAX_NUMNODES;			\
node              381 include/linux/nodemask.h 		(node) = next_node((node), (mask)))
node              383 include/linux/nodemask.h #define for_each_node_mask(node, mask)			\
node              385 include/linux/nodemask.h 		for ((node) = 0; (node) < 1; (node)++)
node              413 include/linux/nodemask.h static inline int node_state(int node, enum node_states state)
node              415 include/linux/nodemask.h 	return node_isset(node, node_states[state]);
node              418 include/linux/nodemask.h static inline void node_set_state(int node, enum node_states state)
node              420 include/linux/nodemask.h 	__node_set(node, &node_states[state]);
node              423 include/linux/nodemask.h static inline void node_clear_state(int node, enum node_states state)
node              425 include/linux/nodemask.h 	__node_clear(node, &node_states[state]);
node              464 include/linux/nodemask.h static inline int node_state(int node, enum node_states state)
node              466 include/linux/nodemask.h 	return node == 0;
node              469 include/linux/nodemask.h static inline void node_set_state(int node, enum node_states state)
node              473 include/linux/nodemask.h static inline void node_clear_state(int node, enum node_states state)
node              482 include/linux/nodemask.h #define for_each_node_state(node, __state) \
node              483 include/linux/nodemask.h 	for ( (node) = 0; (node) == 0; (node) = 1)
node              491 include/linux/nodemask.h #define node_set_online(node)	   node_set_state((node), N_ONLINE)
node              492 include/linux/nodemask.h #define node_set_offline(node)	   node_clear_state((node), N_ONLINE)
node              510 include/linux/nodemask.h #define node_online(node)	node_state((node), N_ONLINE)
node              511 include/linux/nodemask.h #define node_possible(node)	node_state((node), N_POSSIBLE)
node              513 include/linux/nodemask.h #define for_each_node(node)	   for_each_node_state(node, N_POSSIBLE)
node              514 include/linux/nodemask.h #define for_each_online_node(node) for_each_node_state(node, N_ONLINE)
node               45 include/linux/nvmem-consumer.h 	struct list_head	node;
node               94 include/linux/nvmem-provider.h 	struct list_head	node;
node               94 include/linux/of.h 	struct device_node *node;
node              106 include/linux/of.h static inline void of_node_init(struct device_node *node)
node              109 include/linux/of.h 	kobject_init(&node->kobj, &of_node_ktype);
node              111 include/linux/of.h 	node->fwnode.ops = &of_fwnode_ops;
node              121 include/linux/of.h extern struct device_node *of_node_get(struct device_node *node);
node              122 include/linux/of.h extern void of_node_put(struct device_node *node);
node              125 include/linux/of.h static inline struct device_node *of_node_get(struct device_node *node)
node              127 include/linux/of.h 	return node;
node              129 include/linux/of.h static inline void of_node_put(struct device_node *node) { }
node              170 include/linux/of.h #define of_fwnode_handle(node)						\
node              172 include/linux/of.h 		typeof(node) __of_fwnode_handle_node = (node);		\
node              183 include/linux/of.h static inline bool of_node_is_root(const struct device_node *node)
node              185 include/linux/of.h 	return node && (node->parent == NULL);
node              286 include/linux/of.h extern struct device_node *of_get_parent(const struct device_node *node);
node              287 include/linux/of.h extern struct device_node *of_get_next_parent(struct device_node *node);
node              288 include/linux/of.h extern struct device_node *of_get_next_child(const struct device_node *node,
node              291 include/linux/of.h 	const struct device_node *node, struct device_node *prev);
node              295 include/linux/of.h extern struct device_node *of_get_child_by_name(const struct device_node *node,
node              349 include/linux/of.h extern const void *of_get_property(const struct device_node *node,
node              361 include/linux/of.h 	const struct of_device_id *matches, const struct device_node *node);
node              362 include/linux/of.h extern int of_modalias_node(struct device_node *node, char *modalias, int len);
node              626 include/linux/of.h static inline struct device_node *of_get_parent(const struct device_node *node)
node              632 include/linux/of.h 	const struct device_node *node, struct device_node *prev)
node              638 include/linux/of.h 	const struct device_node *node, struct device_node *prev)
node              649 include/linux/of.h #define of_fwnode_handle(node) NULL
node              663 include/linux/of.h 					const struct device_node *node,
node              750 include/linux/of.h static inline const void *of_get_property(const struct device_node *node,
node             1321 include/linux/of.h 	struct list_head node;
node               10 include/linux/of_address.h 	struct device_node *node;
node               52 include/linux/of_address.h 			struct device_node *node);
node               54 include/linux/of_address.h 			struct device_node *node);
node               89 include/linux/of_address.h 			struct device_node *node)
node               95 include/linux/of_address.h 			struct device_node *node)
node              122 include/linux/of_address.h void __iomem *of_iomap(struct device_node *node, int index);
node               39 include/linux/of_fdt.h extern u64 of_flat_dt_translate_address(unsigned long node);
node               45 include/linux/of_fdt.h extern int of_scan_flat_dt(int (*it)(unsigned long node, const char *uname,
node               48 include/linux/of_fdt.h extern int of_scan_flat_dt_subnodes(unsigned long node,
node               49 include/linux/of_fdt.h 				    int (*it)(unsigned long node,
node               53 include/linux/of_fdt.h extern int of_get_flat_dt_subnode_by_name(unsigned long node,
node               55 include/linux/of_fdt.h extern const void *of_get_flat_dt_prop(unsigned long node, const char *name,
node               57 include/linux/of_fdt.h extern int of_flat_dt_is_compatible(unsigned long node, const char *name);
node               59 include/linux/of_fdt.h extern uint32_t of_get_flat_dt_phandle(unsigned long node);
node               61 include/linux/of_fdt.h extern int early_init_dt_scan_chosen(unsigned long node, const char *uname,
node               63 include/linux/of_fdt.h extern int early_init_dt_scan_memory(unsigned long node, const char *uname,
node               68 include/linux/of_fdt.h extern void __init early_init_dt_scan_chosen_arch(unsigned long node);
node               76 include/linux/of_fdt.h extern int early_init_dt_scan_root(unsigned long node, const char *uname,
node               41 include/linux/of_graph.h int of_graph_parse_endpoint(const struct device_node *node,
node               44 include/linux/of_graph.h struct device_node *of_graph_get_port_by_id(struct device_node *node, u32 id);
node               50 include/linux/of_graph.h 					const struct device_node *node);
node               51 include/linux/of_graph.h struct device_node *of_graph_get_port_parent(struct device_node *node);
node               53 include/linux/of_graph.h 					const struct device_node *node);
node               54 include/linux/of_graph.h struct device_node *of_graph_get_remote_port(const struct device_node *node);
node               55 include/linux/of_graph.h struct device_node *of_graph_get_remote_node(const struct device_node *node,
node               59 include/linux/of_graph.h static inline int of_graph_parse_endpoint(const struct device_node *node,
node               71 include/linux/of_graph.h 					struct device_node *node, u32 id)
node               90 include/linux/of_graph.h 					const struct device_node *node)
node               96 include/linux/of_graph.h 	struct device_node *node)
node              102 include/linux/of_graph.h 					const struct device_node *node)
node              108 include/linux/of_graph.h 					const struct device_node *node)
node              113 include/linux/of_graph.h 					const struct device_node *node,
node              108 include/linux/of_irq.h extern unsigned int irq_of_parse_and_map(struct device_node *node, int index);
node               18 include/linux/of_pdt.h 	int (*nextprop)(phandle node, char *prev, char *buf);
node               21 include/linux/of_pdt.h 	int (*getproplen)(phandle node, const char *prop);
node               22 include/linux/of_pdt.h 	int (*getproperty)(phandle node, const char *prop, char *buf,
node               27 include/linux/of_pdt.h 	phandle (*getsibling)(phandle node);
node               30 include/linux/of_pdt.h 	int (*pkg2path)(phandle node, char *buf, const int buflen, int *len);
node               39 include/linux/of_reserved_mem.h void fdt_reserved_mem_save_node(unsigned long node, const char *uname,
node               51 include/linux/of_reserved_mem.h static inline void fdt_reserved_mem_save_node(unsigned long node,
node              571 include/linux/pci.h 	struct list_head node;		/* Node in list of buses */
node              826 include/linux/pci.h 	struct list_head	node;
node             1524 include/linux/pci.h 				      struct device_node *node,
node             1745 include/linux/pci.h 				      struct device_node *node,
node              107 include/linux/perf/arm_pmu.h 	struct hlist_node	node;
node              166 include/linux/phy/phy.h 	struct list_head node;
node              241 include/linux/phy/phy.h struct phy *phy_create(struct device *dev, struct device_node *node,
node              243 include/linux/phy/phy.h struct phy *devm_phy_create(struct device *dev, struct device_node *node,
node              441 include/linux/phy/phy.h 				     struct device_node *node,
node              448 include/linux/phy/phy.h 					  struct device_node *node,
node               59 include/linux/pinctrl/pinctrl.h 	struct list_head node;
node              110 include/linux/plist.h #define PLIST_NODE_INIT(node, __prio)			\
node              113 include/linux/plist.h 	.prio_list = LIST_HEAD_INIT((node).prio_list),	\
node              114 include/linux/plist.h 	.node_list = LIST_HEAD_INIT((node).node_list),	\
node              132 include/linux/plist.h static inline void plist_node_init(struct plist_node *node, int prio)
node              134 include/linux/plist.h 	node->prio = prio;
node              135 include/linux/plist.h 	INIT_LIST_HEAD(&node->prio_list);
node              136 include/linux/plist.h 	INIT_LIST_HEAD(&node->node_list);
node              139 include/linux/plist.h extern void plist_add(struct plist_node *node, struct plist_head *head);
node              140 include/linux/plist.h extern void plist_del(struct plist_node *node, struct plist_head *head);
node              142 include/linux/plist.h extern void plist_requeue(struct plist_node *node, struct plist_head *head);
node              219 include/linux/plist.h static inline int plist_node_empty(const struct plist_node *node)
node              221 include/linux/plist.h 	return list_empty(&node->node_list);
node               42 include/linux/pm_qos.h 	struct plist_node node;
node               48 include/linux/pm_qos.h 	struct list_head node;
node              114 include/linux/pm_qos.h int pm_qos_update_target(struct pm_qos_constraints *c, struct plist_node *node,
node               70 include/linux/posix-timers.h 	struct timerqueue_node	node;
node               81 include/linux/posix-timers.h 	return timerqueue_add(head, &ctmr->node);
node               87 include/linux/posix-timers.h 		timerqueue_del(ctmr->head, &ctmr->node);
node               94 include/linux/posix-timers.h 	return ctmr->node.expires;
node               99 include/linux/posix-timers.h 	ctmr->node.expires = exp;
node              152 include/linux/power/smartreflex.h 	struct list_head		node;
node               72 include/linux/powercap.h 	struct list_head node;
node              389 include/linux/property.h 	const struct software_node *node;
node              422 include/linux/property.h struct fwnode_handle *software_node_fwnode(const struct software_node *node);
node              431 include/linux/property.h int software_node_register(const struct software_node *node);
node              104 include/linux/psi_types.h 	struct list_head node;
node              210 include/linux/pxa2xx_ssp.h 	struct list_head	node;
node               98 include/linux/radix-tree.h 	struct radix_tree_node *node;
node               43 include/linux/rbtree.h #define RB_EMPTY_NODE(node)  \
node               44 include/linux/rbtree.h 	((node)->__rb_parent_color == (unsigned long)(node))
node               45 include/linux/rbtree.h #define RB_CLEAR_NODE(node)  \
node               46 include/linux/rbtree.h 	((node)->__rb_parent_color = (unsigned long)(node))
node               69 include/linux/rbtree.h static inline void rb_link_node(struct rb_node *node, struct rb_node *parent,
node               72 include/linux/rbtree.h 	node->__rb_parent_color = (unsigned long)parent;
node               73 include/linux/rbtree.h 	node->rb_left = node->rb_right = NULL;
node               75 include/linux/rbtree.h 	*rb_link = node;
node               78 include/linux/rbtree.h static inline void rb_link_node_rcu(struct rb_node *node, struct rb_node *parent,
node               81 include/linux/rbtree.h 	node->__rb_parent_color = (unsigned long)parent;
node               82 include/linux/rbtree.h 	node->rb_left = node->rb_right = NULL;
node               84 include/linux/rbtree.h 	rcu_assign_pointer(*rb_link, node);
node              135 include/linux/rbtree.h static inline void rb_insert_color_cached(struct rb_node *node,
node              140 include/linux/rbtree.h 		root->rb_leftmost = node;
node              141 include/linux/rbtree.h 	rb_insert_color(node, &root->rb_root);
node              144 include/linux/rbtree.h static inline void rb_erase_cached(struct rb_node *node,
node              147 include/linux/rbtree.h 	if (root->rb_leftmost == node)
node              148 include/linux/rbtree.h 		root->rb_leftmost = rb_next(node);
node              149 include/linux/rbtree.h 	rb_erase(node, &root->rb_root);
node               28 include/linux/rbtree_augmented.h 	void (*propagate)(struct rb_node *node, struct rb_node *stop);
node               33 include/linux/rbtree_augmented.h extern void __rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node               47 include/linux/rbtree_augmented.h rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node               50 include/linux/rbtree_augmented.h 	__rb_insert_augmented(node, root, augment->rotate);
node               54 include/linux/rbtree_augmented.h rb_insert_augmented_cached(struct rb_node *node,
node               59 include/linux/rbtree_augmented.h 		root->rb_leftmost = node;
node               60 include/linux/rbtree_augmented.h 	rb_insert_augmented(node, &root->rb_root, augment);
node               80 include/linux/rbtree_augmented.h 		RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD);	\
node               81 include/linux/rbtree_augmented.h 		if (RBCOMPUTE(node, true))				\
node               83 include/linux/rbtree_augmented.h 		rb = rb_parent(&node->RBFIELD);				\
node              122 include/linux/rbtree_augmented.h static inline bool RBNAME ## _compute_max(RBSTRUCT *node, bool exit)	      \
node              125 include/linux/rbtree_augmented.h 	RBTYPE max = RBCOMPUTE(node);					      \
node              126 include/linux/rbtree_augmented.h 	if (node->RBFIELD.rb_left) {					      \
node              127 include/linux/rbtree_augmented.h 		child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD);   \
node              131 include/linux/rbtree_augmented.h 	if (node->RBFIELD.rb_right) {					      \
node              132 include/linux/rbtree_augmented.h 		child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD);  \
node              136 include/linux/rbtree_augmented.h 	if (exit && node->RBAUGMENTED == max)				      \
node              138 include/linux/rbtree_augmented.h 	node->RBAUGMENTED = max;					      \
node              198 include/linux/rbtree_augmented.h __rb_erase_augmented(struct rb_node *node, struct rb_root *root,
node              201 include/linux/rbtree_augmented.h 	struct rb_node *child = node->rb_right;
node              202 include/linux/rbtree_augmented.h 	struct rb_node *tmp = node->rb_left;
node              214 include/linux/rbtree_augmented.h 		pc = node->__rb_parent_color;
node              216 include/linux/rbtree_augmented.h 		__rb_change_child(node, child, parent, root);
node              225 include/linux/rbtree_augmented.h 		tmp->__rb_parent_color = pc = node->__rb_parent_color;
node              227 include/linux/rbtree_augmented.h 		__rb_change_child(node, tmp, parent, root);
node              247 include/linux/rbtree_augmented.h 			augment->copy(node, successor);
node              273 include/linux/rbtree_augmented.h 			augment->copy(node, successor);
node              277 include/linux/rbtree_augmented.h 		tmp = node->rb_left;
node              281 include/linux/rbtree_augmented.h 		pc = node->__rb_parent_color;
node              283 include/linux/rbtree_augmented.h 		__rb_change_child(node, successor, tmp, root);
node              302 include/linux/rbtree_augmented.h rb_erase_augmented(struct rb_node *node, struct rb_root *root,
node              305 include/linux/rbtree_augmented.h 	struct rb_node *rebalance = __rb_erase_augmented(node, root, augment);
node              311 include/linux/rbtree_augmented.h rb_erase_augmented_cached(struct rb_node *node, struct rb_root_cached *root,
node              314 include/linux/rbtree_augmented.h 	if (root->rb_leftmost == node)
node              315 include/linux/rbtree_augmented.h 		root->rb_leftmost = rb_next(node);
node              316 include/linux/rbtree_augmented.h 	rb_erase_augmented(node, &root->rb_root, augment);
node               41 include/linux/rbtree_latch.h 	struct rb_node node[2];
node               70 include/linux/rbtree_latch.h __lt_from_rb(struct rb_node *node, int idx)
node               72 include/linux/rbtree_latch.h 	return container_of(node, struct latch_tree_node, node[idx]);
node               81 include/linux/rbtree_latch.h 	struct rb_node *node = &ltn->node[idx];
node               95 include/linux/rbtree_latch.h 	rb_link_node_rcu(node, parent, link);
node               96 include/linux/rbtree_latch.h 	rb_insert_color(node, root);
node              102 include/linux/rbtree_latch.h 	rb_erase(&ltn->node[idx], &ltr->tree[idx]);
node              107 include/linux/rbtree_latch.h 	  int (*comp)(void *key, struct latch_tree_node *node))
node              109 include/linux/rbtree_latch.h 	struct rb_node *node = rcu_dereference_raw(ltr->tree[idx].rb_node);
node              113 include/linux/rbtree_latch.h 	while (node) {
node              114 include/linux/rbtree_latch.h 		ltn = __lt_from_rb(node, idx);
node              118 include/linux/rbtree_latch.h 			node = rcu_dereference_raw(node->rb_left);
node              120 include/linux/rbtree_latch.h 			node = rcu_dereference_raw(node->rb_right);
node              144 include/linux/rbtree_latch.h latch_tree_insert(struct latch_tree_node *node,
node              149 include/linux/rbtree_latch.h 	__lt_insert(node, root, 0, ops->less);
node              151 include/linux/rbtree_latch.h 	__lt_insert(node, root, 1, ops->less);
node              171 include/linux/rbtree_latch.h latch_tree_erase(struct latch_tree_node *node,
node              176 include/linux/rbtree_latch.h 	__lt_erase(node, root, 0);
node              178 include/linux/rbtree_latch.h 	__lt_erase(node, root, 1);
node              203 include/linux/rbtree_latch.h 	struct latch_tree_node *node;
node              208 include/linux/rbtree_latch.h 		node = __lt_find(key, root, seq & 1, ops->comp);
node              211 include/linux/rbtree_latch.h 	return node;
node              503 include/linux/rculist.h #define hlist_next_rcu(node)	(*((struct hlist_node __rcu **)(&(node)->next)))
node              504 include/linux/rculist.h #define hlist_pprev_rcu(node)	(*((struct hlist_node __rcu **)((node)->pprev)))
node               44 include/linux/rculist_nulls.h #define hlist_nulls_next_rcu(node) \
node               45 include/linux/rculist_nulls.h 	(*((struct hlist_nulls_node __rcu __force **)&(node)->next))
node               23 include/linux/regulator/of_regulator.h 				    struct device_node *node,
node               25 include/linux/regulator/of_regulator.h extern int of_regulator_match(struct device *dev, struct device_node *node,
node               31 include/linux/regulator/of_regulator.h 				    struct device_node *node,
node               38 include/linux/regulator/of_regulator.h 				     struct device_node *node,
node              336 include/linux/remoteproc.h 	struct list_head node;
node              440 include/linux/remoteproc.h 	struct list_head node;
node              486 include/linux/remoteproc.h 	struct list_head node;
node              529 include/linux/remoteproc.h 	struct list_head node;
node              580 include/linux/remoteproc.h 	struct list_head node;
node               22 include/linux/reset.h struct reset_control *__of_reset_control_get(struct device_node *node,
node               83 include/linux/reset.h 					struct device_node *node,
node              229 include/linux/reset.h 				struct device_node *node, const char *id)
node              231 include/linux/reset.h 	return __of_reset_control_get(node, id, 0, false, false, true);
node              254 include/linux/reset.h 				struct device_node *node, const char *id)
node              256 include/linux/reset.h 	return __of_reset_control_get(node, id, 0, true, false, false);
node              271 include/linux/reset.h 					struct device_node *node, int index)
node              273 include/linux/reset.h 	return __of_reset_control_get(node, NULL, index, false, false, true);
node              299 include/linux/reset.h 					struct device_node *node, int index)
node              301 include/linux/reset.h 	return __of_reset_control_get(node, NULL, index, true, false, false);
node              412 include/linux/reset.h 				struct device_node *node, const char *id)
node              414 include/linux/reset.h 	return of_reset_control_get_exclusive(node, id);
node              418 include/linux/reset.h 				struct device_node *node, int index)
node              420 include/linux/reset.h 	return of_reset_control_get_exclusive_by_index(node, index);
node              470 include/linux/reset.h of_reset_control_array_get_exclusive(struct device_node *node)
node              472 include/linux/reset.h 	return of_reset_control_array_get(node, false, false, true);
node              476 include/linux/reset.h of_reset_control_array_get_exclusive_released(struct device_node *node)
node              478 include/linux/reset.h 	return of_reset_control_array_get(node, false, false, false);
node              482 include/linux/reset.h of_reset_control_array_get_shared(struct device_node *node)
node              484 include/linux/reset.h 	return of_reset_control_array_get(node, true, false, true);
node              488 include/linux/reset.h of_reset_control_array_get_optional_exclusive(struct device_node *node)
node              490 include/linux/reset.h 	return of_reset_control_array_get(node, false, true, true);
node              494 include/linux/reset.h of_reset_control_array_get_optional_shared(struct device_node *node)
node              496 include/linux/reset.h 	return of_reset_control_array_get(node, true, true, true);
node               24 include/linux/resource_ext.h 	struct list_head	node;
node               37 include/linux/resource_ext.h 	list_add(&entry->node, head);
node               43 include/linux/resource_ext.h 	list_add_tail(&entry->node, head);
node               48 include/linux/resource_ext.h 	list_del(&entry->node);
node               64 include/linux/resource_ext.h 	list_for_each_entry((entry), (list), node)
node               67 include/linux/resource_ext.h 	list_for_each_entry_safe((entry), (tmp), (list), node)
node              208 include/linux/ring_buffer.h int trace_rb_cpu_prepare(unsigned int cpu, struct hlist_node *node);
node               98 include/linux/rio.h 	struct list_head node;
node              232 include/linux/rio.h 	struct list_head node;
node              268 include/linux/rio.h 	struct list_head node;	/* node in global list of ports */
node              321 include/linux/rio.h 	struct list_head node;	/* node in list of networks */
node              456 include/linux/rio.h 	struct list_head node;
node              543 include/linux/rio.h 	struct list_head node;
node               13 include/linux/rpmsg/qcom_glink.h 					    struct device_node *node);
node               20 include/linux/rpmsg/qcom_glink.h 			 struct device_node *node)
node               13 include/linux/rpmsg/qcom_smd.h 					     struct device_node *node);
node               20 include/linux/rpmsg/qcom_smd.h 		       struct device_node *node)
node               90 include/linux/rtc.h 	struct timerqueue_node node;
node              156 include/linux/sbitmap.h 		      gfp_t flags, int node);
node              376 include/linux/sbitmap.h 			    int shift, bool round_robin, gfp_t flags, int node);
node               19 include/linux/sched/numa_balancing.h extern void task_numa_fault(int last_node, int node, int pages, int flags);
node               26 include/linux/sched/numa_balancing.h static inline void task_numa_fault(int last_node, int node, int pages,
node               70 include/linux/sched/signal.h 	struct hlist_node node;
node              383 include/linux/serial_core.h 			     unsigned long node,
node               57 include/linux/serio.h 	struct list_head node;
node               39 include/linux/sh_clk.h 	struct list_head	node;
node               47 include/linux/shdma-base.h 	struct list_head node;
node               10 include/linux/siox.h 	struct list_head node; /* node in smaster->devices */
node             1035 include/linux/skbuff.h 			    int node);
node              302 include/linux/skmsg.h struct sk_psock *sk_psock_init(struct sock *sk, int node);
node              417 include/linux/slab.h void *__kmalloc_node(size_t size, gfp_t flags, int node) __assume_kmalloc_alignment __malloc;
node              418 include/linux/slab.h void *kmem_cache_alloc_node(struct kmem_cache *, gfp_t flags, int node) __assume_slab_alignment __malloc;
node              420 include/linux/slab.h static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node)
node              425 include/linux/slab.h static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node)
node              437 include/linux/slab.h 					   int node, size_t size) __assume_slab_alignment __malloc;
node              442 include/linux/slab.h 			      int node, size_t size)
node              461 include/linux/slab.h 			      int node, size_t size)
node              463 include/linux/slab.h 	void *ret = kmem_cache_alloc_node(s, gfpflags, node);
node              584 include/linux/slab.h static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node)
node              596 include/linux/slab.h 						flags, node, size);
node              599 include/linux/slab.h 	return __kmalloc_node(size, flags, node);
node              645 include/linux/slab.h 				       int node)
node              652 include/linux/slab.h 		return kmalloc_node(bytes, flags, node);
node              653 include/linux/slab.h 	return __kmalloc_node(bytes, flags, node);
node              656 include/linux/slab.h static inline void *kcalloc_node(size_t n, size_t size, gfp_t flags, int node)
node              658 include/linux/slab.h 	return kmalloc_array_node(n, size, flags | __GFP_ZERO, node);
node              664 include/linux/slab.h #define kmalloc_node_track_caller(size, flags, node) \
node              665 include/linux/slab.h 	__kmalloc_node_track_caller(size, flags, node, \
node              670 include/linux/slab.h #define kmalloc_node_track_caller(size, flags, node) \
node              699 include/linux/slab.h static inline void *kzalloc_node(size_t size, gfp_t flags, int node)
node              701 include/linux/slab.h 	return kmalloc_node(size, flags | __GFP_ZERO, node);
node               89 include/linux/slab_def.h 	struct kmem_cache_node *node[MAX_NUMNODES];
node              142 include/linux/slub_def.h 	struct kmem_cache_node *node[MAX_NUMNODES];
node               89 include/linux/soc/qcom/apr.h 	struct list_head node;
node              121 include/linux/soc/qcom/qmi.h 	unsigned int node;
node              151 include/linux/soc/qcom/qmi.h 	void (*bye)(struct qmi_handle *qmi, unsigned int node);
node              153 include/linux/soc/qcom/qmi.h 			   unsigned int node, unsigned int port);
node              562 include/linux/soundwire/sdw.h 	struct list_head node;
node             1386 include/linux/spi/spi.h of_find_spi_device_by_node(struct device_node *node);
node             1391 include/linux/spi/spi.h of_find_spi_device_by_node(struct device_node *node)
node               65 include/linux/srcutree.h 	struct srcu_node node[NUM_RCU_NODES];	/* Combining tree. */
node              496 include/linux/sunrpc/svc.h 					struct svc_pool *pool, int node);
node              498 include/linux/sunrpc/svc.h 					struct svc_pool *pool, int node);
node              118 include/linux/svga.h int svga_compute_pll(const struct svga_pll *pll, u32 f_wanted, u16 *m, u16 *n, u16 *r, int node);
node              119 include/linux/svga.h int svga_check_timings(const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, int node);
node              120 include/linux/svga.h void svga_set_timings(void __iomem *regbase, const struct svga_timing_regs *tm, struct fb_var_screeninfo *var, u32 hmul, u32 hdiv, u32 vmul, u32 vdiv, u32 hborder, int node);
node              315 include/linux/swap.h void workingset_update_node(struct xa_node *node);
node              649 include/linux/swap.h extern void mem_cgroup_throttle_swaprate(struct mem_cgroup *memcg, int node,
node              653 include/linux/swap.h 						int node, gfp_t gfp_mask)
node              732 include/linux/syscalls.h asmlinkage long sys_getcpu(unsigned __user *cpu, unsigned __user *node, struct getcpu_cache __user *cache);
node               14 include/linux/syscore_ops.h 	struct list_head node;
node              137 include/linux/sysctl.h 	struct rb_node node;
node              159 include/linux/sysctl.h 	struct ctl_node *node;
node               81 include/linux/tc.h 	struct list_head node;		/* Node in list of all TC devices. */
node              106 include/linux/tc.h 	struct list_head node;
node              141 include/linux/thermal.h 	struct list_head node;
node              219 include/linux/thermal.h 	struct list_head node;
node               10 include/linux/timerqueue.h 	struct rb_node node;
node               20 include/linux/timerqueue.h 			   struct timerqueue_node *node);
node               22 include/linux/timerqueue.h 			   struct timerqueue_node *node);
node               24 include/linux/timerqueue.h 						struct timerqueue_node *node);
node               38 include/linux/timerqueue.h 	return rb_entry(leftmost, struct timerqueue_node, node);
node               41 include/linux/timerqueue.h static inline void timerqueue_init(struct timerqueue_node *node)
node               43 include/linux/timerqueue.h 	RB_CLEAR_NODE(&node->node);
node               46 include/linux/timerqueue.h static inline bool timerqueue_node_queued(struct timerqueue_node *node)
node               48 include/linux/timerqueue.h 	return !RB_EMPTY_NODE(&node->node);
node               51 include/linux/timerqueue.h static inline bool timerqueue_node_expires(struct timerqueue_node *node)
node               53 include/linux/timerqueue.h 	return node->expires;
node               39 include/linux/topology.h #define nr_cpus_node(node) cpumask_weight(cpumask_of_node(node))
node               42 include/linux/topology.h #define for_each_node_with_cpus(node)			\
node               43 include/linux/topology.h 	for_each_online_node(node)			\
node               44 include/linux/topology.h 		if (nr_cpus_node(node))
node              100 include/linux/topology.h static inline void set_numa_node(int node)
node              102 include/linux/topology.h 	this_cpu_write(numa_node, node);
node              107 include/linux/topology.h static inline void set_cpu_numa_node(int cpu, int node)
node              109 include/linux/topology.h 	per_cpu(numa_node, cpu) = node;
node              136 include/linux/topology.h static inline void set_numa_mem(int node)
node              138 include/linux/topology.h 	this_cpu_write(_numa_mem_, node);
node              139 include/linux/topology.h 	_node_numa_mem_[numa_node_id()] = node;
node              144 include/linux/topology.h static inline int node_to_mem_node(int node)
node              146 include/linux/topology.h 	return _node_numa_mem_[node];
node              166 include/linux/topology.h static inline void set_cpu_numa_mem(int cpu, int node)
node              168 include/linux/topology.h 	per_cpu(_numa_mem_, cpu) = node;
node              169 include/linux/topology.h 	_node_numa_mem_[cpu_to_node(cpu)] = node;
node              184 include/linux/topology.h static inline int node_to_mem_node(int node)
node              186 include/linux/topology.h 	return node;
node             1096 include/linux/usb.h 	struct list_head node;
node               79 include/linux/usb/isp1301.h struct i2c_client *isp1301_get_client(struct device_node *node);
node              224 include/linux/usb/phy.h 	struct device_node *node, struct notifier_block *nb);
node              253 include/linux/usb/phy.h 	struct device_node *node, struct notifier_block *nb)
node               75 include/linux/usb/role.h fwnode_usb_role_switch_get(struct fwnode_handle *node)
node               92 include/linux/user_namespace.h 	struct hlist_node node;
node               26 include/linux/vexpress.h int vexpress_config_get_topo(struct device_node *node, u32 *site,
node               78 include/linux/vmalloc.h 				int node, pgprot_t prot);
node               94 include/linux/vmalloc.h extern void *vmalloc_node(unsigned long size, int node);
node               95 include/linux/vmalloc.h extern void *vzalloc_node(unsigned long size, int node);
node              102 include/linux/vmalloc.h 			pgprot_t prot, unsigned long vm_flags, int node,
node              105 include/linux/vmalloc.h extern void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags);
node              106 include/linux/vmalloc.h static inline void *__vmalloc_node_flags_caller(unsigned long size, int node,
node              109 include/linux/vmalloc.h 	return __vmalloc_node_flags(size, node, flags);
node              113 include/linux/vmalloc.h 					 int node, gfp_t flags, void *caller);
node              232 include/linux/vmstat.h extern unsigned long sum_zone_node_page_state(int node,
node              234 include/linux/vmstat.h extern unsigned long sum_zone_numa_state(int node, enum numa_stat_item item);
node              238 include/linux/vmstat.h #define sum_zone_node_page_state(node, item) global_zone_page_state(item)
node              239 include/linux/vmstat.h #define node_page_state(node, item) global_node_page_state(item)
node              446 include/linux/workqueue.h extern bool queue_work_node(int node, struct workqueue_struct *wq,
node             1115 include/linux/xarray.h #define XA_NODE_BUG_ON(node, x) do {				\
node             1117 include/linux/xarray.h 			if (node) xa_dump_node(node);		\
node             1123 include/linux/xarray.h #define XA_NODE_BUG_ON(node, x)	do { } while (0)
node             1142 include/linux/xarray.h 				const struct xa_node *node, unsigned int offset)
node             1144 include/linux/xarray.h 	XA_NODE_BUG_ON(node, offset >= XA_CHUNK_SIZE);
node             1145 include/linux/xarray.h 	return rcu_dereference_check(node->slots[offset],
node             1151 include/linux/xarray.h 				const struct xa_node *node, unsigned int offset)
node             1153 include/linux/xarray.h 	XA_NODE_BUG_ON(node, offset >= XA_CHUNK_SIZE);
node             1154 include/linux/xarray.h 	return rcu_dereference_protected(node->slots[offset],
node             1160 include/linux/xarray.h 					const struct xa_node *node)
node             1162 include/linux/xarray.h 	return rcu_dereference_check(node->parent,
node             1168 include/linux/xarray.h 					const struct xa_node *node)
node             1170 include/linux/xarray.h 	return rcu_dereference_protected(node->parent,
node             1175 include/linux/xarray.h static inline void *xa_mk_node(const struct xa_node *node)
node             1177 include/linux/xarray.h 	return (void *)((unsigned long)node | 2);
node             1252 include/linux/xarray.h typedef void (*xa_update_node_t)(struct xa_node *node);
node             1403 include/linux/xarray.h static inline bool xas_not_node(struct xa_node *node)
node             1405 include/linux/xarray.h 	return ((unsigned long)node & 3) || !node;
node             1409 include/linux/xarray.h static inline bool xas_frozen(struct xa_node *node)
node             1411 include/linux/xarray.h 	return (unsigned long)node & 2;
node             1415 include/linux/xarray.h static inline bool xas_top(struct xa_node *node)
node             1417 include/linux/xarray.h 	return node <= XAS_RESTART;
node             1489 include/linux/xarray.h 	struct xa_node *node = xas->xa_node;
node             1491 include/linux/xarray.h 	if (node)
node             1492 include/linux/xarray.h 		return xa_entry(xas->xa, node, xas->xa_offset);
node             1557 include/linux/xarray.h 	struct xa_node *node = xas->xa_node;
node             1560 include/linux/xarray.h 	if (unlikely(xas_not_node(node) || node->shift ||
node             1569 include/linux/xarray.h 		entry = xa_entry(xas->xa, node, xas->xa_offset + 1);
node             1615 include/linux/xarray.h 	struct xa_node *node = xas->xa_node;
node             1619 include/linux/xarray.h 	if (unlikely(xas_not_node(node) || node->shift))
node             1628 include/linux/xarray.h 	entry = xa_entry(xas->xa, node, offset);
node             1714 include/linux/xarray.h 	struct xa_node *node = xas->xa_node;
node             1716 include/linux/xarray.h 	if (unlikely(xas_not_node(node) || node->shift ||
node             1722 include/linux/xarray.h 	return xa_entry(xas->xa, node, xas->xa_offset);
node             1743 include/linux/xarray.h 	struct xa_node *node = xas->xa_node;
node             1745 include/linux/xarray.h 	if (unlikely(xas_not_node(node) || node->shift ||
node             1751 include/linux/xarray.h 	return xa_entry(xas->xa, node, xas->xa_offset);
node               93 include/linux/xattr.h 	struct simple_xattr *xattr, *node;
node               95 include/linux/xattr.h 	list_for_each_entry_safe(xattr, node, &xattrs->head, list) {
node               55 include/linux/zorro.h     struct list_head node;
node               70 include/media/v4l2-clk.h #define v4l2_clk_name_of(name, size, node) snprintf(name, size, \
node               71 include/media/v4l2-clk.h 			  "of-%pOF", node)
node              212 include/media/v4l2-ctrls.h 	struct list_head node;
node              288 include/media/v4l2-ctrls.h 	struct list_head node;
node               76 include/media/v4l2-event.h 	struct list_head	node;
node              151 include/net/caif/caif_layer.h 	struct list_head node;
node               16 include/net/datalink.h 	struct list_head node;
node               80 include/net/garp.h 	struct rb_node			node;
node               78 include/net/inet_frag.h 	struct rhash_head	node;
node               89 include/net/inet_hashtables.h 	struct hlist_node	node;
node               99 include/net/inet_hashtables.h 	hlist_for_each_entry(tb, head, node)
node              318 include/net/ip6_fib.h 	struct fib6_node *root, *node;
node               22 include/net/ipx.h 	__u8    node[IPX_NODE_LEN]; 
node               71 include/net/ipx.h 	struct list_head	node; /* node in ipx_interfaces list */
node               79 include/net/ipx.h 	struct list_head	node; /* node in ipx_routes list */
node              102 include/net/ipx.h 	unsigned char		node[IPX_NODE_LEN];
node              144 include/net/ipx.h int ipxitf_send(struct ipx_interface *intrfc, struct sk_buff *skb, char *node);
node              147 include/net/ipx.h 		     unsigned char *node);
node               84 include/net/lapb.h 	struct list_head	node;
node               64 include/net/llc.h 	struct list_head node;
node               91 include/net/mrp.h 	struct rb_node			node;
node               63 include/net/netfilter/nf_flow_table.h 	struct rhash_head		node;
node               19 include/net/ping.h #define ping_portaddr_for_each_entry(__sk, node, list) \
node               20 include/net/ping.h 	hlist_nulls_for_each_entry(__sk, node, list, sk_nulls_node)
node               22 include/net/pkt_cls.h 	int	(*fn)(struct tcf_proto *, void *node, struct tcf_walker *);
node              510 include/net/sctp/sctp.h 	hlist_for_each_entry(epb, head, node)
node               87 include/net/sctp/structs.h 	struct hlist_node	node;
node              776 include/net/sctp/structs.h 	struct rhlist_head node;
node             1226 include/net/sctp/structs.h 	struct hlist_node node;
node               26 include/net/seg6_hmac.h 	struct rhash_head node;
node              567 include/net/sock.h static inline struct sock *sk_entry(const struct hlist_node *node)
node              569 include/net/sock.h 	return hlist_entry(node, struct sock, sk_node);
node              615 include/net/sock.h static inline void sk_node_init(struct hlist_node *node)
node              617 include/net/sock.h 	node->pprev = NULL;
node              620 include/net/sock.h static inline void sk_nulls_node_init(struct hlist_nulls_node *node)
node              622 include/net/sock.h 	node->pprev = NULL;
node              752 include/net/sock.h #define sk_nulls_for_each(__sk, node, list) \
node              753 include/net/sock.h 	hlist_nulls_for_each_entry(__sk, node, list, sk_nulls_node)
node              754 include/net/sock.h #define sk_nulls_for_each_rcu(__sk, node, list) \
node              755 include/net/sock.h 	hlist_nulls_for_each_entry_rcu(__sk, node, list, sk_nulls_node)
node              758 include/net/sock.h #define sk_nulls_for_each_from(__sk, node) \
node              759 include/net/sock.h 	if (__sk && ({ node = &(__sk)->sk_nulls_node; 1; })) \
node              760 include/net/sock.h 		hlist_nulls_for_each_entry_from(__sk, node, sk_nulls_node)
node             1182 include/net/sock.h 	struct list_head	node;
node             1539 include/net/tcp.h 	struct hlist_node	node;
node              130 include/net/x25.h 	struct list_head	node;		
node              138 include/net/x25.h 	struct list_head	node;
node              172 include/net/x25.h 	struct list_head	node;
node               15 include/net/xdp_priv.h 	struct rhash_head node;
node               74 include/net/xdp_sock.h 	struct list_head node;
node              166 include/rdma/ib_umem_odp.h 	struct interval_tree_node *node;
node              168 include/rdma/ib_umem_odp.h 	node = interval_tree_iter_first(root, addr, addr + length - 1);
node              169 include/rdma/ib_umem_odp.h 	if (!node)
node              171 include/rdma/ib_umem_odp.h 	return container_of(node, struct ib_umem_odp, interval_tree);
node              179 include/rdma/rdma_vt.h 	int node;
node               82 include/scsi/scsi_device.h 	struct list_head	node;
node               71 include/soc/at91/atmel_tcb.h 	struct list_head	node;
node               90 include/soc/fsl/dpaa2-io.h 	struct list_head node;
node               57 include/soc/fsl/qe/qe_ic.h void qe_ic_init(struct device_node *node, unsigned int flags,
node               63 include/soc/fsl/qe/qe_ic.h static inline void qe_ic_init(struct device_node *node, unsigned int flags,
node               73 include/sound/simple_card_utils.h 			     struct device_node *node,
node               84 include/sound/simple_card_utils.h #define asoc_simple_parse_clk_cpu(dev, node, dai_link, simple_dai)		\
node               85 include/sound/simple_card_utils.h 	asoc_simple_parse_clk(dev, node, simple_dai, dai_link->cpus)
node               86 include/sound/simple_card_utils.h #define asoc_simple_parse_clk_codec(dev, node, dai_link, simple_dai)	\
node               87 include/sound/simple_card_utils.h 	asoc_simple_parse_clk(dev, node, simple_dai, dai_link->codecs)
node               89 include/sound/simple_card_utils.h 			  struct device_node *node,
node              100 include/sound/simple_card_utils.h #define asoc_simple_parse_cpu(node, dai_link, is_single_link)	\
node              101 include/sound/simple_card_utils.h 	asoc_simple_parse_dai(node, dai_link->cpus, is_single_link)
node              102 include/sound/simple_card_utils.h #define asoc_simple_parse_codec(node, dai_link)	\
node              103 include/sound/simple_card_utils.h 	asoc_simple_parse_dai(node, dai_link->codecs, NULL)
node              104 include/sound/simple_card_utils.h #define asoc_simple_parse_platform(node, dai_link)	\
node              105 include/sound/simple_card_utils.h 	asoc_simple_parse_dai(node, dai_link->platforms, NULL)
node              576 include/sound/soc-dapm.h 		struct snd_soc_dapm_widget *node[2];
node               61 include/trace/events/alarmtimer.h 		__entry->expires = alarm->node.expires;
node              167 include/trace/events/asoc.h 		__string(	pnname,	path->node[dir]->name	)
node              176 include/trace/events/asoc.h 		__assign_str(pnname, path->node[dir]->name);
node              178 include/trace/events/asoc.h 		__entry->path_node = (long)path->node[dir];
node               43 include/trace/events/cpuhp.h 		 struct hlist_node *node),
node               45 include/trace/events/cpuhp.h 	TP_ARGS(cpu, target, idx, fun, node),
node               69 include/trace/events/kmem.h 		 int node),
node               71 include/trace/events/kmem.h 	TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node),
node               79 include/trace/events/kmem.h 		__field(	int,		node		)
node               88 include/trace/events/kmem.h 		__entry->node		= node;
node               97 include/trace/events/kmem.h 		__entry->node)
node              104 include/trace/events/kmem.h 		 gfp_t gfp_flags, int node),
node              106 include/trace/events/kmem.h 	TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
node              113 include/trace/events/kmem.h 		 gfp_t gfp_flags, int node),
node              115 include/trace/events/kmem.h 	TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
node               45 include/trace/events/oom.h 		__field(	int, node)
node               56 include/trace/events/oom.h 		__entry->node = zone_to_nid(zoneref->zone);
node               67 include/trace/events/oom.h 			__entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
node              173 include/trace/events/rtc.h 		__entry->expires = timer->node.expires;
node               38 include/uapi/linux/qrtr.h 			__le32 node;
node               43 include/uapi/linux/qrtr.h 			__le32 node;
node               70 include/uapi/linux/scif_ioctl.h 	__u16 node;
node               50 include/uapi/linux/tipc.h 	__u32 node;
node              269 include/uapi/linux/tipc.h 			      unsigned int node)
node              273 include/uapi/linux/tipc.h 		node;
node              173 include/video/mmp_disp.h 	struct list_head node;
node              211 include/video/mmp_disp.h 	struct list_head node;
node              826 include/video/omapfb_dss.h omapdss_of_find_source_for_first_ep(struct device_node *node);
node               60 include/xen/xenbus.h 	const char *node;
node              137 include/xen/xenbus.h 			const char *dir, const char *node, unsigned int *num);
node              139 include/xen/xenbus.h 		  const char *dir, const char *node, unsigned int *len);
node              141 include/xen/xenbus.h 		 const char *dir, const char *node, const char *string);
node              143 include/xen/xenbus.h 		 const char *dir, const char *node);
node              145 include/xen/xenbus.h 		  const char *dir, const char *node);
node              146 include/xen/xenbus.h int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node);
node              153 include/xen/xenbus.h 		 const char *dir, const char *node, const char *fmt, ...);
node              156 include/xen/xenbus.h unsigned int xenbus_read_unsigned(const char *dir, const char *node,
node              162 include/xen/xenbus.h 		  const char *dir, const char *node, const char *fmt, ...);
node              177 ipc/mqueue.c   	struct rb_node *node = &leaf->rb_node;
node              179 ipc/mqueue.c   	if (info->msg_tree_rightmost == node)
node              180 ipc/mqueue.c   		info->msg_tree_rightmost = rb_prev(node);
node              182 ipc/mqueue.c   	rb_erase(node, &info->msg_tree);
node              166 kernel/async.c 					  int node, struct async_domain *domain)
node              212 kernel/async.c 	queue_work_node(node, system_unbound_wq, &entry->work);
node              231 kernel/async.c async_cookie_t async_schedule_node(async_func_t func, void *data, int node)
node              233 kernel/async.c 	return async_schedule_node_domain(func, data, node, &async_dfl_domain);
node              194 kernel/audit_tree.c 	size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node);
node              274 kernel/audit_tree.c static struct audit_chunk *find_chunk(struct node *p)
node              327 kernel/audit_tree.c static void remove_chunk_node(struct audit_chunk *chunk, struct node *p)
node              464 kernel/audit_tree.c 	struct node *p;
node              575 kernel/audit_tree.c 		struct node *p;
node              579 kernel/audit_tree.c 		p = list_first_entry(&victim->chunks, struct node, list);
node              621 kernel/audit_tree.c 		struct node *node = list_entry(p, struct node, list);
node              623 kernel/audit_tree.c 		if (node->index & (1U<<31)) {
node              689 kernel/audit_tree.c 		struct node *node;
node              708 kernel/audit_tree.c 		list_for_each_entry(node, &tree->chunks, list) {
node              709 kernel/audit_tree.c 			struct audit_chunk *chunk = find_chunk(node);
node              711 kernel/audit_tree.c 			node->index |= 1U<<31;
node              715 kernel/audit_tree.c 				node->index &= ~(1U<<31);
node              845 kernel/audit_tree.c 		struct node *node;
node              847 kernel/audit_tree.c 		list_for_each_entry(node, &tree->chunks, list)
node              848 kernel/audit_tree.c 			node->index &= ~(1U<<31);
node              947 kernel/audit_tree.c 			struct node *node;
node              949 kernel/audit_tree.c 			list_for_each_entry(node, &tree->chunks, list)
node              950 kernel/audit_tree.c 				node->index &= ~(1U<<31);
node               42 kernel/bpf/bpf_lru_list.c static bool bpf_lru_node_is_ref(const struct bpf_lru_node *node)
node               44 kernel/bpf/bpf_lru_list.c 	return node->ref;
node               62 kernel/bpf/bpf_lru_list.c 					struct bpf_lru_node *node,
node               66 kernel/bpf/bpf_lru_list.c 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
node               72 kernel/bpf/bpf_lru_list.c 	if (&node->list == l->next_inactive_rotation)
node               75 kernel/bpf/bpf_lru_list.c 	bpf_lru_list_count_dec(l, node->type);
node               77 kernel/bpf/bpf_lru_list.c 	node->type = tgt_free_type;
node               78 kernel/bpf/bpf_lru_list.c 	list_move(&node->list, free_list);
node               83 kernel/bpf/bpf_lru_list.c 				   struct bpf_lru_node *node,
node               86 kernel/bpf/bpf_lru_list.c 	if (WARN_ON_ONCE(!IS_LOCAL_LIST_TYPE(node->type)) ||
node               91 kernel/bpf/bpf_lru_list.c 	node->type = tgt_type;
node               92 kernel/bpf/bpf_lru_list.c 	node->ref = 0;
node               93 kernel/bpf/bpf_lru_list.c 	list_move(&node->list, &l->lists[tgt_type]);
node              101 kernel/bpf/bpf_lru_list.c 				struct bpf_lru_node *node,
node              104 kernel/bpf/bpf_lru_list.c 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)) ||
node              108 kernel/bpf/bpf_lru_list.c 	if (node->type != tgt_type) {
node              109 kernel/bpf/bpf_lru_list.c 		bpf_lru_list_count_dec(l, node->type);
node              111 kernel/bpf/bpf_lru_list.c 		node->type = tgt_type;
node              113 kernel/bpf/bpf_lru_list.c 	node->ref = 0;
node              118 kernel/bpf/bpf_lru_list.c 	if (&node->list == l->next_inactive_rotation)
node              121 kernel/bpf/bpf_lru_list.c 	list_move(&node->list, &l->lists[tgt_type]);
node              143 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node, *tmp_node, *first_node;
node              147 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_safe_reverse(node, tmp_node, active, list) {
node              148 kernel/bpf/bpf_lru_list.c 		if (bpf_lru_node_is_ref(node))
node              149 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
node              151 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
node              153 kernel/bpf/bpf_lru_list.c 		if (++i == lru->nr_scans || node == first_node)
node              171 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node;
node              188 kernel/bpf/bpf_lru_list.c 		node = list_entry(cur, struct bpf_lru_node, list);
node              190 kernel/bpf/bpf_lru_list.c 		if (bpf_lru_node_is_ref(node))
node              191 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
node              213 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node, *tmp_node;
node              217 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) {
node              218 kernel/bpf/bpf_lru_list.c 		if (bpf_lru_node_is_ref(node)) {
node              219 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_ACTIVE);
node              220 kernel/bpf/bpf_lru_list.c 		} else if (lru->del_from_htab(lru->del_arg, node)) {
node              221 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move_to_free(l, node, free_list,
node              262 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node, *tmp_node;
node              277 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_safe_reverse(node, tmp_node, force_shrink_list,
node              279 kernel/bpf/bpf_lru_list.c 		if (lru->del_from_htab(lru->del_arg, node)) {
node              280 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move_to_free(l, node, free_list,
node              293 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node, *tmp_node;
node              295 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_safe_reverse(node, tmp_node,
node              297 kernel/bpf/bpf_lru_list.c 		if (bpf_lru_node_is_ref(node))
node              298 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move_in(l, node, BPF_LRU_LIST_T_ACTIVE);
node              300 kernel/bpf/bpf_lru_list.c 			__bpf_lru_node_move_in(l, node,
node              306 kernel/bpf/bpf_lru_list.c 				   struct bpf_lru_node *node)
node              310 kernel/bpf/bpf_lru_list.c 	if (WARN_ON_ONCE(IS_LOCAL_LIST_TYPE(node->type)))
node              314 kernel/bpf/bpf_lru_list.c 	__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
node              322 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node, *tmp_node;
node              331 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_safe(node, tmp_node, &l->lists[BPF_LRU_LIST_T_FREE],
node              333 kernel/bpf/bpf_lru_list.c 		__bpf_lru_node_move_to_free(l, node, local_free_list(loc_l),
node              350 kernel/bpf/bpf_lru_list.c 				     struct bpf_lru_node *node,
node              353 kernel/bpf/bpf_lru_list.c 	*(u32 *)((void *)node + lru->hash_offset) = hash;
node              354 kernel/bpf/bpf_lru_list.c 	node->cpu = cpu;
node              355 kernel/bpf/bpf_lru_list.c 	node->type = BPF_LRU_LOCAL_LIST_T_PENDING;
node              356 kernel/bpf/bpf_lru_list.c 	node->ref = 0;
node              357 kernel/bpf/bpf_lru_list.c 	list_add(&node->list, local_pending_list(loc_l));
node              363 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node;
node              365 kernel/bpf/bpf_lru_list.c 	node = list_first_entry_or_null(local_free_list(loc_l),
node              368 kernel/bpf/bpf_lru_list.c 	if (node)
node              369 kernel/bpf/bpf_lru_list.c 		list_del(&node->list);
node              371 kernel/bpf/bpf_lru_list.c 	return node;
node              377 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node;
node              382 kernel/bpf/bpf_lru_list.c 	list_for_each_entry_reverse(node, local_pending_list(loc_l),
node              384 kernel/bpf/bpf_lru_list.c 		if ((!bpf_lru_node_is_ref(node) || force) &&
node              385 kernel/bpf/bpf_lru_list.c 		    lru->del_from_htab(lru->del_arg, node)) {
node              386 kernel/bpf/bpf_lru_list.c 			list_del(&node->list);
node              387 kernel/bpf/bpf_lru_list.c 			return node;
node              403 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node = NULL;
node              420 kernel/bpf/bpf_lru_list.c 		node = list_first_entry(free_list, struct bpf_lru_node, list);
node              421 kernel/bpf/bpf_lru_list.c 		*(u32 *)((void *)node + lru->hash_offset) = hash;
node              422 kernel/bpf/bpf_lru_list.c 		node->ref = 0;
node              423 kernel/bpf/bpf_lru_list.c 		__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_INACTIVE);
node              428 kernel/bpf/bpf_lru_list.c 	return node;
node              436 kernel/bpf/bpf_lru_list.c 	struct bpf_lru_node *node;
node              445 kernel/bpf/bpf_lru_list.c 	node = __local_list_pop_free(loc_l);
node              446 kernel/bpf/bpf_lru_list.c 	if (!node) {
node              448 kernel/bpf/bpf_lru_list.c 		node = __local_list_pop_free(loc_l);
node              451 kernel/bpf/bpf_lru_list.c 	if (node)
node              452 kernel/bpf/bpf_lru_list.c 		__local_list_add_pending(lru, loc_l, cpu, node, hash);
node              456 kernel/bpf/bpf_lru_list.c 	if (node)
node              457 kernel/bpf/bpf_lru_list.c 		return node;
node              474 kernel/bpf/bpf_lru_list.c 		node = __local_list_pop_free(steal_loc_l);
node              475 kernel/bpf/bpf_lru_list.c 		if (!node)
node              476 kernel/bpf/bpf_lru_list.c 			node = __local_list_pop_pending(lru, steal_loc_l);
node              481 kernel/bpf/bpf_lru_list.c 	} while (!node && steal != first_steal);
node              485 kernel/bpf/bpf_lru_list.c 	if (node) {
node              487 kernel/bpf/bpf_lru_list.c 		__local_list_add_pending(lru, loc_l, cpu, node, hash);
node              491 kernel/bpf/bpf_lru_list.c 	return node;
node              503 kernel/bpf/bpf_lru_list.c 				     struct bpf_lru_node *node)
node              507 kernel/bpf/bpf_lru_list.c 	if (WARN_ON_ONCE(node->type == BPF_LRU_LIST_T_FREE) ||
node              508 kernel/bpf/bpf_lru_list.c 	    WARN_ON_ONCE(node->type == BPF_LRU_LOCAL_LIST_T_FREE))
node              511 kernel/bpf/bpf_lru_list.c 	if (node->type == BPF_LRU_LOCAL_LIST_T_PENDING) {
node              514 kernel/bpf/bpf_lru_list.c 		loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
node              518 kernel/bpf/bpf_lru_list.c 		if (unlikely(node->type != BPF_LRU_LOCAL_LIST_T_PENDING)) {
node              523 kernel/bpf/bpf_lru_list.c 		node->type = BPF_LRU_LOCAL_LIST_T_FREE;
node              524 kernel/bpf/bpf_lru_list.c 		node->ref = 0;
node              525 kernel/bpf/bpf_lru_list.c 		list_move(&node->list, local_free_list(loc_l));
node              532 kernel/bpf/bpf_lru_list.c 	bpf_lru_list_push_free(&lru->common_lru.lru_list, node);
node              536 kernel/bpf/bpf_lru_list.c 				     struct bpf_lru_node *node)
node              541 kernel/bpf/bpf_lru_list.c 	l = per_cpu_ptr(lru->percpu_lru, node->cpu);
node              545 kernel/bpf/bpf_lru_list.c 	__bpf_lru_node_move(l, node, BPF_LRU_LIST_T_FREE);
node              550 kernel/bpf/bpf_lru_list.c void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node)
node              553 kernel/bpf/bpf_lru_list.c 		bpf_percpu_lru_push_free(lru, node);
node              555 kernel/bpf/bpf_lru_list.c 		bpf_common_lru_push_free(lru, node);
node              566 kernel/bpf/bpf_lru_list.c 		struct bpf_lru_node *node;
node              568 kernel/bpf/bpf_lru_list.c 		node = (struct bpf_lru_node *)(buf + node_offset);
node              569 kernel/bpf/bpf_lru_list.c 		node->type = BPF_LRU_LIST_T_FREE;
node              570 kernel/bpf/bpf_lru_list.c 		node->ref = 0;
node              571 kernel/bpf/bpf_lru_list.c 		list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
node              589 kernel/bpf/bpf_lru_list.c 		struct bpf_lru_node *node;
node              593 kernel/bpf/bpf_lru_list.c 		node = (struct bpf_lru_node *)(buf + node_offset);
node              594 kernel/bpf/bpf_lru_list.c 		node->cpu = cpu;
node              595 kernel/bpf/bpf_lru_list.c 		node->type = BPF_LRU_LIST_T_FREE;
node              596 kernel/bpf/bpf_lru_list.c 		node->ref = 0;
node              597 kernel/bpf/bpf_lru_list.c 		list_add(&node->list, &l->lists[BPF_LRU_LIST_T_FREE]);
node               50 kernel/bpf/bpf_lru_list.h typedef bool (*del_from_htab_func)(void *arg, struct bpf_lru_node *node);
node               64 kernel/bpf/bpf_lru_list.h static inline void bpf_lru_node_set_ref(struct bpf_lru_node *node)
node               69 kernel/bpf/bpf_lru_list.h 	if (!node->ref)
node               70 kernel/bpf/bpf_lru_list.h 		node->ref = 1;
node               79 kernel/bpf/bpf_lru_list.h void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
node               80 kernel/bpf/bpf_lru_list.h void bpf_lru_promote(struct bpf_lru *lru, struct bpf_lru_node *node);
node               50 kernel/bpf/cgroup.c 		list_for_each_entry_safe(pl, tmp, progs, node) {
node               51 kernel/bpf/cgroup.c 			list_del(&pl->node);
node               96 kernel/bpf/cgroup.c 	list_for_each_entry(pl, head, node) {
node              166 kernel/bpf/cgroup.c 		list_for_each_entry(pl, &p->bpf.progs[type], node) {
node              341 kernel/bpf/cgroup.c 		list_for_each_entry(pl, progs, node) {
node              361 kernel/bpf/cgroup.c 		list_add_tail(&pl->node, progs);
node              371 kernel/bpf/cgroup.c 			list_add_tail(&pl->node, progs);
node              373 kernel/bpf/cgroup.c 			pl = list_first_entry(progs, typeof(*pl), node);
node              415 kernel/bpf/cgroup.c 		list_del(&pl->node);
node              454 kernel/bpf/cgroup.c 		list_for_each_entry(pl, progs, node) {
node              470 kernel/bpf/cgroup.c 		pl = list_first_entry(progs, typeof(*pl), node);
node              480 kernel/bpf/cgroup.c 	list_del(&pl->node);
node              538 kernel/bpf/cgroup.c 		list_for_each_entry(pl, progs, node) {
node               60 kernel/bpf/hashtab.c static bool htab_lru_map_delete_node(void *arg, struct bpf_lru_node *node);
node              122 kernel/bpf/hashtab.c 	struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash);
node              125 kernel/bpf/hashtab.c 	if (node) {
node              126 kernel/bpf/hashtab.c 		l = container_of(node, struct htab_elem, lru_node);
node              570 kernel/bpf/hashtab.c static bool htab_lru_map_delete_node(void *arg, struct bpf_lru_node *node)
node              579 kernel/bpf/hashtab.c 	tgt_l = container_of(node, struct htab_elem, lru_node);
node               53 kernel/bpf/local_storage.c 	struct rb_node *node;
node               58 kernel/bpf/local_storage.c 	node = root->rb_node;
node               59 kernel/bpf/local_storage.c 	while (node) {
node               62 kernel/bpf/local_storage.c 		storage = container_of(node, struct bpf_cgroup_storage, node);
node               66 kernel/bpf/local_storage.c 			node = node->rb_left;
node               69 kernel/bpf/local_storage.c 			node = node->rb_right;
node               93 kernel/bpf/local_storage.c 		this = container_of(*new, struct bpf_cgroup_storage, node);
node              108 kernel/bpf/local_storage.c 	rb_link_node(&storage->node, parent, new);
node              109 kernel/bpf/local_storage.c 	rb_insert_color(&storage->node, root);
node              594 kernel/bpf/local_storage.c 	rb_erase(&storage->node, root);
node              165 kernel/bpf/lpm_trie.c 				   const struct lpm_trie_node *node,
node              168 kernel/bpf/lpm_trie.c 	u32 limit = min(node->prefixlen, key->prefixlen);
node              180 kernel/bpf/lpm_trie.c 		u64 diff = be64_to_cpu(*(__be64 *)node->data ^
node              193 kernel/bpf/lpm_trie.c 		u32 diff = be32_to_cpu(*(__be32 *)&node->data[i] ^
node              205 kernel/bpf/lpm_trie.c 		u16 diff = be16_to_cpu(*(__be16 *)&node->data[i] ^
node              217 kernel/bpf/lpm_trie.c 		prefixlen += 8 - fls(node->data[i] ^ key->data[i]);
node              230 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node, *found = NULL;
node              235 kernel/bpf/lpm_trie.c 	for (node = rcu_dereference(trie->root); node;) {
node              243 kernel/bpf/lpm_trie.c 		matchlen = longest_prefix_match(trie, node, key);
node              245 kernel/bpf/lpm_trie.c 			found = node;
node              253 kernel/bpf/lpm_trie.c 		if (matchlen < node->prefixlen)
node              259 kernel/bpf/lpm_trie.c 		if (!(node->flags & LPM_TREE_NODE_FLAG_IM))
node              260 kernel/bpf/lpm_trie.c 			found = node;
node              266 kernel/bpf/lpm_trie.c 		next_bit = extract_bit(key->data, node->prefixlen);
node              267 kernel/bpf/lpm_trie.c 		node = rcu_dereference(node->child[next_bit]);
node              279 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node;
node              285 kernel/bpf/lpm_trie.c 	node = kmalloc_node(size, GFP_ATOMIC | __GFP_NOWARN,
node              287 kernel/bpf/lpm_trie.c 	if (!node)
node              290 kernel/bpf/lpm_trie.c 	node->flags = 0;
node              293 kernel/bpf/lpm_trie.c 		memcpy(node->data + trie->data_size, value,
node              296 kernel/bpf/lpm_trie.c 	return node;
node              304 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node, *im_node = NULL, *new_node = NULL;
node              347 kernel/bpf/lpm_trie.c 	while ((node = rcu_dereference_protected(*slot,
node              349 kernel/bpf/lpm_trie.c 		matchlen = longest_prefix_match(trie, node, key);
node              351 kernel/bpf/lpm_trie.c 		if (node->prefixlen != matchlen ||
node              352 kernel/bpf/lpm_trie.c 		    node->prefixlen == key->prefixlen ||
node              353 kernel/bpf/lpm_trie.c 		    node->prefixlen == trie->max_prefixlen)
node              356 kernel/bpf/lpm_trie.c 		next_bit = extract_bit(key->data, node->prefixlen);
node              357 kernel/bpf/lpm_trie.c 		slot = &node->child[next_bit];
node              363 kernel/bpf/lpm_trie.c 	if (!node) {
node              371 kernel/bpf/lpm_trie.c 	if (node->prefixlen == matchlen) {
node              372 kernel/bpf/lpm_trie.c 		new_node->child[0] = node->child[0];
node              373 kernel/bpf/lpm_trie.c 		new_node->child[1] = node->child[1];
node              375 kernel/bpf/lpm_trie.c 		if (!(node->flags & LPM_TREE_NODE_FLAG_IM))
node              379 kernel/bpf/lpm_trie.c 		kfree_rcu(node, rcu);
node              388 kernel/bpf/lpm_trie.c 		next_bit = extract_bit(node->data, matchlen);
node              389 kernel/bpf/lpm_trie.c 		rcu_assign_pointer(new_node->child[next_bit], node);
node              402 kernel/bpf/lpm_trie.c 	memcpy(im_node->data, node->data, trie->data_size);
node              406 kernel/bpf/lpm_trie.c 		rcu_assign_pointer(im_node->child[0], node);
node              410 kernel/bpf/lpm_trie.c 		rcu_assign_pointer(im_node->child[1], node);
node              436 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node, *parent;
node              456 kernel/bpf/lpm_trie.c 	while ((node = rcu_dereference_protected(
node              458 kernel/bpf/lpm_trie.c 		matchlen = longest_prefix_match(trie, node, key);
node              460 kernel/bpf/lpm_trie.c 		if (node->prefixlen != matchlen ||
node              461 kernel/bpf/lpm_trie.c 		    node->prefixlen == key->prefixlen)
node              464 kernel/bpf/lpm_trie.c 		parent = node;
node              466 kernel/bpf/lpm_trie.c 		next_bit = extract_bit(key->data, node->prefixlen);
node              467 kernel/bpf/lpm_trie.c 		trim = &node->child[next_bit];
node              470 kernel/bpf/lpm_trie.c 	if (!node || node->prefixlen != key->prefixlen ||
node              471 kernel/bpf/lpm_trie.c 	    node->prefixlen != matchlen ||
node              472 kernel/bpf/lpm_trie.c 	    (node->flags & LPM_TREE_NODE_FLAG_IM)) {
node              482 kernel/bpf/lpm_trie.c 	if (rcu_access_pointer(node->child[0]) &&
node              483 kernel/bpf/lpm_trie.c 	    rcu_access_pointer(node->child[1])) {
node              484 kernel/bpf/lpm_trie.c 		node->flags |= LPM_TREE_NODE_FLAG_IM;
node              496 kernel/bpf/lpm_trie.c 	    !node->child[0] && !node->child[1]) {
node              497 kernel/bpf/lpm_trie.c 		if (node == rcu_access_pointer(parent->child[0]))
node              504 kernel/bpf/lpm_trie.c 		kfree_rcu(node, rcu);
node              512 kernel/bpf/lpm_trie.c 	if (node->child[0])
node              513 kernel/bpf/lpm_trie.c 		rcu_assign_pointer(*trim, rcu_access_pointer(node->child[0]));
node              514 kernel/bpf/lpm_trie.c 	else if (node->child[1])
node              515 kernel/bpf/lpm_trie.c 		rcu_assign_pointer(*trim, rcu_access_pointer(node->child[1]));
node              518 kernel/bpf/lpm_trie.c 	kfree_rcu(node, rcu);
node              590 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node;
node              606 kernel/bpf/lpm_trie.c 			node = rcu_dereference_protected(*slot, 1);
node              607 kernel/bpf/lpm_trie.c 			if (!node)
node              610 kernel/bpf/lpm_trie.c 			if (rcu_access_pointer(node->child[0])) {
node              611 kernel/bpf/lpm_trie.c 				slot = &node->child[0];
node              615 kernel/bpf/lpm_trie.c 			if (rcu_access_pointer(node->child[1])) {
node              616 kernel/bpf/lpm_trie.c 				slot = &node->child[1];
node              620 kernel/bpf/lpm_trie.c 			kfree(node);
node              632 kernel/bpf/lpm_trie.c 	struct lpm_trie_node *node, *next_node = NULL, *parent, *search_root;
node              667 kernel/bpf/lpm_trie.c 	for (node = search_root; node;) {
node              668 kernel/bpf/lpm_trie.c 		node_stack[++stack_ptr] = node;
node              669 kernel/bpf/lpm_trie.c 		matchlen = longest_prefix_match(trie, node, key);
node              670 kernel/bpf/lpm_trie.c 		if (node->prefixlen != matchlen ||
node              671 kernel/bpf/lpm_trie.c 		    node->prefixlen == key->prefixlen)
node              674 kernel/bpf/lpm_trie.c 		next_bit = extract_bit(key->data, node->prefixlen);
node              675 kernel/bpf/lpm_trie.c 		node = rcu_dereference(node->child[next_bit]);
node              677 kernel/bpf/lpm_trie.c 	if (!node || node->prefixlen != key->prefixlen ||
node              678 kernel/bpf/lpm_trie.c 	    (node->flags & LPM_TREE_NODE_FLAG_IM))
node              684 kernel/bpf/lpm_trie.c 	node = node_stack[stack_ptr];
node              687 kernel/bpf/lpm_trie.c 		if (rcu_dereference(parent->child[0]) == node) {
node              697 kernel/bpf/lpm_trie.c 		node = parent;
node              709 kernel/bpf/lpm_trie.c 	for (node = search_root; node;) {
node              710 kernel/bpf/lpm_trie.c 		if (node->flags & LPM_TREE_NODE_FLAG_IM) {
node              711 kernel/bpf/lpm_trie.c 			node = rcu_dereference(node->child[0]);
node              713 kernel/bpf/lpm_trie.c 			next_node = node;
node              714 kernel/bpf/lpm_trie.c 			node = rcu_dereference(node->child[0]);
node              715 kernel/bpf/lpm_trie.c 			if (!node)
node              716 kernel/bpf/lpm_trie.c 				node = rcu_dereference(next_node->child[1]);
node               29 kernel/bpf/percpu_freelist.c 					 struct pcpu_freelist_node *node)
node               32 kernel/bpf/percpu_freelist.c 	node->next = head->first;
node               33 kernel/bpf/percpu_freelist.c 	head->first = node;
node               38 kernel/bpf/percpu_freelist.c 			struct pcpu_freelist_node *node)
node               42 kernel/bpf/percpu_freelist.c 	___pcpu_freelist_push(head, node);
node               46 kernel/bpf/percpu_freelist.c 			struct pcpu_freelist_node *node)
node               51 kernel/bpf/percpu_freelist.c 	__pcpu_freelist_push(s, node);
node               87 kernel/bpf/percpu_freelist.c 	struct pcpu_freelist_node *node;
node               94 kernel/bpf/percpu_freelist.c 		node = head->first;
node               95 kernel/bpf/percpu_freelist.c 		if (node) {
node               96 kernel/bpf/percpu_freelist.c 			head->first = node->next;
node               98 kernel/bpf/percpu_freelist.c 			return node;
node               35 kernel/bpf/xskmap.c 	struct xsk_map_node *node;
node               38 kernel/bpf/xskmap.c 	node = kzalloc(sizeof(*node), GFP_ATOMIC | __GFP_NOWARN);
node               39 kernel/bpf/xskmap.c 	if (!node)
node               44 kernel/bpf/xskmap.c 		kfree(node);
node               48 kernel/bpf/xskmap.c 	node->map = map;
node               49 kernel/bpf/xskmap.c 	node->map_entry = map_entry;
node               50 kernel/bpf/xskmap.c 	return node;
node               53 kernel/bpf/xskmap.c static void xsk_map_node_free(struct xsk_map_node *node)
node               55 kernel/bpf/xskmap.c 	xsk_map_put(node->map);
node               56 kernel/bpf/xskmap.c 	kfree(node);
node               59 kernel/bpf/xskmap.c static void xsk_map_sock_add(struct xdp_sock *xs, struct xsk_map_node *node)
node               62 kernel/bpf/xskmap.c 	list_add_tail(&node->node, &xs->map_list);
node               72 kernel/bpf/xskmap.c 	list_for_each_entry_safe(n, tmp, &xs->map_list, node) {
node               74 kernel/bpf/xskmap.c 			list_del(&n->node);
node              221 kernel/bpf/xskmap.c 	struct xsk_map_node *node;
node              247 kernel/bpf/xskmap.c 	node = xsk_map_node_alloc(m, map_entry);
node              248 kernel/bpf/xskmap.c 	if (IS_ERR(node)) {
node              250 kernel/bpf/xskmap.c 		return PTR_ERR(node);
node              265 kernel/bpf/xskmap.c 	xsk_map_sock_add(xs, node);
node              276 kernel/bpf/xskmap.c 	xsk_map_node_free(node);
node             1668 kernel/cgroup/cgroup.c 		list_for_each_entry(cfts, &css->ss->cfts, node)
node             1698 kernel/cgroup/cgroup.c 		list_for_each_entry(cfts, &css->ss->cfts, node) {
node             1711 kernel/cgroup/cgroup.c 	list_for_each_entry(cfts, &css->ss->cfts, node) {
node             4038 kernel/cgroup/cgroup.c 	list_del(&cfts->node);
node             4095 kernel/cgroup/cgroup.c 	list_add_tail(&cfts->node, &ss->cfts);
node             3417 kernel/cgroup/cpuset.c bool __cpuset_node_allowed(int node, gfp_t gfp_mask)
node             3425 kernel/cgroup/cpuset.c 	if (node_isset(node, current->mems_allowed))
node             3444 kernel/cgroup/cpuset.c 	allowed = node_isset(node, cs->mems_allowed);
node               65 kernel/cpu.c   	struct hlist_node	*node;
node              117 kernel/cpu.c   					 struct hlist_node *node);
node              122 kernel/cpu.c   					 struct hlist_node *node);
node              148 kernel/cpu.c   				 bool bringup, struct hlist_node *node,
node              153 kernel/cpu.c   	int (*cbm)(unsigned int cpu, struct hlist_node *node);
node              181 kernel/cpu.c   	if (node) {
node              183 kernel/cpu.c   		trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node);
node              184 kernel/cpu.c   		ret = cbm(cpu, node);
node              191 kernel/cpu.c   	hlist_for_each(node, &step->list) {
node              192 kernel/cpu.c   		if (lastp && node == *lastp)
node              195 kernel/cpu.c   		trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node);
node              196 kernel/cpu.c   		ret = cbm(cpu, node);
node              202 kernel/cpu.c   			*lastp = node;
node              216 kernel/cpu.c   	hlist_for_each(node, &step->list) {
node              220 kernel/cpu.c   		trace_cpuhp_multi_enter(cpu, st->target, state, cbm, node);
node              221 kernel/cpu.c   		ret = cbm(cpu, node);
node              687 kernel/cpu.c   		st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last);
node              695 kernel/cpu.c   		st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last);
node              718 kernel/cpu.c   			 struct hlist_node *node)
node              737 kernel/cpu.c   		return cpuhp_invoke_callback(cpu, state, bringup, node, NULL);
node              742 kernel/cpu.c   	st->node = node;
node              763 kernel/cpu.c   	st->node = st->last = NULL;
node             1622 kernel/cpu.c   			    struct hlist_node *node)
node             1640 kernel/cpu.c   		ret = cpuhp_invoke_ap_callback(cpu, state, bringup, node);
node             1642 kernel/cpu.c   		ret = cpuhp_invoke_callback(cpu, state, bringup, node, NULL);
node             1644 kernel/cpu.c   	ret = cpuhp_invoke_callback(cpu, state, bringup, node, NULL);
node             1656 kernel/cpu.c   				   struct hlist_node *node)
node             1670 kernel/cpu.c   			cpuhp_issue_call(cpu, state, false, node);
node             1675 kernel/cpu.c   					  struct hlist_node *node,
node             1704 kernel/cpu.c   		ret = cpuhp_issue_call(cpu, state, true, node);
node             1707 kernel/cpu.c   				cpuhp_rollback_install(cpu, state, node);
node             1713 kernel/cpu.c   	hlist_add_head(node, &sp->list);
node             1719 kernel/cpu.c   int __cpuhp_state_add_instance(enum cpuhp_state state, struct hlist_node *node,
node             1725 kernel/cpu.c   	ret = __cpuhp_state_add_instance_cpuslocked(state, node, invoke);
node             1824 kernel/cpu.c   				  struct hlist_node *node, bool invoke)
node             1849 kernel/cpu.c   			cpuhp_issue_call(cpu, state, false, node);
node             1853 kernel/cpu.c   	hlist_del(node);
node              337 kernel/dma/coherent.c 	unsigned long node = rmem->fdt_node;
node              339 kernel/dma/coherent.c 	if (of_get_flat_dt_prop(node, "reusable", NULL))
node              343 kernel/dma/coherent.c 	if (!of_get_flat_dt_prop(node, "no-map", NULL)) {
node              348 kernel/dma/coherent.c 	if (of_get_flat_dt_prop(node, "linux,dma-default", NULL)) {
node              303 kernel/dma/contiguous.c 	unsigned long node = rmem->fdt_node;
node              307 kernel/dma/contiguous.c 	if (!of_get_flat_dt_prop(node, "reusable", NULL) ||
node              308 kernel/dma/contiguous.c 	    of_get_flat_dt_prop(node, "no-map", NULL))
node              324 kernel/dma/contiguous.c 	if (of_get_flat_dt_prop(node, "linux,cma-default", NULL))
node               90 kernel/dma/direct.c 	int node = dev_to_node(dev);
node              108 kernel/dma/direct.c 		page = alloc_pages_node(node, gfp, get_order(alloc_size));
node             1557 kernel/events/core.c 	struct rb_node **node;
node             1561 kernel/events/core.c 	node = &groups->tree.rb_node;
node             1562 kernel/events/core.c 	parent = *node;
node             1564 kernel/events/core.c 	while (*node) {
node             1565 kernel/events/core.c 		parent = *node;
node             1566 kernel/events/core.c 		node_event = container_of(*node, struct perf_event, group_node);
node             1569 kernel/events/core.c 			node = &parent->rb_left;
node             1571 kernel/events/core.c 			node = &parent->rb_right;
node             1574 kernel/events/core.c 	rb_link_node(&event->group_node, parent, node);
node             1623 kernel/events/core.c 	struct rb_node *node = groups->tree.rb_node;
node             1625 kernel/events/core.c 	while (node) {
node             1626 kernel/events/core.c 		node_event = container_of(node, struct perf_event, group_node);
node             1629 kernel/events/core.c 			node = node->rb_left;
node             1631 kernel/events/core.c 			node = node->rb_right;
node             1634 kernel/events/core.c 			node = node->rb_left;
node             9158 kernel/events/core.c 	int node = cpu_to_node(event->cpu == -1 ? 0 : event->cpu);
node             9161 kernel/events/core.c 	filter = kzalloc_node(sizeof(*filter), GFP_KERNEL, node);
node              567 kernel/events/ring_buffer.c static struct page *rb_alloc_aux_page(int node, int order)
node              575 kernel/events/ring_buffer.c 		page = alloc_pages_node(node, PERF_AUX_GFP, order);
node              633 kernel/events/ring_buffer.c 	int node = (event->cpu == -1) ? -1 : cpu_to_node(event->cpu);
node              657 kernel/events/ring_buffer.c 				     node);
node              667 kernel/events/ring_buffer.c 		page = rb_alloc_aux_page(node, order);
node              747 kernel/events/ring_buffer.c 	int node;
node              749 kernel/events/ring_buffer.c 	node = (cpu == -1) ? cpu : cpu_to_node(cpu);
node              750 kernel/events/ring_buffer.c 	page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
node              167 kernel/fork.c  static inline struct task_struct *alloc_task_struct_node(int node)
node              169 kernel/fork.c  	return kmem_cache_alloc_node(task_struct_cachep, GFP_KERNEL, node);
node              213 kernel/fork.c  static unsigned long *alloc_thread_stack_node(struct task_struct *tsk, int node)
node              244 kernel/fork.c  				     0, node, __builtin_return_address(0));
node              257 kernel/fork.c  	struct page *page = alloc_pages_node(node, THREADINFO_GFP,
node              303 kernel/fork.c  						  int node)
node              306 kernel/fork.c  	stack = kmem_cache_alloc_node(thread_stack_cache, THREADINFO_GFP, node);
node              856 kernel/fork.c  static struct task_struct *dup_task_struct(struct task_struct *orig, int node)
node              863 kernel/fork.c  	if (node == NUMA_NO_NODE)
node              864 kernel/fork.c  		node = tsk_fork_get_node(orig);
node              865 kernel/fork.c  	tsk = alloc_task_struct_node(node);
node              869 kernel/fork.c  	stack = alloc_thread_stack_node(tsk, node);
node             1763 kernel/fork.c  					int node,
node             1835 kernel/fork.c  	INIT_HLIST_NODE(&delayed.node);
node             1839 kernel/fork.c  		hlist_add_head(&delayed.node, &current->signal->multiprocess);
node             1847 kernel/fork.c  	p = dup_task_struct(current, node);
node             2211 kernel/fork.c  	hlist_del_init(&delayed.node);
node             2285 kernel/fork.c  	hlist_del_init(&delayed.node);
node              145 kernel/gcov/fs.c static struct gcov_info *get_node_info(struct gcov_node *node)
node              147 kernel/gcov/fs.c 	if (node->num_loaded > 0)
node              148 kernel/gcov/fs.c 		return node->loaded_info[0];
node              150 kernel/gcov/fs.c 	return node->unloaded_info;
node              157 kernel/gcov/fs.c static struct gcov_info *get_accumulated_info(struct gcov_node *node)
node              162 kernel/gcov/fs.c 	if (node->unloaded_info)
node              163 kernel/gcov/fs.c 		info = gcov_info_dup(node->unloaded_info);
node              165 kernel/gcov/fs.c 		info = gcov_info_dup(node->loaded_info[i++]);
node              168 kernel/gcov/fs.c 	for (; i < node->num_loaded; i++)
node              169 kernel/gcov/fs.c 		gcov_info_add(info, node->loaded_info[i]);
node              180 kernel/gcov/fs.c 	struct gcov_node *node = inode->i_private;
node              192 kernel/gcov/fs.c 	info = get_accumulated_info(node);
node              240 kernel/gcov/fs.c 	struct gcov_node *node;
node              243 kernel/gcov/fs.c 	list_for_each_entry(node, &all_head, all) {
node              244 kernel/gcov/fs.c 		info = get_node_info(node);
node              246 kernel/gcov/fs.c 			return node;
node              255 kernel/gcov/fs.c static void reset_node(struct gcov_node *node)
node              259 kernel/gcov/fs.c 	if (node->unloaded_info)
node              260 kernel/gcov/fs.c 		gcov_info_reset(node->unloaded_info);
node              261 kernel/gcov/fs.c 	for (i = 0; i < node->num_loaded; i++)
node              262 kernel/gcov/fs.c 		gcov_info_reset(node->loaded_info[i]);
node              265 kernel/gcov/fs.c static void remove_node(struct gcov_node *node);
node              277 kernel/gcov/fs.c 	struct gcov_node *node;
node              282 kernel/gcov/fs.c 	node = get_node_by_name(gcov_info_filename(info));
node              283 kernel/gcov/fs.c 	if (node) {
node              285 kernel/gcov/fs.c 		if (node->num_loaded == 0)
node              286 kernel/gcov/fs.c 			remove_node(node);
node              288 kernel/gcov/fs.c 			reset_node(node);
node              366 kernel/gcov/fs.c static void add_links(struct gcov_node *node, struct dentry *parent)
node              375 kernel/gcov/fs.c 	node->links = kcalloc(num, sizeof(struct dentry *), GFP_KERNEL);
node              376 kernel/gcov/fs.c 	if (!node->links)
node              380 kernel/gcov/fs.c 				gcov_info_filename(get_node_info(node)),
node              387 kernel/gcov/fs.c 		node->links[i] = debugfs_create_symlink(deskew(basename),
node              396 kernel/gcov/fs.c 		debugfs_remove(node->links[i]);
node              397 kernel/gcov/fs.c 	kfree(node->links);
node              398 kernel/gcov/fs.c 	node->links = NULL;
node              410 kernel/gcov/fs.c static void init_node(struct gcov_node *node, struct gcov_info *info,
node              413 kernel/gcov/fs.c 	INIT_LIST_HEAD(&node->list);
node              414 kernel/gcov/fs.c 	INIT_LIST_HEAD(&node->children);
node              415 kernel/gcov/fs.c 	INIT_LIST_HEAD(&node->all);
node              416 kernel/gcov/fs.c 	if (node->loaded_info) {
node              417 kernel/gcov/fs.c 		node->loaded_info[0] = info;
node              418 kernel/gcov/fs.c 		node->num_loaded = 1;
node              420 kernel/gcov/fs.c 	node->parent = parent;
node              422 kernel/gcov/fs.c 		strcpy(node->name, name);
node              432 kernel/gcov/fs.c 	struct gcov_node *node;
node              434 kernel/gcov/fs.c 	node = kzalloc(sizeof(struct gcov_node) + strlen(name) + 1, GFP_KERNEL);
node              435 kernel/gcov/fs.c 	if (!node)
node              438 kernel/gcov/fs.c 		node->loaded_info = kcalloc(1, sizeof(struct gcov_info *),
node              440 kernel/gcov/fs.c 		if (!node->loaded_info)
node              443 kernel/gcov/fs.c 	init_node(node, info, name, parent);
node              446 kernel/gcov/fs.c 		node->dentry = debugfs_create_file(deskew(node->name), 0600,
node              447 kernel/gcov/fs.c 					parent->dentry, node, &gcov_data_fops);
node              449 kernel/gcov/fs.c 		node->dentry = debugfs_create_dir(node->name, parent->dentry);
node              451 kernel/gcov/fs.c 		add_links(node, parent->dentry);
node              452 kernel/gcov/fs.c 	list_add(&node->list, &parent->children);
node              453 kernel/gcov/fs.c 	list_add(&node->all, &all_head);
node              455 kernel/gcov/fs.c 	return node;
node              458 kernel/gcov/fs.c 	kfree(node);
node              464 kernel/gcov/fs.c static void remove_links(struct gcov_node *node)
node              468 kernel/gcov/fs.c 	if (!node->links)
node              471 kernel/gcov/fs.c 		debugfs_remove(node->links[i]);
node              472 kernel/gcov/fs.c 	kfree(node->links);
node              473 kernel/gcov/fs.c 	node->links = NULL;
node              480 kernel/gcov/fs.c static void release_node(struct gcov_node *node)
node              482 kernel/gcov/fs.c 	list_del(&node->list);
node              483 kernel/gcov/fs.c 	list_del(&node->all);
node              484 kernel/gcov/fs.c 	debugfs_remove(node->dentry);
node              485 kernel/gcov/fs.c 	remove_links(node);
node              486 kernel/gcov/fs.c 	kfree(node->loaded_info);
node              487 kernel/gcov/fs.c 	if (node->unloaded_info)
node              488 kernel/gcov/fs.c 		gcov_info_free(node->unloaded_info);
node              489 kernel/gcov/fs.c 	kfree(node);
node              493 kernel/gcov/fs.c static void remove_node(struct gcov_node *node)
node              497 kernel/gcov/fs.c 	while ((node != &root_node) && list_empty(&node->children)) {
node              498 kernel/gcov/fs.c 		parent = node->parent;
node              499 kernel/gcov/fs.c 		release_node(node);
node              500 kernel/gcov/fs.c 		node = parent;
node              511 kernel/gcov/fs.c 	struct gcov_node *node;
node              513 kernel/gcov/fs.c 	list_for_each_entry(node, &parent->children, list) {
node              514 kernel/gcov/fs.c 		if (strcmp(node->name, name) == 0)
node              515 kernel/gcov/fs.c 			return node;
node              528 kernel/gcov/fs.c 	struct gcov_node *node;
node              532 kernel/gcov/fs.c 	list_for_each_entry(node, &all_head, all) {
node              533 kernel/gcov/fs.c 		if (node->num_loaded > 0)
node              534 kernel/gcov/fs.c 			reset_node(node);
node              535 kernel/gcov/fs.c 		else if (list_empty(&node->children)) {
node              536 kernel/gcov/fs.c 			remove_node(node);
node              570 kernel/gcov/fs.c 	struct gcov_node *node;
node              589 kernel/gcov/fs.c 		node = get_child_by_name(parent, curr);
node              590 kernel/gcov/fs.c 		if (!node) {
node              591 kernel/gcov/fs.c 			node = new_node(parent, NULL, curr);
node              592 kernel/gcov/fs.c 			if (!node)
node              595 kernel/gcov/fs.c 		parent = node;
node              598 kernel/gcov/fs.c 	node = new_node(parent, info, curr);
node              599 kernel/gcov/fs.c 	if (!node)
node              614 kernel/gcov/fs.c static void add_info(struct gcov_node *node, struct gcov_info *info)
node              617 kernel/gcov/fs.c 	int num = node->num_loaded;
node              630 kernel/gcov/fs.c 	memcpy(loaded_info, node->loaded_info,
node              639 kernel/gcov/fs.c 		if (!gcov_info_is_compatible(node->unloaded_info, info)) {
node              643 kernel/gcov/fs.c 			gcov_info_free(node->unloaded_info);
node              644 kernel/gcov/fs.c 			node->unloaded_info = NULL;
node              651 kernel/gcov/fs.c 		if (!gcov_info_is_compatible(node->loaded_info[0], info)) {
node              659 kernel/gcov/fs.c 	kfree(node->loaded_info);
node              660 kernel/gcov/fs.c 	node->loaded_info = loaded_info;
node              661 kernel/gcov/fs.c 	node->num_loaded = num + 1;
node              667 kernel/gcov/fs.c static int get_info_index(struct gcov_node *node, struct gcov_info *info)
node              671 kernel/gcov/fs.c 	for (i = 0; i < node->num_loaded; i++) {
node              672 kernel/gcov/fs.c 		if (node->loaded_info[i] == info)
node              681 kernel/gcov/fs.c static void save_info(struct gcov_node *node, struct gcov_info *info)
node              683 kernel/gcov/fs.c 	if (node->unloaded_info)
node              684 kernel/gcov/fs.c 		gcov_info_add(node->unloaded_info, info);
node              686 kernel/gcov/fs.c 		node->unloaded_info = gcov_info_dup(info);
node              687 kernel/gcov/fs.c 		if (!node->unloaded_info) {
node              699 kernel/gcov/fs.c static void remove_info(struct gcov_node *node, struct gcov_info *info)
node              703 kernel/gcov/fs.c 	i = get_info_index(node, info);
node              710 kernel/gcov/fs.c 		save_info(node, info);
node              712 kernel/gcov/fs.c 	node->loaded_info[i] = node->loaded_info[node->num_loaded - 1];
node              713 kernel/gcov/fs.c 	node->num_loaded--;
node              714 kernel/gcov/fs.c 	if (node->num_loaded > 0)
node              717 kernel/gcov/fs.c 	kfree(node->loaded_info);
node              718 kernel/gcov/fs.c 	node->loaded_info = NULL;
node              719 kernel/gcov/fs.c 	node->num_loaded = 0;
node              720 kernel/gcov/fs.c 	if (!node->unloaded_info)
node              721 kernel/gcov/fs.c 		remove_node(node);
node              730 kernel/gcov/fs.c 	struct gcov_node *node;
node              733 kernel/gcov/fs.c 	node = get_node_by_name(gcov_info_filename(info));
node              736 kernel/gcov/fs.c 		if (node)
node              737 kernel/gcov/fs.c 			add_info(node, info);
node              742 kernel/gcov/fs.c 		if (node)
node              743 kernel/gcov/fs.c 			remove_info(node, info);
node               46 kernel/irq/affinity.c 	int node;
node               52 kernel/irq/affinity.c 	for (node = 0; node < nr_node_ids; node++) {
node               53 kernel/irq/affinity.c 		if (!zalloc_cpumask_var(&masks[node], GFP_KERNEL))
node               60 kernel/irq/affinity.c 	while (--node >= 0)
node               61 kernel/irq/affinity.c 		free_cpumask_var(masks[node]);
node               68 kernel/irq/affinity.c 	int node;
node               70 kernel/irq/affinity.c 	for (node = 0; node < nr_node_ids; node++)
node               71 kernel/irq/affinity.c 		free_cpumask_var(masks[node]);
node              179 kernel/irq/devres.c 			   unsigned int cnt, int node, struct module *owner,
node              189 kernel/irq/devres.c 	base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity);
node              104 kernel/irq/internals.h extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr);
node               55 kernel/irq/irqdesc.c static int alloc_masks(struct irq_desc *desc, int node)
node               58 kernel/irq/irqdesc.c 				     GFP_KERNEL, node))
node               63 kernel/irq/irqdesc.c 				     GFP_KERNEL, node)) {
node               70 kernel/irq/irqdesc.c 	if (!zalloc_cpumask_var_node(&desc->pending_mask, GFP_KERNEL, node)) {
node               81 kernel/irq/irqdesc.c static void desc_smp_init(struct irq_desc *desc, int node,
node               92 kernel/irq/irqdesc.c 	desc->irq_common_data.node = node;
node               98 kernel/irq/irqdesc.c alloc_masks(struct irq_desc *desc, int node) { return 0; }
node              100 kernel/irq/irqdesc.c desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { }
node              103 kernel/irq/irqdesc.c static void desc_set_defaults(unsigned int irq, struct irq_desc *desc, int node,
node              127 kernel/irq/irqdesc.c 	desc_smp_init(desc, node, affinity);
node              387 kernel/irq/irqdesc.c static struct irq_desc *alloc_desc(int irq, int node, unsigned int flags,
node              393 kernel/irq/irqdesc.c 	desc = kzalloc_node(sizeof(*desc), GFP_KERNEL, node);
node              401 kernel/irq/irqdesc.c 	if (alloc_masks(desc, node))
node              409 kernel/irq/irqdesc.c 	desc_set_defaults(irq, desc, node, affinity, owner);
node              466 kernel/irq/irqdesc.c static int alloc_descs(unsigned int start, unsigned int cnt, int node,
node              491 kernel/irq/irqdesc.c 			node = cpu_to_node(cpumask_first(mask));
node              495 kernel/irq/irqdesc.c 		desc = alloc_desc(start + i, node, flags, mask, owner);
node              521 kernel/irq/irqdesc.c 	int i, initcnt, node = first_online_node;
node              541 kernel/irq/irqdesc.c 		desc = alloc_desc(i, node, 0, NULL, NULL);
node              560 kernel/irq/irqdesc.c 	int count, i, node = first_online_node;
node              572 kernel/irq/irqdesc.c 		alloc_masks(&desc[i], node);
node              576 kernel/irq/irqdesc.c 		desc_set_defaults(i, &desc[i], node, NULL, NULL);
node              597 kernel/irq/irqdesc.c static inline int alloc_descs(unsigned int start, unsigned int cnt, int node,
node              766 kernel/irq/irqdesc.c __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node,
node              800 kernel/irq/irqdesc.c 	ret = alloc_descs(start, cnt, node, affinity, owner);
node              815 kernel/irq/irqdesc.c unsigned int irq_alloc_hwirqs(int cnt, int node)
node              817 kernel/irq/irqdesc.c 	int i, irq = __irq_alloc_descs(-1, 0, cnt, node, NULL, NULL);
node              823 kernel/irq/irqdesc.c 		if (arch_setup_hwirq(i, node))
node             1011 kernel/irq/irqdomain.c 			   int node, const struct irq_affinity_desc *affinity)
node             1016 kernel/irq/irqdomain.c 		virq = __irq_alloc_descs(virq, virq, cnt, node, THIS_MODULE,
node             1022 kernel/irq/irqdomain.c 		virq = __irq_alloc_descs(-1, hint, cnt, node, THIS_MODULE,
node             1025 kernel/irq/irqdomain.c 			virq = __irq_alloc_descs(-1, 1, cnt, node, THIS_MODULE,
node             1327 kernel/irq/irqdomain.c 			    unsigned int nr_irqs, int node, void *arg,
node             1341 kernel/irq/irqdomain.c 		virq = irq_domain_alloc_descs(irq_base, nr_irqs, 0, node,
node              409 kernel/irq/manage.c 	int ret, node = irq_desc_get_node(desc);
node              435 kernel/irq/manage.c 	if (node != NUMA_NO_NODE) {
node              436 kernel/irq/manage.c 		const struct cpumask *nodemask = cpumask_of_node(node);
node               37 kernel/kthread.c 	int node;
node              275 kernel/kthread.c 	current->pref_node_fork = create->node;
node              294 kernel/kthread.c 						    void *data, int node,
node              307 kernel/kthread.c 	create->node = node;
node              380 kernel/kthread.c 					   void *data, int node,
node              388 kernel/kthread.c 	task = __kthread_create_on_node(threadfn, data, node, namefmt, args);
node              663 kernel/kthread.c 					struct kthread_work, node);
node              664 kernel/kthread.c 		list_del_init(&work->node);
node              687 kernel/kthread.c 	int node = NUMA_NO_NODE;
node              696 kernel/kthread.c 		node = cpu_to_node(cpu);
node              699 kernel/kthread.c 						node, namefmt, args);
node              781 kernel/kthread.c 	return !list_empty(&work->node) || work->canceling;
node              788 kernel/kthread.c 	WARN_ON_ONCE(!list_empty(&work->node));
node              800 kernel/kthread.c 	list_add_tail(&work->node, pos);
node              861 kernel/kthread.c 	WARN_ON_ONCE(list_empty(&work->node));
node              862 kernel/kthread.c 	list_del_init(&work->node);
node              892 kernel/kthread.c 	list_add(&work->node, &worker->delayed_work_list);
node              968 kernel/kthread.c 	if (!list_empty(&work->node))
node              969 kernel/kthread.c 		kthread_insert_work(worker, &fwork.work, work->node.next);
node             1019 kernel/kthread.c 	if (!list_empty(&work->node)) {
node             1020 kernel/kthread.c 		list_del_init(&work->node);
node              583 kernel/livepatch/core.c 		list_del(&func->node);
node              613 kernel/livepatch/core.c 		list_del(&obj->node);
node              802 kernel/livepatch/core.c 	list_add_tail(&func->node, &obj->func_list);
node              810 kernel/livepatch/core.c 	list_add_tail(&obj->node, &patch->obj_list);
node               30 kernel/livepatch/patch.c 	list_for_each_entry(ops, &klp_ops, node) {
node              160 kernel/livepatch/patch.c 		list_del(&ops->node);
node              201 kernel/livepatch/patch.c 		list_add(&ops->node, &klp_ops);
node              232 kernel/livepatch/patch.c 	list_del(&ops->node);
node               23 kernel/livepatch/patch.h 	struct list_head node;
node               55 kernel/livepatch/shadow.c 	struct hlist_node node;
node               89 kernel/livepatch/shadow.c 	hash_for_each_possible_rcu(klp_shadow_hash, shadow, node,
node              157 kernel/livepatch/shadow.c 	hash_add_rcu(klp_shadow_hash, &new_shadow->node,
node              237 kernel/livepatch/shadow.c 	hash_del_rcu(&shadow->node);
node              261 kernel/livepatch/shadow.c 	hash_for_each_possible(klp_shadow_hash, shadow, node,
node              292 kernel/livepatch/shadow.c 	hash_for_each(klp_shadow_hash, i, shadow, node) {
node               65 kernel/locking/mcs_spinlock.h void mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node)
node               70 kernel/locking/mcs_spinlock.h 	node->locked = 0;
node               71 kernel/locking/mcs_spinlock.h 	node->next   = NULL;
node               79 kernel/locking/mcs_spinlock.h 	prev = xchg(lock, node);
node               91 kernel/locking/mcs_spinlock.h 	WRITE_ONCE(prev->next, node);
node               94 kernel/locking/mcs_spinlock.h 	arch_mcs_spin_lock_contended(&node->locked);
node              102 kernel/locking/mcs_spinlock.h void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *node)
node              104 kernel/locking/mcs_spinlock.h 	struct mcs_spinlock *next = READ_ONCE(node->next);
node              110 kernel/locking/mcs_spinlock.h 		if (likely(cmpxchg_release(lock, node, NULL) == node))
node              113 kernel/locking/mcs_spinlock.h 		while (!(next = READ_ONCE(node->next)))
node               25 kernel/locking/osq_lock.c static inline int node_cpu(struct optimistic_spin_node *node)
node               27 kernel/locking/osq_lock.c 	return node->cpu - 1;
node               43 kernel/locking/osq_lock.c 	      struct optimistic_spin_node *node,
node               78 kernel/locking/osq_lock.c 		if (node->next) {
node               79 kernel/locking/osq_lock.c 			next = xchg(&node->next, NULL);
node               92 kernel/locking/osq_lock.c 	struct optimistic_spin_node *node = this_cpu_ptr(&osq_node);
node               97 kernel/locking/osq_lock.c 	node->locked = 0;
node               98 kernel/locking/osq_lock.c 	node->next = NULL;
node               99 kernel/locking/osq_lock.c 	node->cpu = curr;
node              112 kernel/locking/osq_lock.c 	node->prev = prev;
node              126 kernel/locking/osq_lock.c 	WRITE_ONCE(prev->next, node);
node              137 kernel/locking/osq_lock.c 	while (!READ_ONCE(node->locked)) {
node              143 kernel/locking/osq_lock.c 		if (need_resched() || vcpu_is_preempted(node_cpu(node->prev)))
node              160 kernel/locking/osq_lock.c 		if (prev->next == node &&
node              161 kernel/locking/osq_lock.c 		    cmpxchg(&prev->next, node, NULL) == node)
node              169 kernel/locking/osq_lock.c 		if (smp_load_acquire(&node->locked))
node              178 kernel/locking/osq_lock.c 		prev = READ_ONCE(node->prev);
node              188 kernel/locking/osq_lock.c 	next = osq_wait_next(lock, node, prev);
node              208 kernel/locking/osq_lock.c 	struct optimistic_spin_node *node, *next;
node              221 kernel/locking/osq_lock.c 	node = this_cpu_ptr(&osq_node);
node              222 kernel/locking/osq_lock.c 	next = xchg(&node->next, NULL);
node              228 kernel/locking/osq_lock.c 	next = osq_wait_next(lock, node, NULL);
node              271 kernel/locking/qspinlock.c static __always_inline void __pv_init_node(struct mcs_spinlock *node) { }
node              272 kernel/locking/qspinlock.c static __always_inline void __pv_wait_node(struct mcs_spinlock *node,
node              275 kernel/locking/qspinlock.c 					   struct mcs_spinlock *node) { }
node              277 kernel/locking/qspinlock.c 						   struct mcs_spinlock *node)
node              316 kernel/locking/qspinlock.c 	struct mcs_spinlock *prev, *next, *node;
node              399 kernel/locking/qspinlock.c 	node = this_cpu_ptr(&qnodes[0].mcs);
node              400 kernel/locking/qspinlock.c 	idx = node->count++;
node              419 kernel/locking/qspinlock.c 	node = grab_mcs_node(node, idx);
node              433 kernel/locking/qspinlock.c 	node->locked = 0;
node              434 kernel/locking/qspinlock.c 	node->next = NULL;
node              435 kernel/locking/qspinlock.c 	pv_init_node(node);
node              470 kernel/locking/qspinlock.c 		WRITE_ONCE(prev->next, node);
node              472 kernel/locking/qspinlock.c 		pv_wait_node(node, prev);
node              473 kernel/locking/qspinlock.c 		arch_mcs_spin_lock_contended(&node->locked);
node              481 kernel/locking/qspinlock.c 		next = READ_ONCE(node->next);
node              507 kernel/locking/qspinlock.c 	if ((val = pv_wait_head_or_lock(lock, node)))
node              550 kernel/locking/qspinlock.c 		next = smp_cond_load_relaxed(&node->next, (VAL));
node              173 kernel/locking/qspinlock_paravirt.h 	struct pv_node   *node;
node              212 kernel/locking/qspinlock_paravirt.h static struct qspinlock **pv_hash(struct qspinlock *lock, struct pv_node *node)
node              221 kernel/locking/qspinlock_paravirt.h 			WRITE_ONCE(he->node, node);
node              243 kernel/locking/qspinlock_paravirt.h 	struct pv_node *node;
node              247 kernel/locking/qspinlock_paravirt.h 			node = READ_ONCE(he->node);
node              249 kernel/locking/qspinlock_paravirt.h 			return node;
node              278 kernel/locking/qspinlock_paravirt.h static void pv_init_node(struct mcs_spinlock *node)
node              280 kernel/locking/qspinlock_paravirt.h 	struct pv_node *pn = (struct pv_node *)node;
node              293 kernel/locking/qspinlock_paravirt.h static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev)
node              295 kernel/locking/qspinlock_paravirt.h 	struct pv_node *pn = (struct pv_node *)node;
node              302 kernel/locking/qspinlock_paravirt.h 			if (READ_ONCE(node->locked))
node              322 kernel/locking/qspinlock_paravirt.h 		if (!READ_ONCE(node->locked)) {
node              343 kernel/locking/qspinlock_paravirt.h 				  !READ_ONCE(node->locked));
node              360 kernel/locking/qspinlock_paravirt.h static void pv_kick_node(struct qspinlock *lock, struct mcs_spinlock *node)
node              362 kernel/locking/qspinlock_paravirt.h 	struct pv_node *pn = (struct pv_node *)node;
node              403 kernel/locking/qspinlock_paravirt.h pv_wait_head_or_lock(struct qspinlock *lock, struct mcs_spinlock *node)
node              405 kernel/locking/qspinlock_paravirt.h 	struct pv_node *pn = (struct pv_node *)node;
node              495 kernel/locking/qspinlock_paravirt.h 	struct pv_node *node;
node              517 kernel/locking/qspinlock_paravirt.h 	node = pv_unhash(lock);
node              533 kernel/locking/qspinlock_paravirt.h 	pv_kick(node->cpu);
node              107 kernel/module.c 	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
node              114 kernel/module.c 	struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
node              158 kernel/module.c static noinline void __mod_tree_insert(struct mod_tree_node *node)
node              160 kernel/module.c 	latch_tree_insert(&node->node, &mod_tree.root, &mod_tree_ops);
node              163 kernel/module.c static void __mod_tree_remove(struct mod_tree_node *node)
node              165 kernel/module.c 	latch_tree_erase(&node->node, &mod_tree.root, &mod_tree_ops);
node              202 kernel/module.c 	return container_of(ltn, struct mod_tree_node, node)->mod;
node             3527 kernel/module.c 	struct llist_node node;
node             3541 kernel/module.c 		initfree = container_of(pos, struct mod_initfree, node);
node             3650 kernel/module.c 	if (llist_add(&freeinit->node, &init_free_list))
node              780 kernel/padata.c static int padata_cpu_online(unsigned int cpu, struct hlist_node *node)
node              785 kernel/padata.c 	pinst = hlist_entry_safe(node, struct padata_instance, cpu_online_node);
node              795 kernel/padata.c static int padata_cpu_dead(unsigned int cpu, struct hlist_node *node)
node              800 kernel/padata.c 	pinst = hlist_entry_safe(node, struct padata_instance, cpu_dead_node);
node              104 kernel/power/qos.c 	struct plist_node *node;
node              118 kernel/power/qos.c 		plist_for_each(node, &c->list)
node              119 kernel/power/qos.c 			total_value += node->prio;
node              181 kernel/power/qos.c 	plist_for_each_entry(req, &c->list, node) {
node              184 kernel/power/qos.c 		if ((req->node).prio != c->default_value) {
node              190 kernel/power/qos.c 			   (req->node).prio, state);
node              214 kernel/power/qos.c int pm_qos_update_target(struct pm_qos_constraints *c, struct plist_node *node,
node              230 kernel/power/qos.c 		plist_del(node, &c->list);
node              238 kernel/power/qos.c 		plist_del(node, &c->list);
node              241 kernel/power/qos.c 		plist_node_init(node, new_value);
node              242 kernel/power/qos.c 		plist_add(node, &c->list);
node              277 kernel/power/qos.c 	list_del(&req->node);
node              278 kernel/power/qos.c 	list_for_each_entry(req, &pqf->list, node)
node              315 kernel/power/qos.c 		INIT_LIST_HEAD(&req->node);
node              316 kernel/power/qos.c 		list_add_tail(&req->node, &pqf->list);
node              355 kernel/power/qos.c 	if (new_value != req->node.prio)
node              358 kernel/power/qos.c 			&req->node, PM_QOS_UPDATE_REQ, new_value);
node              403 kernel/power/qos.c 			     &req->node, PM_QOS_ADD_REQ, value);
node              454 kernel/power/qos.c 	if (new_value != req->node.prio)
node              457 kernel/power/qos.c 			&req->node, PM_QOS_UPDATE_REQ, new_value);
node              485 kernel/power/qos.c 			     &req->node, PM_QOS_REMOVE_REQ,
node              372 kernel/power/snapshot.c 	struct rtree_node *node;
node              406 kernel/power/snapshot.c 	struct rtree_node *node;
node              408 kernel/power/snapshot.c 	node = chain_alloc(ca, sizeof(struct rtree_node));
node              409 kernel/power/snapshot.c 	if (!node)
node              412 kernel/power/snapshot.c 	node->data = get_image_page(gfp_mask, safe_needed);
node              413 kernel/power/snapshot.c 	if (!node->data)
node              416 kernel/power/snapshot.c 	list_add_tail(&node->list, list);
node              418 kernel/power/snapshot.c 	return node;
node              431 kernel/power/snapshot.c 	struct rtree_node *node, *block, **dst;
node              446 kernel/power/snapshot.c 		node = alloc_rtree_node(gfp_mask, safe_needed, ca,
node              448 kernel/power/snapshot.c 		if (!node)
node              451 kernel/power/snapshot.c 		node->data[0] = (unsigned long)zone->rtree;
node              452 kernel/power/snapshot.c 		zone->rtree = node;
node              462 kernel/power/snapshot.c 	node = zone->rtree;
node              468 kernel/power/snapshot.c 		if (!node) {
node              469 kernel/power/snapshot.c 			node = alloc_rtree_node(gfp_mask, safe_needed, ca,
node              471 kernel/power/snapshot.c 			if (!node)
node              473 kernel/power/snapshot.c 			*dst = node;
node              479 kernel/power/snapshot.c 		node = *dst;
node              539 kernel/power/snapshot.c 	struct rtree_node *node;
node              541 kernel/power/snapshot.c 	list_for_each_entry(node, &zone->nodes, list)
node              542 kernel/power/snapshot.c 		free_image_page(node->data, clear_nosave_free);
node              544 kernel/power/snapshot.c 	list_for_each_entry(node, &zone->leaves, list)
node              545 kernel/power/snapshot.c 		free_image_page(node->data, clear_nosave_free);
node              552 kernel/power/snapshot.c 	bm->cur.node = list_entry(bm->cur.zone->leaves.next,
node              711 kernel/power/snapshot.c 	struct rtree_node *node;
node              743 kernel/power/snapshot.c 	node = bm->cur.node;
node              748 kernel/power/snapshot.c 	node      = zone->rtree;
node              756 kernel/power/snapshot.c 		BUG_ON(node->data[index] == 0);
node              757 kernel/power/snapshot.c 		node = (struct rtree_node *)node->data[index];
node              763 kernel/power/snapshot.c 	bm->cur.node = node;
node              767 kernel/power/snapshot.c 	*addr = node->data;
node              813 kernel/power/snapshot.c 	clear_bit(bit, bm->cur.node->data);
node              847 kernel/power/snapshot.c 	if (!list_is_last(&bm->cur.node->list, &bm->cur.zone->leaves)) {
node              848 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.node->list.next,
node              860 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.zone->leaves.next,
node              890 kernel/power/snapshot.c 		bit	  = find_next_bit(bm->cur.node->data, bits,
node              916 kernel/power/snapshot.c 	struct rtree_node *node;
node              918 kernel/power/snapshot.c 	list_for_each_entry(node, &zone->nodes, list)
node              919 kernel/power/snapshot.c 		recycle_safe_page(node->data);
node              921 kernel/power/snapshot.c 	list_for_each_entry(node, &zone->leaves, list)
node              922 kernel/power/snapshot.c 		recycle_safe_page(node->data);
node              123 kernel/power/swap.c 	struct rb_node node;
node              138 kernel/power/swap.c 		ext = rb_entry(*new, struct swsusp_extent, node);
node              166 kernel/power/swap.c 	rb_link_node(&ext->node, parent, new);
node              167 kernel/power/swap.c 	rb_insert_color(&ext->node, &swsusp_extents);
node              198 kernel/power/swap.c 	struct rb_node *node;
node              200 kernel/power/swap.c 	while ((node = swsusp_extents.rb_node)) {
node              204 kernel/power/swap.c 		ext = rb_entry(node, struct swsusp_extent, node);
node              205 kernel/power/swap.c 		rb_erase(node, &swsusp_extents);
node               29 kernel/power/wakelock.c 	struct rb_node		node;
node               40 kernel/power/wakelock.c 	struct rb_node *node;
node               47 kernel/power/wakelock.c 	for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) {
node               48 kernel/power/wakelock.c 		wl = rb_entry(node, struct wakelock, node);
node              125 kernel/power/wakelock.c 			rb_erase(&wl->node, &wakelocks_tree);
node              153 kernel/power/wakelock.c 	struct rb_node **node = &wakelocks_tree.rb_node;
node              154 kernel/power/wakelock.c 	struct rb_node *parent = *node;
node              157 kernel/power/wakelock.c 	while (*node) {
node              160 kernel/power/wakelock.c 		parent = *node;
node              161 kernel/power/wakelock.c 		wl = rb_entry(*node, struct wakelock, node);
node              170 kernel/power/wakelock.c 			node = &(*node)->rb_left;
node              172 kernel/power/wakelock.c 			node = &(*node)->rb_right;
node              199 kernel/power/wakelock.c 	rb_link_node(&wl->node, parent, node);
node              200 kernel/power/wakelock.c 	rb_insert_color(&wl->node, &wakelocks_tree);
node              354 kernel/profile.c 	int i, node = cpu_to_mem(cpu);
node              363 kernel/profile.c 		page = __alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
node              326 kernel/rcu/rcu.h #define rcu_is_last_leaf_node(rnp) ((rnp) == &rcu_state.node[rcu_num_nodes - 1])
node              334 kernel/rcu/rcu.h 	for ((rnp) = &(sp)->node[0]; \
node              335 kernel/rcu/rcu.h 	     (rnp) < &(sp)->node[rcu_num_nodes]; (rnp)++)
node              347 kernel/rcu/rcu.h 	     (rnp) < &rcu_state.node[rcu_num_nodes]; (rnp)++)
node               94 kernel/rcu/srcutree.c 	ssp->level[0] = &ssp->node[0];
node              111 kernel/rcu/srcutree.c 		if (snp == &ssp->node[0]) {
node               88 kernel/rcu/tree.c 	.level = { &rcu_state.node[0] },
node              527 kernel/rcu/tree.c 	return &rcu_state.node[0];
node              291 kernel/rcu/tree.h 	struct rcu_node node[NUM_RCU_NODES];	/* Hierarchy. */
node             1629 kernel/resource.c 		INIT_LIST_HEAD(&entry->node);
node             1641 kernel/resource.c 	list_for_each_entry_safe(entry, tmp, head, node)
node              415 kernel/sched/core.c 	struct wake_q_node *node = &task->wake_q;
node              426 kernel/sched/core.c 	if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL)))
node              432 kernel/sched/core.c 	*head->lastp = node;
node              433 kernel/sched/core.c 	head->lastp = &node->next;
node              480 kernel/sched/core.c 	struct wake_q_node *node = head->first;
node              482 kernel/sched/core.c 	while (node != WAKE_Q_TAIL) {
node              485 kernel/sched/core.c 		task = container_of(node, struct task_struct, wake_q);
node              488 kernel/sched/core.c 		node = node->next;
node              821 kernel/sched/debug.c void print_numa_stats(struct seq_file *m, int node, unsigned long tsf,
node              824 kernel/sched/debug.c 	SEQ_printf(m, "numa_faults node=%d ", node);
node             1265 kernel/sched/fair.c 	int node;
node             1267 kernel/sched/fair.c 	for_each_online_node(node) {
node             1268 kernel/sched/fair.c 		faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)];
node             1277 kernel/sched/fair.c 	int node;
node             1279 kernel/sched/fair.c 	for_each_online_node(node) {
node             1280 kernel/sched/fair.c 		faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)];
node             1303 kernel/sched/fair.c 	int node;
node             1316 kernel/sched/fair.c 	for_each_online_node(node) {
node             1318 kernel/sched/fair.c 		int dist = node_distance(nid, node);
node             1324 kernel/sched/fair.c 		if (dist == sched_max_numa_distance || node == nid)
node             1340 kernel/sched/fair.c 			faults = task_faults(p, node);
node             1342 kernel/sched/fair.c 			faults = group_faults(p, node);
node             2056 kernel/sched/fair.c 		int node, max_node = nid;
node             2060 kernel/sched/fair.c 		for_each_online_node(node) {
node             2061 kernel/sched/fair.c 			score = group_weight(p, node, dist);
node             2064 kernel/sched/fair.c 				max_node = node;
node             10524 kernel/sched/fair.c 	int node;
node             10530 kernel/sched/fair.c 	for_each_online_node(node) {
node             10532 kernel/sched/fair.c 			tsf = p->numa_faults[task_faults_idx(NUMA_MEM, node, 0)];
node             10533 kernel/sched/fair.c 			tpf = p->numa_faults[task_faults_idx(NUMA_MEM, node, 1)];
node             10536 kernel/sched/fair.c 			gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)],
node             10537 kernel/sched/fair.c 			gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)];
node             10539 kernel/sched/fair.c 		print_numa_stats(m, node, tsf, tpf, gsf, gpf);
node              495 kernel/sched/psi.c 	list_for_each_entry(t, &group->triggers, node)
node              513 kernel/sched/psi.c 	list_for_each_entry(t, &group->triggers, node) {
node             1071 kernel/sched/psi.c 	list_add(&t->node, &group->triggers);
node             1099 kernel/sched/psi.c 	if (!list_empty(&t->node)) {
node             1103 kernel/sched/psi.c 		list_del(&t->node);
node             1108 kernel/sched/psi.c 		list_for_each_entry(tmp, &group->triggers, node)
node             1300 kernel/sched/sched.h extern void sched_setnuma(struct task_struct *p, int node);
node             2194 kernel/sched/sched.h print_numa_stats(struct seq_file *m, int node, unsigned long tsf,
node             1707 kernel/sched/topology.c 	int node = cpu_to_node(cpu);
node             1712 kernel/sched/topology.c 			if (node_distance(j, node) <= sched_domains_numa_distance[i])
node             1169 kernel/signal.c 		hlist_for_each_entry(delayed, &t->signal->multiprocess, node) {
node             1144 kernel/sysctl_binary.c 		unsigned long area, node;
node             1162 kernel/sysctl_binary.c 		node = simple_strtoul(nodep, NULL, 10);
node             1165 kernel/sysctl_binary.c 		if ((area > 63)||(node > 1023))
node             1168 kernel/sysctl_binary.c 		dnaddr = cpu_to_le16((area << 10) | node);
node              172 kernel/time/alarmtimer.c 		timerqueue_del(&base->timerqueue, &alarm->node);
node              174 kernel/time/alarmtimer.c 	timerqueue_add(&base->timerqueue, &alarm->node);
node              192 kernel/time/alarmtimer.c 	timerqueue_del(&base->timerqueue, &alarm->node);
node              223 kernel/time/alarmtimer.c 		hrtimer_set_expires(&alarm->timer, alarm->node.expires);
node              237 kernel/time/alarmtimer.c 	return ktime_sub(alarm->node.expires, base->gettime());
node              338 kernel/time/alarmtimer.c 	timerqueue_init(&alarm->node);
node              371 kernel/time/alarmtimer.c 	alarm->node.expires = start;
node              373 kernel/time/alarmtimer.c 	hrtimer_start(&alarm->timer, alarm->node.expires, HRTIMER_MODE_ABS);
node              400 kernel/time/alarmtimer.c 	hrtimer_set_expires(&alarm->timer, alarm->node.expires);
node              455 kernel/time/alarmtimer.c 	delta = ktime_sub(now, alarm->node.expires);
node              465 kernel/time/alarmtimer.c 		alarm->node.expires = ktime_add_ns(alarm->node.expires,
node              468 kernel/time/alarmtimer.c 		if (alarm->node.expires > now)
node              477 kernel/time/alarmtimer.c 	alarm->node.expires = ktime_add_safe(alarm->node.expires, interval);
node              581 kernel/time/alarmtimer.c 	alarm_start(alarm, alarm->node.expires);
node              605 kernel/time/alarmtimer.c 	return ktime_sub(alarm->node.expires, now);
node              646 kernel/time/alarmtimer.c 		alarm->node.expires = expires;
node              516 kernel/time/hrtimer.c 		timer = container_of(next, struct hrtimer, node);
node              523 kernel/time/hrtimer.c 			timer = container_of(next, struct hrtimer, node);
node              972 kernel/time/hrtimer.c 	return timerqueue_add(&base->active, &timer->node);
node              997 kernel/time/hrtimer.c 	if (!timerqueue_del(&base->active, &timer->node))
node             1409 kernel/time/hrtimer.c 	timerqueue_init(&timer->node);
node             1554 kernel/time/hrtimer.c 		struct timerqueue_node *node;
node             1559 kernel/time/hrtimer.c 		while ((node = timerqueue_getnext(&base->active))) {
node             1562 kernel/time/hrtimer.c 			timer = container_of(node, struct hrtimer, node);
node             2016 kernel/time/hrtimer.c 	struct timerqueue_node *node;
node             2018 kernel/time/hrtimer.c 	while ((node = timerqueue_getnext(&old_base->active))) {
node             2019 kernel/time/hrtimer.c 		timer = container_of(node, struct hrtimer, node);
node              127 kernel/time/posix-cpu-timers.c 	u64 delta, incr, expires = timer->it.cpu.node.expires;
node              147 kernel/time/posix-cpu-timers.c 		timer->it.cpu.node.expires += incr;
node              151 kernel/time/posix-cpu-timers.c 	return timer->it.cpu.node.expires;
node              395 kernel/time/posix-cpu-timers.c 	timerqueue_init(&new_timer->it.cpu.node);
node              427 kernel/time/posix-cpu-timers.c 		WARN_ON_ONCE(ctmr->head || timerqueue_node_queued(&ctmr->node));
node              445 kernel/time/posix-cpu-timers.c 	struct timerqueue_node *node;
node              448 kernel/time/posix-cpu-timers.c 	while ((node = timerqueue_getnext(head))) {
node              449 kernel/time/posix-cpu-timers.c 		timerqueue_del(head, node);
node              450 kernel/time/posix-cpu-timers.c 		ctmr = container_of(node, struct cpu_timer, node);
node              773 kernel/time/posix-cpu-timers.c 		ctmr = container_of(next, struct cpu_timer, node);
node              100 kernel/time/timer_list.c 		timer = container_of(curr, struct hrtimer, node);
node              390 kernel/trace/ftrace.c 	struct hlist_node		node;
node              706 kernel/trace/ftrace.c 	hlist_for_each_entry_rcu_notrace(rec, hhd, node) {
node              720 kernel/trace/ftrace.c 	hlist_add_head_rcu(&rec->node, &stat->hash[key]);
node              498 kernel/trace/ring_buffer.c 	struct hlist_node		node;
node             1421 kernel/trace/ring_buffer.c 	ret = cpuhp_state_add_instance(CPUHP_TRACE_RB_PREPARE, &buffer->node);
node             1454 kernel/trace/ring_buffer.c 	cpuhp_state_remove_instance(CPUHP_TRACE_RB_PREPARE, &buffer->node);
node             4869 kernel/trace/ring_buffer.c int trace_rb_cpu_prepare(unsigned int cpu, struct hlist_node *node)
node             4876 kernel/trace/ring_buffer.c 	buffer = container_of(node, struct ring_buffer, node);
node             1188 kernel/trace/trace_events.c 	struct list_head *node = v;
node             1194 kernel/trace/trace_events.c 		node = common_head;
node             1198 kernel/trace/trace_events.c 		node = head;
node             1206 kernel/trace/trace_events.c 	node = node->prev;
node             1207 kernel/trace/trace_events.c 	if (node == common_head)
node             1209 kernel/trace/trace_events.c 	else if (node == head)
node             1212 kernel/trace/trace_events.c 		return node;
node              653 kernel/trace/trace_output.c 	hlist_for_each_entry(event, &event_hash[key], node) {
node              772 kernel/trace/trace_output.c 	hlist_add_head(&event->node, &event_hash[key]);
node              787 kernel/trace/trace_output.c 	hlist_del(&event->node);
node               27 kernel/trace/trace_stat.c 	struct rb_node		node;
node               51 kernel/trace/trace_stat.c 	rbtree_postorder_for_each_entry_safe(snode, n, &session->stat_root, node) {
node               95 kernel/trace/trace_stat.c 		this = container_of(*new, struct stat_node, node);
node              105 kernel/trace/trace_stat.c 	rb_link_node(&data->node, parent, new);
node              106 kernel/trace/trace_stat.c 	rb_insert_color(&data->node, root);
node              176 kernel/trace/trace_stat.c 	struct rb_node *node;
node              190 kernel/trace/trace_stat.c 	node = rb_first(&session->stat_root);
node              191 kernel/trace/trace_stat.c 	for (i = 0; node && i < n; i++)
node              192 kernel/trace/trace_stat.c 		node = rb_next(node);
node              194 kernel/trace/trace_stat.c 	return node;
node              200 kernel/trace/trace_stat.c 	struct rb_node *node = p;
node              207 kernel/trace/trace_stat.c 	return rb_next(node);
node              219 kernel/trace/trace_stat.c 	struct stat_node *l = container_of(v, struct stat_node, node);
node              312 kernel/trace/trace_stat.c 	struct stat_session *session, *node;
node              323 kernel/trace/trace_stat.c 	list_for_each_entry(node, &all_stat_sessions, session_list) {
node              324 kernel/trace/trace_stat.c 		if (node->ts == trace)
node              355 kernel/trace/trace_stat.c 	struct stat_session *node, *tmp;
node              358 kernel/trace/trace_stat.c 	list_for_each_entry_safe(node, tmp, &all_stat_sessions, session_list) {
node              359 kernel/trace/trace_stat.c 		if (node->ts == trace) {
node              360 kernel/trace/trace_stat.c 			list_del(&node->session_list);
node              361 kernel/trace/trace_stat.c 			destroy_session(node);
node              118 kernel/ucount.c 	hlist_for_each_entry(ucounts, hashent, node) {
node              148 kernel/ucount.c 			hlist_add_head(&new->node, hashent);
node              167 kernel/ucount.c 		hlist_del_init(&ucounts->node);
node              150 kernel/workqueue.c 	int			node;		/* I: the associated node ID */
node              411 kernel/workqueue.c 	list_for_each_entry((worker), &(pool)->workers, node)		\
node              565 kernel/workqueue.c 						  int node)
node              575 kernel/workqueue.c 	if (unlikely(node == NUMA_NO_NODE))
node              578 kernel/workqueue.c 	return rcu_dereference_raw(wq->numa_pwq_tbl[node]);
node             1539 kernel/workqueue.c static int workqueue_select_cpu_near(int node)
node             1548 kernel/workqueue.c 	if (node < 0 || node >= MAX_NUMNODES || !node_online(node))
node             1553 kernel/workqueue.c 	if (node == cpu_to_node(cpu))
node             1557 kernel/workqueue.c 	cpu = cpumask_any_and(cpumask_of_node(node), cpu_online_mask);
node             1583 kernel/workqueue.c bool queue_work_node(int node, struct workqueue_struct *wq,
node             1603 kernel/workqueue.c 		int cpu = workqueue_select_cpu_near(node);
node             1819 kernel/workqueue.c static struct worker *alloc_worker(int node)
node             1823 kernel/workqueue.c 	worker = kzalloc_node(sizeof(*worker), GFP_KERNEL, node);
node             1827 kernel/workqueue.c 		INIT_LIST_HEAD(&worker->node);
node             1862 kernel/workqueue.c 	list_add_tail(&worker->node, &pool->workers);
node             1883 kernel/workqueue.c 	list_del(&worker->node);
node             1920 kernel/workqueue.c 	worker = alloc_worker(pool->node);
node             1932 kernel/workqueue.c 	worker->task = kthread_create_on_node(worker_thread, worker, pool->node,
node             3424 kernel/workqueue.c 	pool->node = NUMA_NO_NODE;
node             3591 kernel/workqueue.c 	int node;
node             3606 kernel/workqueue.c 		for_each_node(node) {
node             3608 kernel/workqueue.c 					   wq_numa_possible_cpumask[node])) {
node             3609 kernel/workqueue.c 				target_node = node;
node             3622 kernel/workqueue.c 	pool->node = target_node;
node             3789 kernel/workqueue.c 	pwq = kmem_cache_alloc_node(pwq_cache, GFP_KERNEL, pool->node);
node             3821 kernel/workqueue.c static bool wq_calc_node_cpumask(const struct workqueue_attrs *attrs, int node,
node             3828 kernel/workqueue.c 	cpumask_and(cpumask, cpumask_of_node(node), attrs->cpumask);
node             3836 kernel/workqueue.c 	cpumask_and(cpumask, attrs->cpumask, wq_numa_possible_cpumask[node]);
node             3853 kernel/workqueue.c 						   int node,
node             3864 kernel/workqueue.c 	old_pwq = rcu_access_pointer(wq->numa_pwq_tbl[node]);
node             3865 kernel/workqueue.c 	rcu_assign_pointer(wq->numa_pwq_tbl[node], pwq);
node             3882 kernel/workqueue.c 		int node;
node             3884 kernel/workqueue.c 		for_each_node(node)
node             3885 kernel/workqueue.c 			put_pwq_unlocked(ctx->pwq_tbl[node]);
node             3901 kernel/workqueue.c 	int node;
node             3938 kernel/workqueue.c 	for_each_node(node) {
node             3939 kernel/workqueue.c 		if (wq_calc_node_cpumask(new_attrs, node, -1, tmp_attrs->cpumask)) {
node             3940 kernel/workqueue.c 			ctx->pwq_tbl[node] = alloc_unbound_pwq(wq, tmp_attrs);
node             3941 kernel/workqueue.c 			if (!ctx->pwq_tbl[node])
node             3945 kernel/workqueue.c 			ctx->pwq_tbl[node] = ctx->dfl_pwq;
node             3968 kernel/workqueue.c 	int node;
node             3976 kernel/workqueue.c 	for_each_node(node)
node             3977 kernel/workqueue.c 		ctx->pwq_tbl[node] = numa_pwq_tbl_install(ctx->wq, node,
node             3978 kernel/workqueue.c 							  ctx->pwq_tbl[node]);
node             4085 kernel/workqueue.c 	int node = cpu_to_node(cpu);
node             4106 kernel/workqueue.c 	pwq = unbound_pwq_by_node(wq, node);
node             4114 kernel/workqueue.c 	if (wq_calc_node_cpumask(wq->dfl_pwq->pool->attrs, node, cpu_off, cpumask)) {
node             4131 kernel/workqueue.c 	old_pwq = numa_pwq_tbl_install(wq, node, pwq);
node             4139 kernel/workqueue.c 	old_pwq = numa_pwq_tbl_install(wq, node, wq->dfl_pwq);
node             4335 kernel/workqueue.c 	int node;
node             4404 kernel/workqueue.c 		for_each_node(node) {
node             4405 kernel/workqueue.c 			pwq = rcu_access_pointer(wq->numa_pwq_tbl[node]);
node             4406 kernel/workqueue.c 			RCU_INIT_POINTER(wq->numa_pwq_tbl[node], NULL);
node             4635 kernel/workqueue.c 	if (pool->node != NUMA_NO_NODE)
node             4636 kernel/workqueue.c 		pr_cont(" node=%d", pool->node);
node             5393 kernel/workqueue.c 	int node, written = 0;
node             5397 kernel/workqueue.c 	for_each_node(node) {
node             5399 kernel/workqueue.c 				     "%s%d:%d", delim, node,
node             5400 kernel/workqueue.c 				     unbound_pwq_by_node(wq, node)->pool->id);
node             5842 kernel/workqueue.c 	int node, cpu;
node             5863 kernel/workqueue.c 	for_each_node(node)
node             5864 kernel/workqueue.c 		BUG_ON(!zalloc_cpumask_var_node(&tbl[node], GFP_KERNEL,
node             5865 kernel/workqueue.c 				node_online(node) ? node : NUMA_NO_NODE));
node             5868 kernel/workqueue.c 		node = cpu_to_node(cpu);
node             5869 kernel/workqueue.c 		if (WARN_ON(node == NUMA_NO_NODE)) {
node             5874 kernel/workqueue.c 		cpumask_set_cpu(cpu, tbl[node]);
node             5914 kernel/workqueue.c 			pool->node = cpu_to_node(cpu);
node             5992 kernel/workqueue.c 			pool->node = cpu_to_node(cpu);
node               41 kernel/workqueue_internal.h 	struct list_head	node;		/* A: anchored at pool->workers */
node               68 lib/842/842_compress.c 	struct hlist_node node;
node               74 lib/842/842_compress.c 	struct hlist_node node;
node               80 lib/842/842_compress.c 	struct hlist_node node;
node              115 lib/842/842_compress.c 	for (_i = 0; _i < ARRAY_SIZE((p)->node##b); _i++) {	\
node              116 lib/842/842_compress.c 		(p)->node##b[_i].index = _i;			\
node              117 lib/842/842_compress.c 		(p)->node##b[_i].data = 0;			\
node              118 lib/842/842_compress.c 		INIT_HLIST_NODE(&(p)->node##b[_i].node);	\
node              125 lib/842/842_compress.c 	hash_for_each_possible(p->htable##b, _n, node, p->data##b[n]) {	\
node              140 lib/842/842_compress.c 	struct sw842_hlist_node##b *_n = &(p)->node##b[(i)+(d)];	\
node              141 lib/842/842_compress.c 	hash_del(&_n->node);						\
node              147 lib/842/842_compress.c 	hash_add((p)->htable##b, &_n->node, _n->data);			\
node               26 lib/assoc_array.c 	const struct assoc_array_node *node;
node               40 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node               52 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
node               79 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node               81 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
node               90 lib/assoc_array.c 	parent = READ_ONCE(node->back_pointer); /* Address dependency. */
node               91 lib/assoc_array.c 	slot = node->parent_slot;
node              153 lib/assoc_array.c 		struct assoc_array_node	*node;	/* Node in which leaf might be found */
node              176 lib/assoc_array.c 	struct assoc_array_node *node;
node              206 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node              209 lib/assoc_array.c 	ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
node              218 lib/assoc_array.c 		result->terminal_node.node = node;
node              309 lib/assoc_array.c 	const struct assoc_array_node *node;
node              318 lib/assoc_array.c 	node = result.terminal_node.node;
node              324 lib/assoc_array.c 		ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */
node              347 lib/assoc_array.c 	struct assoc_array_node *node;
node              374 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node              375 lib/assoc_array.c 	BUG_ON(node->back_pointer != parent);
node              376 lib/assoc_array.c 	BUG_ON(slot != -1 && node->parent_slot != slot);
node              380 lib/assoc_array.c 	pr_devel("Node %p [back=%p]\n", node, node->back_pointer);
node              382 lib/assoc_array.c 		struct assoc_array_ptr *ptr = node->slots[slot];
node              397 lib/assoc_array.c 	parent = node->back_pointer;
node              398 lib/assoc_array.c 	slot = node->parent_slot;
node              400 lib/assoc_array.c 	kfree(node);
node              423 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node              480 lib/assoc_array.c 	struct assoc_array_node *node, *new_n0, *new_n1, *side;
node              488 lib/assoc_array.c 	node	= result->terminal_node.node;
node              505 lib/assoc_array.c 		ptr = node->slots[i];
node              514 lib/assoc_array.c 			edit->leaf_p = &node->slots[i];
node              515 lib/assoc_array.c 			edit->dead_leaf = node->slots[i];
node              526 lib/assoc_array.c 		edit->leaf_p = &node->slots[free_slot];
node              527 lib/assoc_array.c 		edit->adjust_count_on = node;
node              552 lib/assoc_array.c 		ptr = node->slots[i];
node              614 lib/assoc_array.c 	new_n0->back_pointer = node->back_pointer;
node              615 lib/assoc_array.c 	new_n0->parent_slot = node->parent_slot;
node              622 lib/assoc_array.c 	new_n0->nr_leaves_on_branch = node->nr_leaves_on_branch;
node              648 lib/assoc_array.c 		if (assoc_array_ptr_is_meta(node->slots[i]))
node              649 lib/assoc_array.c 			new_n0->slots[i] = node->slots[i];
node              659 lib/assoc_array.c 		if (assoc_array_ptr_is_meta(node->slots[i]))
node              662 lib/assoc_array.c 			new_n1->slots[next_slot++] = node->slots[i];
node              668 lib/assoc_array.c 			new_n0->slots[free_slot] = node->slots[i];
node              690 lib/assoc_array.c 			ptr = node->slots[i];
node              702 lib/assoc_array.c 	ptr = node->back_pointer;
node              706 lib/assoc_array.c 		edit->set[0].ptr = &assoc_array_ptr_to_node(ptr)->slots[node->parent_slot];
node              709 lib/assoc_array.c 	edit->excised_meta[0] = assoc_array_node_to_ptr(node);
node              731 lib/assoc_array.c 		int x = ops->diff_objects(assoc_array_ptr_to_leaf(node->slots[i]),
node              751 lib/assoc_array.c 	new_s0->back_pointer = node->back_pointer;
node              752 lib/assoc_array.c 	new_s0->parent_slot = node->parent_slot;
node              777 lib/assoc_array.c 		ptr = node->slots[i];
node              798 lib/assoc_array.c 	struct assoc_array_node *node, *new_n0, *side;
node              827 lib/assoc_array.c 		node = assoc_array_ptr_to_node(shortcut->back_pointer);
node              828 lib/assoc_array.c 		edit->set[0].ptr = &node->slots[shortcut->parent_slot];
node             1038 lib/assoc_array.c 	struct assoc_array_node	*node;
node             1056 lib/assoc_array.c 	collapse->node->slots[collapse->slot++] = assoc_array_leaf_to_ptr(leaf);
node             1085 lib/assoc_array.c 	struct assoc_array_node *node, *new_n0;
node             1106 lib/assoc_array.c 		node = result.terminal_node.node;
node             1109 lib/assoc_array.c 			ptr = node->slots[slot];
node             1131 lib/assoc_array.c 	edit->dead_leaf = node->slots[slot];
node             1132 lib/assoc_array.c 	edit->set[0].ptr = &node->slots[slot];
node             1134 lib/assoc_array.c 	edit->adjust_count_on = node;
node             1158 lib/assoc_array.c 	if (node->nr_leaves_on_branch <= ASSOC_ARRAY_FAN_OUT + 1) {
node             1168 lib/assoc_array.c 			ptr = node->slots[i];
node             1176 lib/assoc_array.c 			 node->nr_leaves_on_branch - 1, has_meta);
node             1181 lib/assoc_array.c 		parent = node;
node             1206 lib/assoc_array.c 		if (has_meta || parent != node) {
node             1207 lib/assoc_array.c 			node = parent;
node             1215 lib/assoc_array.c 			new_n0->back_pointer = node->back_pointer;
node             1216 lib/assoc_array.c 			new_n0->parent_slot = node->parent_slot;
node             1217 lib/assoc_array.c 			new_n0->nr_leaves_on_branch = node->nr_leaves_on_branch;
node             1220 lib/assoc_array.c 			collapse.node = new_n0;
node             1223 lib/assoc_array.c 			assoc_array_subtree_iterate(assoc_array_node_to_ptr(node),
node             1224 lib/assoc_array.c 						    node->back_pointer,
node             1230 lib/assoc_array.c 			if (!node->back_pointer) {
node             1232 lib/assoc_array.c 			} else if (assoc_array_ptr_is_leaf(node->back_pointer)) {
node             1234 lib/assoc_array.c 			} else if (assoc_array_ptr_is_node(node->back_pointer)) {
node             1236 lib/assoc_array.c 					assoc_array_ptr_to_node(node->back_pointer);
node             1237 lib/assoc_array.c 				edit->set[1].ptr = &p->slots[node->parent_slot];
node             1238 lib/assoc_array.c 			} else if (assoc_array_ptr_is_shortcut(node->back_pointer)) {
node             1240 lib/assoc_array.c 					assoc_array_ptr_to_shortcut(node->back_pointer);
node             1244 lib/assoc_array.c 			edit->excised_subtree = assoc_array_node_to_ptr(node);
node             1349 lib/assoc_array.c 	struct assoc_array_node *node;
node             1377 lib/assoc_array.c 		node = edit->adjust_count_on;
node             1379 lib/assoc_array.c 			node->nr_leaves_on_branch += edit->adjust_count_by;
node             1381 lib/assoc_array.c 			ptr = node->back_pointer;
node             1391 lib/assoc_array.c 			node = assoc_array_ptr_to_node(ptr);
node             1460 lib/assoc_array.c 	struct assoc_array_node *node, *new_n;
node             1508 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node             1512 lib/assoc_array.c 	pr_devel("dup node %p -> %p\n", node, new_n);
node             1514 lib/assoc_array.c 	new_n->parent_slot = node->parent_slot;
node             1522 lib/assoc_array.c 		ptr = node->slots[slot];
node             1693 lib/assoc_array.c 	ptr = node->back_pointer;
node             1701 lib/assoc_array.c 		slot = node->parent_slot;
node             1705 lib/assoc_array.c 	node = assoc_array_ptr_to_node(cursor);
node               95 lib/btree.c    	unsigned long *node;
node               97 lib/btree.c    	node = mempool_alloc(head->mempool, gfp);
node               98 lib/btree.c    	if (likely(node))
node               99 lib/btree.c    		memset(node, 0, NODESIZE);
node              100 lib/btree.c    	return node;
node              148 lib/btree.c    static unsigned long *bkey(struct btree_geo *geo, unsigned long *node, int n)
node              150 lib/btree.c    	return &node[n * geo->keylen];
node              153 lib/btree.c    static void *bval(struct btree_geo *geo, unsigned long *node, int n)
node              155 lib/btree.c    	return (void *)node[geo->no_longs + n];
node              158 lib/btree.c    static void setkey(struct btree_geo *geo, unsigned long *node, int n,
node              161 lib/btree.c    	longcpy(bkey(geo, node, n), key, geo->keylen);
node              164 lib/btree.c    static void setval(struct btree_geo *geo, unsigned long *node, int n,
node              167 lib/btree.c    	node[geo->no_longs + n] = (unsigned long) val;
node              170 lib/btree.c    static void clearpair(struct btree_geo *geo, unsigned long *node, int n)
node              172 lib/btree.c    	longset(bkey(geo, node, n), 0, geo->keylen);
node              173 lib/btree.c    	node[geo->no_longs + n] = 0;
node              178 lib/btree.c    	head->node = NULL;
node              201 lib/btree.c    	mempool_free(head->node, head->mempool);
node              211 lib/btree.c    	unsigned long *node = head->node;
node              217 lib/btree.c    		node = bval(geo, node, 0);
node              219 lib/btree.c    	longcpy(key, bkey(geo, node, 0), geo->keylen);
node              220 lib/btree.c    	return bval(geo, node, 0);
node              224 lib/btree.c    static int keycmp(struct btree_geo *geo, unsigned long *node, int pos,
node              227 lib/btree.c    	return longcmp(bkey(geo, node, pos), key, geo->keylen);
node              245 lib/btree.c    	unsigned long *node = head->node;
node              252 lib/btree.c    			if (keycmp(geo, node, i, key) <= 0)
node              256 lib/btree.c    		node = bval(geo, node, i);
node              257 lib/btree.c    		if (!node)
node              261 lib/btree.c    	if (!node)
node              265 lib/btree.c    		if (keycmp(geo, node, i, key) == 0)
node              266 lib/btree.c    			return bval(geo, node, i);
node              275 lib/btree.c    	unsigned long *node = head->node;
node              282 lib/btree.c    			if (keycmp(geo, node, i, key) <= 0)
node              286 lib/btree.c    		node = bval(geo, node, i);
node              287 lib/btree.c    		if (!node)
node              291 lib/btree.c    	if (!node)
node              295 lib/btree.c    		if (keycmp(geo, node, i, key) == 0) {
node              296 lib/btree.c    			setval(geo, node, i, val);
node              315 lib/btree.c    	unsigned long *node, *oldnode;
node              327 lib/btree.c    	node = head->node;
node              330 lib/btree.c    			if (keycmp(geo, node, i, key) <= 0)
node              334 lib/btree.c    		oldnode = node;
node              335 lib/btree.c    		node = bval(geo, node, i);
node              336 lib/btree.c    		if (!node)
node              341 lib/btree.c    	if (!node)
node              345 lib/btree.c    		if (keycmp(geo, node, i, key) <= 0) {
node              346 lib/btree.c    			if (bval(geo, node, i)) {
node              347 lib/btree.c    				longcpy(__key, bkey(geo, node, i), geo->keylen);
node              348 lib/btree.c    				return bval(geo, node, i);
node              363 lib/btree.c    static int getpos(struct btree_geo *geo, unsigned long *node,
node              369 lib/btree.c    		if (keycmp(geo, node, i, key) <= 0)
node              375 lib/btree.c    static int getfill(struct btree_geo *geo, unsigned long *node, int start)
node              380 lib/btree.c    		if (!bval(geo, node, i))
node              391 lib/btree.c    	unsigned long *node = head->node;
node              396 lib/btree.c    			if (keycmp(geo, node, i, key) <= 0)
node              399 lib/btree.c    		if ((i == geo->no_pairs) || !bval(geo, node, i)) {
node              404 lib/btree.c    			setkey(geo, node, i, key);
node              407 lib/btree.c    		node = bval(geo, node, i);
node              409 lib/btree.c    	BUG_ON(!node);
node              410 lib/btree.c    	return node;
node              416 lib/btree.c    	unsigned long *node;
node              419 lib/btree.c    	node = btree_node_alloc(head, gfp);
node              420 lib/btree.c    	if (!node)
node              422 lib/btree.c    	if (head->node) {
node              423 lib/btree.c    		fill = getfill(geo, head->node, 0);
node              424 lib/btree.c    		setkey(geo, node, 0, bkey(geo, head->node, fill - 1));
node              425 lib/btree.c    		setval(geo, node, 0, head->node);
node              427 lib/btree.c    	head->node = node;
node              434 lib/btree.c    	unsigned long *node;
node              440 lib/btree.c    	node = head->node;
node              441 lib/btree.c    	fill = getfill(geo, node, 0);
node              443 lib/btree.c    	head->node = bval(geo, node, 0);
node              445 lib/btree.c    	mempool_free(node, head->mempool);
node              452 lib/btree.c    	unsigned long *node;
node              463 lib/btree.c    	node = find_level(head, geo, key, level);
node              464 lib/btree.c    	pos = getpos(geo, node, key);
node              465 lib/btree.c    	fill = getfill(geo, node, pos);
node              467 lib/btree.c    	BUG_ON(pos < fill && keycmp(geo, node, pos, key) == 0);
node              477 lib/btree.c    				bkey(geo, node, fill / 2 - 1),
node              484 lib/btree.c    			setkey(geo, new, i, bkey(geo, node, i));
node              485 lib/btree.c    			setval(geo, new, i, bval(geo, node, i));
node              486 lib/btree.c    			setkey(geo, node, i, bkey(geo, node, i + fill / 2));
node              487 lib/btree.c    			setval(geo, node, i, bval(geo, node, i + fill / 2));
node              488 lib/btree.c    			clearpair(geo, node, i + fill / 2);
node              491 lib/btree.c    			setkey(geo, node, i, bkey(geo, node, fill - 1));
node              492 lib/btree.c    			setval(geo, node, i, bval(geo, node, fill - 1));
node              493 lib/btree.c    			clearpair(geo, node, fill - 1);
node              501 lib/btree.c    		setkey(geo, node, i, bkey(geo, node, i - 1));
node              502 lib/btree.c    		setval(geo, node, i, bval(geo, node, i - 1));
node              504 lib/btree.c    	setkey(geo, node, pos, key);
node              505 lib/btree.c    	setval(geo, node, pos, val);
node              594 lib/btree.c    	unsigned long *node;
node              601 lib/btree.c    		head->node = NULL;
node              605 lib/btree.c    	node = find_level(head, geo, key, level);
node              606 lib/btree.c    	pos = getpos(geo, node, key);
node              607 lib/btree.c    	fill = getfill(geo, node, pos);
node              608 lib/btree.c    	if ((level == 1) && (keycmp(geo, node, pos, key) != 0))
node              610 lib/btree.c    	ret = bval(geo, node, pos);
node              614 lib/btree.c    		setkey(geo, node, i, bkey(geo, node, i + 1));
node              615 lib/btree.c    		setval(geo, node, i, bval(geo, node, i + 1));
node              617 lib/btree.c    	clearpair(geo, node, fill - 1);
node              621 lib/btree.c    			rebalance(head, geo, key, level, node, fill - 1);
node              649 lib/btree.c    	if (!(target->node)) {
node              651 lib/btree.c    		target->node = victim->node;
node              677 lib/btree.c    			       unsigned long *node, unsigned long opaque,
node              687 lib/btree.c    		child = bval(geo, node, i);
node              694 lib/btree.c    			func(child, opaque, bkey(geo, node, i), count++,
node              698 lib/btree.c    		mempool_free(node, head->mempool);
node              757 lib/btree.c    	if (head->node)
node              758 lib/btree.c    		count = __btree_for_each(head, geo, head->node, opaque, func,
node              775 lib/btree.c    	if (head->node)
node              776 lib/btree.c    		count = __btree_for_each(head, geo, head->node, opaque, func,
node              113 lib/cpumask.c  bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node)
node              115 lib/cpumask.c  	*mask = kmalloc_node(cpumask_size(), flags, node);
node              128 lib/cpumask.c  bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node)
node              130 lib/cpumask.c  	return alloc_cpumask_var_node(mask, flags | __GFP_ZERO, node);
node              206 lib/cpumask.c  unsigned int cpumask_local_spread(unsigned int i, int node)
node              213 lib/cpumask.c  	if (node == NUMA_NO_NODE) {
node              219 lib/cpumask.c  		for_each_cpu_and(cpu, cpumask_of_node(node), cpu_online_mask)
node              225 lib/cpumask.c  			if (cpumask_test_cpu(cpu, cpumask_of_node(node)))
node              153 lib/debugobjects.c 			obj = hlist_entry(obj_to_free.first, typeof(*obj), node);
node              154 lib/debugobjects.c 			hlist_del(&obj->node);
node              156 lib/debugobjects.c 			hlist_add_head(&obj->node, &obj_pool);
node              179 lib/debugobjects.c 			hlist_add_head(&new[--cnt]->node, &obj_pool);
node              195 lib/debugobjects.c 	hlist_for_each_entry(obj, &b->list, node) {
node              214 lib/debugobjects.c 		obj = hlist_entry(list->first, typeof(*obj), node);
node              215 lib/debugobjects.c 		hlist_del(&obj->node);
node              258 lib/debugobjects.c 				hlist_add_head(&obj2->node,
node              280 lib/debugobjects.c 		hlist_add_head(&obj->node, &b->list);
node              313 lib/debugobjects.c 		obj = hlist_entry(obj_to_free.first, typeof(*obj), node);
node              314 lib/debugobjects.c 		hlist_del(&obj->node);
node              315 lib/debugobjects.c 		hlist_add_head(&obj->node, &obj_pool);
node              335 lib/debugobjects.c 	hlist_for_each_entry_safe(obj, tmp, &tofree, node) {
node              336 lib/debugobjects.c 		hlist_del(&obj->node);
node              358 lib/debugobjects.c 		hlist_add_head(&obj->node, &percpu_pool->free_objs);
node              383 lib/debugobjects.c 		hlist_add_head(&obj->node, &obj_to_free);
node              388 lib/debugobjects.c 				hlist_add_head(&objs[--lookahead_count]->node,
node              402 lib/debugobjects.c 				hlist_add_head(&obj->node, &obj_to_free);
node              409 lib/debugobjects.c 		hlist_add_head(&obj->node, &obj_pool);
node              414 lib/debugobjects.c 				hlist_add_head(&objs[--lookahead_count]->node,
node              457 lib/debugobjects.c 		hlist_for_each_entry_safe(obj, tmp, &freelist, node) {
node              458 lib/debugobjects.c 			hlist_del(&obj->node);
node              826 lib/debugobjects.c 		hlist_del(&obj->node);
node              956 lib/debugobjects.c 		hlist_for_each_entry_safe(obj, tmp, &db->list, node) {
node              972 lib/debugobjects.c 				hlist_del(&obj->node);
node             1289 lib/debugobjects.c 		hlist_add_head(&obj_static_pool[i].node, &obj_pool);
node             1307 lib/debugobjects.c 		hlist_add_head(&obj->node, &objects);
node             1317 lib/debugobjects.c 	hlist_for_each_entry_safe(obj, tmp, &obj_pool, node)
node             1318 lib/debugobjects.c 		hlist_del(&obj->node);
node             1326 lib/debugobjects.c 		hlist_for_each_entry(obj, &objects, node) {
node             1327 lib/debugobjects.c 			new = hlist_entry(obj_pool.first, typeof(*obj), node);
node             1328 lib/debugobjects.c 			hlist_del(&new->node);
node             1331 lib/debugobjects.c 			hlist_add_head(&new->node, &db->list);
node             1340 lib/debugobjects.c 	hlist_for_each_entry_safe(obj, tmp, &objects, node) {
node             1341 lib/debugobjects.c 		hlist_del(&obj->node);
node              206 lib/devres.c   void __iomem *devm_of_iomap(struct device *dev, struct device_node *node, int index,
node              211 lib/devres.c   	if (of_address_to_resource(node, index, &res))
node               81 lib/generic-radix-tree.c 	struct genradix_node *node;
node               83 lib/generic-radix-tree.c 	node = (struct genradix_node *)__get_free_page(gfp_mask|__GFP_ZERO);
node               90 lib/generic-radix-tree.c 	kmemleak_alloc(node, PAGE_SIZE, 1, gfp_mask);
node               91 lib/generic-radix-tree.c 	return node;
node               94 lib/generic-radix-tree.c static inline void genradix_free_node(struct genradix_node *node)
node               96 lib/generic-radix-tree.c 	kmemleak_free(node);
node               97 lib/generic-radix-tree.c 	free_page((unsigned long)node);
node              292 lib/idr.c      	struct radix_tree_node *node;
node              298 lib/idr.c      	entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot);
node              302 lib/idr.c      	__radix_tree_replace(&idr->idr_rt, node, slot, ptr);
node              564 lib/idr.c      		struct xa_node *node = xa_to_node(entry);
node              565 lib/idr.c      		unsigned int shift = node->shift + IDA_CHUNK_SHIFT +
node              569 lib/idr.c      		xa_dump_node(node);
node              571 lib/idr.c      			ida_dump_entry(node->slots[i],
node              572 lib/idr.c      					index | (i << node->shift));
node                7 lib/interval_tree.c #define START(node) ((node)->start)
node                8 lib/interval_tree.c #define LAST(node)  ((node)->last)
node               32 lib/interval_tree_test.c 	struct interval_tree_node *node;
node               35 lib/interval_tree_test.c 	for (node = interval_tree_iter_first(root, start, last); node;
node               36 lib/interval_tree_test.c 	     node = interval_tree_iter_next(node, start, last))
node              175 lib/klist.c    	struct klist_node *node;
node              192 lib/klist.c    		if (waiter->node != n)
node              242 lib/klist.c    	waiter.node = n;
node                6 lib/nodemask.c int __next_node_in(int node, const nodemask_t *srcp)
node                8 lib/nodemask.c 	int ret = __next_node(node, srcp);
node              737 lib/objagg.c   	struct objagg_tmp_node *node = &graph->nodes[index];
node              738 lib/objagg.c   	unsigned int weight = node->objagg_obj->stats.user_count;
node              748 lib/objagg.c   		node = &graph->nodes[j];
node              749 lib/objagg.c   		if (node->crossed_out)
node              751 lib/objagg.c   		weight += node->objagg_obj->stats.user_count;
node              758 lib/objagg.c   	struct objagg_tmp_node *node;
node              765 lib/objagg.c   		node = &graph->nodes[i];
node              766 lib/objagg.c   		if (node->crossed_out)
node              781 lib/objagg.c   	struct objagg_tmp_node *node;
node              804 lib/objagg.c   		node = &graph->nodes[i++];
node              805 lib/objagg.c   		node->objagg_obj = objagg_obj;
node              816 lib/objagg.c   			node = &graph->nodes[j];
node              819 lib/objagg.c   						     node->objagg_obj->obj)) {
node              847 lib/objagg.c   	struct objagg_tmp_node *node;
node              860 lib/objagg.c   		node = &graph->nodes[index];
node              861 lib/objagg.c   		node->crossed_out = true;
node              863 lib/objagg.c   						 node->objagg_obj,
node              874 lib/objagg.c   			node = &graph->nodes[j];
node              875 lib/objagg.c   			if (node->crossed_out)
node              877 lib/objagg.c   			node->crossed_out = true;
node              879 lib/objagg.c   							 node->objagg_obj,
node               73 lib/plist.c    void plist_add(struct plist_node *node, struct plist_head *head)
node               79 lib/plist.c    	WARN_ON(!plist_node_empty(node));
node               80 lib/plist.c    	WARN_ON(!list_empty(&node->prio_list));
node               88 lib/plist.c    		if (node->prio < iter->prio) {
node               98 lib/plist.c    	if (!prev || prev->prio != node->prio)
node               99 lib/plist.c    		list_add_tail(&node->prio_list, &iter->prio_list);
node              101 lib/plist.c    	list_add_tail(&node->node_list, node_next);
node              112 lib/plist.c    void plist_del(struct plist_node *node, struct plist_head *head)
node              116 lib/plist.c    	if (!list_empty(&node->prio_list)) {
node              117 lib/plist.c    		if (node->node_list.next != &head->node_list) {
node              120 lib/plist.c    			next = list_entry(node->node_list.next,
node              125 lib/plist.c    				list_add(&next->prio_list, &node->prio_list);
node              127 lib/plist.c    		list_del_init(&node->prio_list);
node              130 lib/plist.c    	list_del_init(&node->node_list);
node              145 lib/plist.c    void plist_requeue(struct plist_node *node, struct plist_head *head)
node              152 lib/plist.c    	BUG_ON(plist_node_empty(node));
node              154 lib/plist.c    	if (node == plist_last(head))
node              157 lib/plist.c    	iter = plist_next(node);
node              159 lib/plist.c    	if (node->prio != iter->prio)
node              162 lib/plist.c    	plist_del(node, head);
node              165 lib/plist.c    		if (node->prio != iter->prio) {
node              170 lib/plist.c    	list_add_tail(&node->node_list, node_next);
node              212 lib/plist.c    static void __init plist_test_requeue(struct plist_node *node)
node              214 lib/plist.c    	plist_requeue(node, &test_head);
node              216 lib/plist.c    	if (node != plist_last(&test_head))
node              217 lib/plist.c    		BUG_ON(node->prio == plist_next(node)->prio);
node              109 lib/radix-tree.c static inline void tag_set(struct radix_tree_node *node, unsigned int tag,
node              112 lib/radix-tree.c 	__set_bit(offset, node->tags[tag]);
node              115 lib/radix-tree.c static inline void tag_clear(struct radix_tree_node *node, unsigned int tag,
node              118 lib/radix-tree.c 	__clear_bit(offset, node->tags[tag]);
node              121 lib/radix-tree.c static inline int tag_get(const struct radix_tree_node *node, unsigned int tag,
node              124 lib/radix-tree.c 	return test_bit(offset, node->tags[tag]);
node              161 lib/radix-tree.c static inline int any_tag_set(const struct radix_tree_node *node,
node              166 lib/radix-tree.c 		if (node->tags[tag][idx])
node              172 lib/radix-tree.c static inline void all_tag_set(struct radix_tree_node *node, unsigned int tag)
node              174 lib/radix-tree.c 	bitmap_fill(node->tags[tag], RADIX_TREE_MAP_SIZE);
node              189 lib/radix-tree.c radix_tree_find_next_bit(struct radix_tree_node *node, unsigned int tag,
node              192 lib/radix-tree.c 	const unsigned long *addr = node->tags[tag];
node              225 lib/radix-tree.c static inline unsigned long node_maxindex(const struct radix_tree_node *node)
node              227 lib/radix-tree.c 	return shift_maxindex(node->shift);
node              231 lib/radix-tree.c 				const struct radix_tree_node *node,
node              234 lib/radix-tree.c 	return (index & ~node_maxindex(node)) + (offset << node->shift);
node              301 lib/radix-tree.c 	struct radix_tree_node *node =
node              309 lib/radix-tree.c 	memset(node->slots, 0, sizeof(node->slots));
node              310 lib/radix-tree.c 	memset(node->tags, 0, sizeof(node->tags));
node              311 lib/radix-tree.c 	INIT_LIST_HEAD(&node->private_list);
node              313 lib/radix-tree.c 	kmem_cache_free(radix_tree_node_cachep, node);
node              317 lib/radix-tree.c radix_tree_node_free(struct radix_tree_node *node)
node              319 lib/radix-tree.c 	call_rcu(&node->rcu_head, radix_tree_node_rcu_free);
node              334 lib/radix-tree.c 	struct radix_tree_node *node;
node              347 lib/radix-tree.c 		node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
node              348 lib/radix-tree.c 		if (node == NULL)
node              353 lib/radix-tree.c 			node->parent = rtp->nodes;
node              354 lib/radix-tree.c 			rtp->nodes = node;
node              357 lib/radix-tree.c 			kmem_cache_free(radix_tree_node_cachep, node);
node              400 lib/radix-tree.c 	struct radix_tree_node *node = rcu_dereference_raw(root->xa_head);
node              402 lib/radix-tree.c 	*nodep = node;
node              404 lib/radix-tree.c 	if (likely(radix_tree_is_internal_node(node))) {
node              405 lib/radix-tree.c 		node = entry_to_node(node);
node              406 lib/radix-tree.c 		*maxindex = node_maxindex(node);
node              407 lib/radix-tree.c 		return node->shift + RADIX_TREE_MAP_SHIFT;
node              434 lib/radix-tree.c 		struct radix_tree_node *node = radix_tree_node_alloc(gfp, NULL,
node              436 lib/radix-tree.c 		if (!node)
node              440 lib/radix-tree.c 			all_tag_set(node, IDR_FREE);
node              442 lib/radix-tree.c 				tag_clear(node, IDR_FREE, 0);
node              449 lib/radix-tree.c 					tag_set(node, tag, 0);
node              455 lib/radix-tree.c 			entry_to_node(entry)->parent = node;
node              458 lib/radix-tree.c 			node->nr_values = 1;
node              464 lib/radix-tree.c 		node->slots[0] = (void __rcu *)entry;
node              465 lib/radix-tree.c 		entry = node_to_entry(node);
node              482 lib/radix-tree.c 		struct radix_tree_node *node = rcu_dereference_raw(root->xa_head);
node              485 lib/radix-tree.c 		if (!radix_tree_is_internal_node(node))
node              487 lib/radix-tree.c 		node = entry_to_node(node);
node              493 lib/radix-tree.c 		if (node->count != 1)
node              495 lib/radix-tree.c 		child = rcu_dereference_raw(node->slots[0]);
node              504 lib/radix-tree.c 		if (!node->shift && is_idr(root))
node              518 lib/radix-tree.c 		if (is_idr(root) && !tag_get(node, IDR_FREE, 0))
node              539 lib/radix-tree.c 		node->count = 0;
node              541 lib/radix-tree.c 			node->slots[0] = (void __rcu *)RADIX_TREE_RETRY;
node              544 lib/radix-tree.c 		WARN_ON_ONCE(!list_empty(&node->private_list));
node              545 lib/radix-tree.c 		radix_tree_node_free(node);
node              553 lib/radix-tree.c 			struct radix_tree_node *node)
node              560 lib/radix-tree.c 		if (node->count) {
node              561 lib/radix-tree.c 			if (node_to_entry(node) ==
node              567 lib/radix-tree.c 		parent = node->parent;
node              569 lib/radix-tree.c 			parent->slots[node->offset] = NULL;
node              581 lib/radix-tree.c 		WARN_ON_ONCE(!list_empty(&node->private_list));
node              582 lib/radix-tree.c 		radix_tree_node_free(node);
node              585 lib/radix-tree.c 		node = parent;
node              586 lib/radix-tree.c 	} while (node);
node              611 lib/radix-tree.c 	struct radix_tree_node *node = NULL, *child;
node              633 lib/radix-tree.c 			child = radix_tree_node_alloc(gfp, node, root, shift,
node              638 lib/radix-tree.c 			if (node)
node              639 lib/radix-tree.c 				node->count++;
node              644 lib/radix-tree.c 		node = entry_to_node(child);
node              645 lib/radix-tree.c 		offset = radix_tree_descend(node, &child, index);
node              646 lib/radix-tree.c 		slot = &node->slots[offset];
node              650 lib/radix-tree.c 		*nodep = node;
node              665 lib/radix-tree.c static void radix_tree_free_nodes(struct radix_tree_node *node)
node              668 lib/radix-tree.c 	struct radix_tree_node *child = entry_to_node(node);
node              684 lib/radix-tree.c 			if (old == entry_to_node(node))
node              690 lib/radix-tree.c static inline int insert_entries(struct radix_tree_node *node,
node              696 lib/radix-tree.c 	if (node) {
node              697 lib/radix-tree.c 		node->count++;
node              699 lib/radix-tree.c 			node->nr_values++;
node              715 lib/radix-tree.c 	struct radix_tree_node *node;
node              721 lib/radix-tree.c 	error = __radix_tree_create(root, index, &node, &slot);
node              725 lib/radix-tree.c 	error = insert_entries(node, slot, item, false);
node              729 lib/radix-tree.c 	if (node) {
node              730 lib/radix-tree.c 		unsigned offset = get_slot_offset(node, slot);
node              731 lib/radix-tree.c 		BUG_ON(tag_get(node, 0, offset));
node              732 lib/radix-tree.c 		BUG_ON(tag_get(node, 1, offset));
node              733 lib/radix-tree.c 		BUG_ON(tag_get(node, 2, offset));
node              760 lib/radix-tree.c 	struct radix_tree_node *node, *parent;
node              767 lib/radix-tree.c 	radix_tree_load_root(root, &node, &maxindex);
node              771 lib/radix-tree.c 	while (radix_tree_is_internal_node(node)) {
node              774 lib/radix-tree.c 		parent = entry_to_node(node);
node              775 lib/radix-tree.c 		offset = radix_tree_descend(parent, &node, index);
node              777 lib/radix-tree.c 		if (node == RADIX_TREE_RETRY)
node              787 lib/radix-tree.c 	return node;
node              833 lib/radix-tree.c 		struct radix_tree_node *node, int count, int values)
node              835 lib/radix-tree.c 	if (node && (count || values)) {
node              836 lib/radix-tree.c 		node->count += count;
node              837 lib/radix-tree.c 		node->nr_values += values;
node              844 lib/radix-tree.c 				const struct radix_tree_node *node,
node              847 lib/radix-tree.c 	if (node)
node              848 lib/radix-tree.c 		return tag_get(node, tag, offset);
node              860 lib/radix-tree.c 				struct radix_tree_node *node, void __rcu **slot,
node              864 lib/radix-tree.c 		unsigned offset = get_slot_offset(node, slot);
node              865 lib/radix-tree.c 		bool free = node_tag_get(root, node, IDR_FREE, offset);
node              885 lib/radix-tree.c 			  struct radix_tree_node *node,
node              890 lib/radix-tree.c 	int count = calculate_count(root, node, slot, item, old);
node              897 lib/radix-tree.c 	WARN_ON_ONCE(!node && (slot != (void __rcu **)&root->xa_head) &&
node              899 lib/radix-tree.c 	replace_slot(slot, item, node, count, values);
node              901 lib/radix-tree.c 	if (!node)
node              904 lib/radix-tree.c 	delete_node(root, node);
node              943 lib/radix-tree.c 	__radix_tree_replace(root, iter->node, slot, item);
node              947 lib/radix-tree.c 				struct radix_tree_node *node,
node              950 lib/radix-tree.c 	while (node) {
node              951 lib/radix-tree.c 		if (tag_get(node, tag, offset))
node              953 lib/radix-tree.c 		tag_set(node, tag, offset);
node              954 lib/radix-tree.c 		offset = node->offset;
node              955 lib/radix-tree.c 		node = node->parent;
node              978 lib/radix-tree.c 	struct radix_tree_node *node, *parent;
node              981 lib/radix-tree.c 	radix_tree_load_root(root, &node, &maxindex);
node              984 lib/radix-tree.c 	while (radix_tree_is_internal_node(node)) {
node              987 lib/radix-tree.c 		parent = entry_to_node(node);
node              988 lib/radix-tree.c 		offset = radix_tree_descend(parent, &node, index);
node              989 lib/radix-tree.c 		BUG_ON(!node);
node              999 lib/radix-tree.c 	return node;
node             1004 lib/radix-tree.c 				struct radix_tree_node *node,
node             1007 lib/radix-tree.c 	while (node) {
node             1008 lib/radix-tree.c 		if (!tag_get(node, tag, offset))
node             1010 lib/radix-tree.c 		tag_clear(node, tag, offset);
node             1011 lib/radix-tree.c 		if (any_tag_set(node, tag))
node             1014 lib/radix-tree.c 		offset = node->offset;
node             1015 lib/radix-tree.c 		node = node->parent;
node             1040 lib/radix-tree.c 	struct radix_tree_node *node, *parent;
node             1044 lib/radix-tree.c 	radix_tree_load_root(root, &node, &maxindex);
node             1050 lib/radix-tree.c 	while (radix_tree_is_internal_node(node)) {
node             1051 lib/radix-tree.c 		parent = entry_to_node(node);
node             1052 lib/radix-tree.c 		offset = radix_tree_descend(parent, &node, index);
node             1055 lib/radix-tree.c 	if (node)
node             1058 lib/radix-tree.c 	return node;
node             1071 lib/radix-tree.c 	node_tag_clear(root, iter->node, tag, iter_offset(iter));
node             1092 lib/radix-tree.c 	struct radix_tree_node *node, *parent;
node             1098 lib/radix-tree.c 	radix_tree_load_root(root, &node, &maxindex);
node             1102 lib/radix-tree.c 	while (radix_tree_is_internal_node(node)) {
node             1105 lib/radix-tree.c 		parent = entry_to_node(node);
node             1106 lib/radix-tree.c 		offset = radix_tree_descend(parent, &node, index);
node             1110 lib/radix-tree.c 		if (node == RADIX_TREE_RETRY)
node             1120 lib/radix-tree.c 				struct radix_tree_node *node, unsigned offset,
node             1126 lib/radix-tree.c 	if (!node) {
node             1131 lib/radix-tree.c 	iter->tags = node->tags[tag][tag_long] >> tag_bit;
node             1137 lib/radix-tree.c 			iter->tags |= node->tags[tag][tag_long + 1] <<
node             1167 lib/radix-tree.c 	struct radix_tree_node *node, *child;
node             1198 lib/radix-tree.c 		iter->node = NULL;
node             1203 lib/radix-tree.c 		node = entry_to_node(child);
node             1204 lib/radix-tree.c 		offset = radix_tree_descend(node, &child, index);
node             1207 lib/radix-tree.c 				!tag_get(node, tag, offset) : !child) {
node             1213 lib/radix-tree.c 				offset = radix_tree_find_next_bit(node, tag,
node             1218 lib/radix-tree.c 							node->slots[offset]);
node             1222 lib/radix-tree.c 			index &= ~node_maxindex(node);
node             1223 lib/radix-tree.c 			index += offset << node->shift;
node             1229 lib/radix-tree.c 			child = rcu_dereference_raw(node->slots[offset]);
node             1236 lib/radix-tree.c 	} while (node->shift && radix_tree_is_internal_node(child));
node             1239 lib/radix-tree.c 	iter->index = (index &~ node_maxindex(node)) | offset;
node             1240 lib/radix-tree.c 	iter->next_index = (index | node_maxindex(node)) + 1;
node             1241 lib/radix-tree.c 	iter->node = node;
node             1244 lib/radix-tree.c 		set_iter_tags(iter, node, offset, tag);
node             1246 lib/radix-tree.c 	return node->slots + offset;
node             1374 lib/radix-tree.c 				struct radix_tree_node *node, void __rcu **slot)
node             1378 lib/radix-tree.c 	unsigned offset = get_slot_offset(node, slot);
node             1382 lib/radix-tree.c 		node_tag_set(root, node, IDR_FREE, offset);
node             1385 lib/radix-tree.c 			node_tag_clear(root, node, tag, offset);
node             1387 lib/radix-tree.c 	replace_slot(slot, NULL, node, -1, values);
node             1388 lib/radix-tree.c 	return node && delete_node(root, node);
node             1406 lib/radix-tree.c 	if (__radix_tree_delete(root, iter->node, slot))
node             1425 lib/radix-tree.c 	struct radix_tree_node *node = NULL;
node             1429 lib/radix-tree.c 	entry = __radix_tree_lookup(root, index, &node, &slot);
node             1432 lib/radix-tree.c 	if (!entry && (!is_idr(root) || node_tag_get(root, node, IDR_FREE,
node             1433 lib/radix-tree.c 						get_slot_offset(node, slot))))
node             1439 lib/radix-tree.c 	__radix_tree_delete(root, node, slot);
node             1489 lib/radix-tree.c 	struct radix_tree_node *node = NULL, *child;
node             1515 lib/radix-tree.c 			child = radix_tree_node_alloc(gfp, node, root, shift,
node             1521 lib/radix-tree.c 			if (node)
node             1522 lib/radix-tree.c 				node->count++;
node             1526 lib/radix-tree.c 		node = entry_to_node(child);
node             1527 lib/radix-tree.c 		offset = radix_tree_descend(node, &child, start);
node             1528 lib/radix-tree.c 		if (!tag_get(node, IDR_FREE, offset)) {
node             1529 lib/radix-tree.c 			offset = radix_tree_find_next_bit(node, IDR_FREE,
node             1531 lib/radix-tree.c 			start = next_index(start, node, offset);
node             1535 lib/radix-tree.c 				offset = node->offset + 1;
node             1536 lib/radix-tree.c 				node = node->parent;
node             1537 lib/radix-tree.c 				if (!node)
node             1539 lib/radix-tree.c 				shift = node->shift;
node             1541 lib/radix-tree.c 			child = rcu_dereference_raw(node->slots[offset]);
node             1543 lib/radix-tree.c 		slot = &node->slots[offset];
node             1547 lib/radix-tree.c 	if (node)
node             1548 lib/radix-tree.c 		iter->next_index = 1 + min(max, (start | node_maxindex(node)));
node             1551 lib/radix-tree.c 	iter->node = node;
node             1552 lib/radix-tree.c 	set_iter_tags(iter, node, offset, IDR_FREE);
node             1570 lib/radix-tree.c 	struct radix_tree_node *node = rcu_dereference_raw(idr->idr_rt.xa_head);
node             1571 lib/radix-tree.c 	if (radix_tree_is_internal_node(node))
node             1572 lib/radix-tree.c 		radix_tree_free_nodes(node);
node             1581 lib/radix-tree.c 	struct radix_tree_node *node = arg;
node             1583 lib/radix-tree.c 	memset(node, 0, sizeof(*node));
node             1584 lib/radix-tree.c 	INIT_LIST_HEAD(&node->private_list);
node             1590 lib/radix-tree.c 	struct radix_tree_node *node;
node             1595 lib/radix-tree.c 		node = rtp->nodes;
node             1596 lib/radix-tree.c 		rtp->nodes = node->parent;
node             1597 lib/radix-tree.c 		kmem_cache_free(radix_tree_node_cachep, node);
node               85 lib/rbtree.c   __rb_insert(struct rb_node *node, struct rb_root *root,
node               88 lib/rbtree.c   	struct rb_node *parent = rb_red_parent(node), *gparent, *tmp;
node              100 lib/rbtree.c   			rb_set_parent_color(node, NULL, RB_BLACK);
node              133 lib/rbtree.c   				node = gparent;
node              134 lib/rbtree.c   				parent = rb_parent(node);
node              135 lib/rbtree.c   				rb_set_parent_color(node, parent, RB_RED);
node              140 lib/rbtree.c   			if (node == tmp) {
node              154 lib/rbtree.c   				tmp = node->rb_left;
node              156 lib/rbtree.c   				WRITE_ONCE(node->rb_left, parent);
node              160 lib/rbtree.c   				rb_set_parent_color(parent, node, RB_RED);
node              161 lib/rbtree.c   				augment_rotate(parent, node);
node              162 lib/rbtree.c   				parent = node;
node              163 lib/rbtree.c   				tmp = node->rb_right;
node              189 lib/rbtree.c   				node = gparent;
node              190 lib/rbtree.c   				parent = rb_parent(node);
node              191 lib/rbtree.c   				rb_set_parent_color(node, parent, RB_RED);
node              196 lib/rbtree.c   			if (node == tmp) {
node              198 lib/rbtree.c   				tmp = node->rb_right;
node              200 lib/rbtree.c   				WRITE_ONCE(node->rb_right, parent);
node              204 lib/rbtree.c   				rb_set_parent_color(parent, node, RB_RED);
node              205 lib/rbtree.c   				augment_rotate(parent, node);
node              206 lib/rbtree.c   				parent = node;
node              207 lib/rbtree.c   				tmp = node->rb_left;
node              230 lib/rbtree.c   	struct rb_node *node = NULL, *sibling, *tmp1, *tmp2;
node              241 lib/rbtree.c   		if (node != sibling) {	/* node == parent->rb_left */
node              285 lib/rbtree.c   						node = parent;
node              286 lib/rbtree.c   						parent = rb_parent(node);
node              375 lib/rbtree.c   						node = parent;
node              376 lib/rbtree.c   						parent = rb_parent(node);
node              424 lib/rbtree.c   static inline void dummy_propagate(struct rb_node *node, struct rb_node *stop) {}
node              434 lib/rbtree.c   void rb_insert_color(struct rb_node *node, struct rb_root *root)
node              436 lib/rbtree.c   	__rb_insert(node, root, dummy_rotate);
node              440 lib/rbtree.c   void rb_erase(struct rb_node *node, struct rb_root *root)
node              443 lib/rbtree.c   	rebalance = __rb_erase_augmented(node, root, &dummy_callbacks);
node              456 lib/rbtree.c   void __rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node              459 lib/rbtree.c   	__rb_insert(node, root, augment_rotate);
node              492 lib/rbtree.c   struct rb_node *rb_next(const struct rb_node *node)
node              496 lib/rbtree.c   	if (RB_EMPTY_NODE(node))
node              503 lib/rbtree.c   	if (node->rb_right) {
node              504 lib/rbtree.c   		node = node->rb_right;
node              505 lib/rbtree.c   		while (node->rb_left)
node              506 lib/rbtree.c   			node=node->rb_left;
node              507 lib/rbtree.c   		return (struct rb_node *)node;
node              517 lib/rbtree.c   	while ((parent = rb_parent(node)) && node == parent->rb_right)
node              518 lib/rbtree.c   		node = parent;
node              524 lib/rbtree.c   struct rb_node *rb_prev(const struct rb_node *node)
node              528 lib/rbtree.c   	if (RB_EMPTY_NODE(node))
node              535 lib/rbtree.c   	if (node->rb_left) {
node              536 lib/rbtree.c   		node = node->rb_left;
node              537 lib/rbtree.c   		while (node->rb_right)
node              538 lib/rbtree.c   			node=node->rb_right;
node              539 lib/rbtree.c   		return (struct rb_node *)node;
node              546 lib/rbtree.c   	while ((parent = rb_parent(node)) && node == parent->rb_left)
node              547 lib/rbtree.c   		node = parent;
node              592 lib/rbtree.c   static struct rb_node *rb_left_deepest_node(const struct rb_node *node)
node              595 lib/rbtree.c   		if (node->rb_left)
node              596 lib/rbtree.c   			node = node->rb_left;
node              597 lib/rbtree.c   		else if (node->rb_right)
node              598 lib/rbtree.c   			node = node->rb_right;
node              600 lib/rbtree.c   			return (struct rb_node *)node;
node              604 lib/rbtree.c   struct rb_node *rb_next_postorder(const struct rb_node *node)
node              607 lib/rbtree.c   	if (!node)
node              609 lib/rbtree.c   	parent = rb_parent(node);
node              612 lib/rbtree.c   	if (parent && node == parent->rb_left && parent->rb_right) {
node               32 lib/rbtree_test.c static void insert(struct test_node *node, struct rb_root_cached *root)
node               35 lib/rbtree_test.c 	u32 key = node->key;
node               45 lib/rbtree_test.c 	rb_link_node(&node->rb, parent, new);
node               46 lib/rbtree_test.c 	rb_insert_color(&node->rb, &root->rb_root);
node               49 lib/rbtree_test.c static void insert_cached(struct test_node *node, struct rb_root_cached *root)
node               52 lib/rbtree_test.c 	u32 key = node->key;
node               65 lib/rbtree_test.c 	rb_link_node(&node->rb, parent, new);
node               66 lib/rbtree_test.c 	rb_insert_color_cached(&node->rb, root, leftmost);
node               69 lib/rbtree_test.c static inline void erase(struct test_node *node, struct rb_root_cached *root)
node               71 lib/rbtree_test.c 	rb_erase(&node->rb, &root->rb_root);
node               74 lib/rbtree_test.c static inline void erase_cached(struct test_node *node, struct rb_root_cached *root)
node               76 lib/rbtree_test.c 	rb_erase_cached(&node->rb, root);
node               80 lib/rbtree_test.c #define NODE_VAL(node) ((node)->val)
node               85 lib/rbtree_test.c static void insert_augmented(struct test_node *node,
node               89 lib/rbtree_test.c 	u32 key = node->key;
node               90 lib/rbtree_test.c 	u32 val = node->val;
node              104 lib/rbtree_test.c 	node->augmented = val;
node              105 lib/rbtree_test.c 	rb_link_node(&node->rb, rb_parent, new);
node              106 lib/rbtree_test.c 	rb_insert_augmented(&node->rb, &root->rb_root, &augment_callbacks);
node              109 lib/rbtree_test.c static void insert_augmented_cached(struct test_node *node,
node              113 lib/rbtree_test.c 	u32 key = node->key;
node              114 lib/rbtree_test.c 	u32 val = node->val;
node              131 lib/rbtree_test.c 	node->augmented = val;
node              132 lib/rbtree_test.c 	rb_link_node(&node->rb, rb_parent, new);
node              133 lib/rbtree_test.c 	rb_insert_augmented_cached(&node->rb, root,
node              138 lib/rbtree_test.c static void erase_augmented(struct test_node *node, struct rb_root_cached *root)
node              140 lib/rbtree_test.c 	rb_erase_augmented(&node->rb, &root->rb_root, &augment_callbacks);
node              143 lib/rbtree_test.c static void erase_augmented_cached(struct test_node *node,
node              146 lib/rbtree_test.c 	rb_erase_augmented_cached(&node->rb, root, &augment_callbacks);
node              198 lib/rbtree_test.c 		struct test_node *node = rb_entry(rb, struct test_node, rb);
node              199 lib/rbtree_test.c 		WARN_ON_ONCE(node->key < prev_key);
node              207 lib/rbtree_test.c 		prev_key = node->key;
node              224 lib/rbtree_test.c 		struct test_node *node = rb_entry(rb, struct test_node, rb);
node              225 lib/rbtree_test.c 		u32 subtree, max = node->val;
node              226 lib/rbtree_test.c 		if (node->rb.rb_left) {
node              227 lib/rbtree_test.c 			subtree = rb_entry(node->rb.rb_left, struct test_node,
node              232 lib/rbtree_test.c 		if (node->rb.rb_right) {
node              233 lib/rbtree_test.c 			subtree = rb_entry(node->rb.rb_right, struct test_node,
node              238 lib/rbtree_test.c 		WARN_ON_ONCE(node->augmented != max);
node              246 lib/rbtree_test.c 	struct rb_node *node;
node              295 lib/rbtree_test.c 		for (node = rb_first(&root.rb_root); node; node = rb_next(node))
node              309 lib/rbtree_test.c 		node = rb_first(&root.rb_root);
node              321 lib/rbtree_test.c 		node = rb_first_cached(&root);
node               45 lib/sbitmap.c  		      gfp_t flags, int node)
node               76 lib/sbitmap.c  	sb->map = kcalloc_node(sb->map_nr, sizeof(*sb->map), flags, node);
node              370 lib/sbitmap.c  			    int shift, bool round_robin, gfp_t flags, int node)
node              375 lib/sbitmap.c  	ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node);
node              395 lib/sbitmap.c  	sbq->ws = kzalloc_node(SBQ_WAIT_QUEUES * sizeof(*sbq->ws), flags, node);
node               64 lib/test_rhashtable.c 	struct rhash_head	node;
node               95 lib/test_rhashtable.c 	.head_offset = offsetof(struct test_obj, node),
node              123 lib/test_rhashtable.c 		err = rhashtable_insert_fast(ht, &obj->node, params);
node              254 lib/test_rhashtable.c 			rhashtable_remove_fast(ht, &obj->node, test_rht_params);
node              670 lib/test_rhashtable.c 			err = rhashtable_remove_fast(&ht, &tdata->objs[i].node,
node              305 lib/test_xarray.c 	struct xa_node *node;
node              319 lib/test_xarray.c 	node = xas.xa_node;
node              320 lib/test_xarray.c 	XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != xa_mk_value(0));
node              324 lib/test_xarray.c 	XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != XA_RETRY_ENTRY);
node              337 lib/test_xarray.c 		node = xa_head(xa);
node              342 lib/test_xarray.c 		XA_BUG_ON(xa, xa_head(xa) == node);
node              346 lib/test_xarray.c 		XA_BUG_ON(xa, xa->xa_head != node);
node             1560 lib/test_xarray.c static void test_update_node(struct xa_node *node)
node             1562 lib/test_xarray.c 	if (node->count && node->count == node->nr_values) {
node             1563 lib/test_xarray.c 		if (list_empty(&node->private_list))
node             1564 lib/test_xarray.c 			list_add(&shadow_nodes, &node->private_list);
node             1566 lib/test_xarray.c 		if (!list_empty(&node->private_list))
node             1567 lib/test_xarray.c 			list_del_init(&node->private_list);
node             1573 lib/test_xarray.c 	struct xa_node *node;
node             1576 lib/test_xarray.c 	while ((node = list_first_entry_or_null(&shadow_nodes,
node             1578 lib/test_xarray.c 		XA_STATE(xas, node->array, 0);
node             1579 lib/test_xarray.c 		XA_BUG_ON(xa, node->array != xa);
node             1580 lib/test_xarray.c 		list_del_init(&node->private_list);
node             1581 lib/test_xarray.c 		xas.xa_node = xa_parent_locked(node->array, node);
node             1582 lib/test_xarray.c 		xas.xa_offset = node->offset;
node             1583 lib/test_xarray.c 		xas.xa_shift = node->shift + XA_CHUNK_SHIFT;
node               27 lib/timerqueue.c bool timerqueue_add(struct timerqueue_head *head, struct timerqueue_node *node)
node               35 lib/timerqueue.c 	WARN_ON_ONCE(!RB_EMPTY_NODE(&node->node));
node               39 lib/timerqueue.c 		ptr = rb_entry(parent, struct timerqueue_node, node);
node               40 lib/timerqueue.c 		if (node->expires < ptr->expires) {
node               47 lib/timerqueue.c 	rb_link_node(&node->node, parent, p);
node               48 lib/timerqueue.c 	rb_insert_color_cached(&node->node, &head->rb_root, leftmost);
node               63 lib/timerqueue.c bool timerqueue_del(struct timerqueue_head *head, struct timerqueue_node *node)
node               65 lib/timerqueue.c 	WARN_ON_ONCE(RB_EMPTY_NODE(&node->node));
node               67 lib/timerqueue.c 	rb_erase_cached(&node->node, &head->rb_root);
node               68 lib/timerqueue.c 	RB_CLEAR_NODE(&node->node);
node               83 lib/timerqueue.c struct timerqueue_node *timerqueue_iterate_next(struct timerqueue_node *node)
node               87 lib/timerqueue.c 	if (!node)
node               89 lib/timerqueue.c 	next = rb_next(&node->node);
node               92 lib/timerqueue.c 	return container_of(next, struct timerqueue_node, node);
node               78 lib/xarray.c   static inline unsigned long *node_marks(struct xa_node *node, xa_mark_t mark)
node               80 lib/xarray.c   	return node->marks[(__force unsigned)mark];
node               83 lib/xarray.c   static inline bool node_get_mark(struct xa_node *node,
node               86 lib/xarray.c   	return test_bit(offset, node_marks(node, mark));
node               90 lib/xarray.c   static inline bool node_set_mark(struct xa_node *node, unsigned int offset,
node               93 lib/xarray.c   	return __test_and_set_bit(offset, node_marks(node, mark));
node               97 lib/xarray.c   static inline bool node_clear_mark(struct xa_node *node, unsigned int offset,
node              100 lib/xarray.c   	return __test_and_clear_bit(offset, node_marks(node, mark));
node              103 lib/xarray.c   static inline bool node_any_mark(struct xa_node *node, xa_mark_t mark)
node              105 lib/xarray.c   	return !bitmap_empty(node_marks(node, mark), XA_CHUNK_SIZE);
node              108 lib/xarray.c   static inline void node_mark_all(struct xa_node *node, xa_mark_t mark)
node              110 lib/xarray.c   	bitmap_fill(node_marks(node, mark), XA_CHUNK_SIZE);
node              142 lib/xarray.c   static unsigned int get_offset(unsigned long index, struct xa_node *node)
node              144 lib/xarray.c   	return (index >> node->shift) & XA_CHUNK_MASK;
node              201 lib/xarray.c   static void *xas_descend(struct xa_state *xas, struct xa_node *node)
node              203 lib/xarray.c   	unsigned int offset = get_offset(xas->xa_index, node);
node              204 lib/xarray.c   	void *entry = xa_entry(xas->xa, node, offset);
node              206 lib/xarray.c   	xas->xa_node = node;
node              209 lib/xarray.c   		entry = xa_entry(xas->xa, node, offset);
node              236 lib/xarray.c   		struct xa_node *node = xa_to_node(entry);
node              238 lib/xarray.c   		if (xas->xa_shift > node->shift)
node              240 lib/xarray.c   		entry = xas_descend(xas, node);
node              241 lib/xarray.c   		if (node->shift == 0)
node              254 lib/xarray.c   static void xa_node_free(struct xa_node *node)
node              256 lib/xarray.c   	XA_NODE_BUG_ON(node, !list_empty(&node->private_list));
node              257 lib/xarray.c   	node->array = XA_RCU_FREE;
node              258 lib/xarray.c   	call_rcu(&node->rcu_head, radix_tree_node_rcu_free);
node              269 lib/xarray.c   	struct xa_node *node = xas->xa_alloc;
node              271 lib/xarray.c   	if (!node)
node              273 lib/xarray.c   	XA_NODE_BUG_ON(node, !list_empty(&node->private_list));
node              274 lib/xarray.c   	kmem_cache_free(radix_tree_node_cachep, node);
node              347 lib/xarray.c   static void xas_update(struct xa_state *xas, struct xa_node *node)
node              350 lib/xarray.c   		xas->xa_update(node);
node              352 lib/xarray.c   		XA_NODE_BUG_ON(node, !list_empty(&node->private_list));
node              358 lib/xarray.c   	struct xa_node *node = xas->xa_alloc;
node              363 lib/xarray.c   	if (node) {
node              371 lib/xarray.c   		node = kmem_cache_alloc(radix_tree_node_cachep, gfp);
node              372 lib/xarray.c   		if (!node) {
node              379 lib/xarray.c   		node->offset = xas->xa_offset;
node              381 lib/xarray.c   		XA_NODE_BUG_ON(node, parent->count > XA_CHUNK_SIZE);
node              384 lib/xarray.c   	XA_NODE_BUG_ON(node, shift > BITS_PER_LONG);
node              385 lib/xarray.c   	XA_NODE_BUG_ON(node, !list_empty(&node->private_list));
node              386 lib/xarray.c   	node->shift = shift;
node              387 lib/xarray.c   	node->count = 0;
node              388 lib/xarray.c   	node->nr_values = 0;
node              389 lib/xarray.c   	RCU_INIT_POINTER(node->parent, xas->xa_node);
node              390 lib/xarray.c   	node->array = xas->xa;
node              392 lib/xarray.c   	return node;
node              436 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              441 lib/xarray.c   		XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE);
node              442 lib/xarray.c   		if (node->count != 1)
node              444 lib/xarray.c   		entry = xa_entry_locked(xa, node, 0);
node              447 lib/xarray.c   		if (!xa_is_node(entry) && node->shift)
node              454 lib/xarray.c   		if (xa_track_free(xa) && !node_get_mark(node, 0, XA_FREE_MARK))
node              457 lib/xarray.c   		node->count = 0;
node              458 lib/xarray.c   		node->nr_values = 0;
node              460 lib/xarray.c   			RCU_INIT_POINTER(node->slots[0], XA_RETRY_ENTRY);
node              461 lib/xarray.c   		xas_update(xas, node);
node              462 lib/xarray.c   		xa_node_free(node);
node              465 lib/xarray.c   		node = xa_to_node(entry);
node              466 lib/xarray.c   		node->parent = NULL;
node              479 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              484 lib/xarray.c   		XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE);
node              485 lib/xarray.c   		if (node->count)
node              488 lib/xarray.c   		parent = xa_parent_locked(xas->xa, node);
node              490 lib/xarray.c   		xas->xa_offset = node->offset;
node              491 lib/xarray.c   		xa_node_free(node);
node              502 lib/xarray.c   		node = parent;
node              503 lib/xarray.c   		xas_update(xas, node);
node              506 lib/xarray.c   	if (!node->parent)
node              522 lib/xarray.c   	struct xa_node *node = top;
node              525 lib/xarray.c   		void *entry = xa_entry_locked(xas->xa, node, offset);
node              527 lib/xarray.c   		if (node->shift && xa_is_node(entry)) {
node              528 lib/xarray.c   			node = xa_to_node(entry);
node              533 lib/xarray.c   			RCU_INIT_POINTER(node->slots[offset], XA_RETRY_ENTRY);
node              538 lib/xarray.c   			parent = xa_parent_locked(xas->xa, node);
node              539 lib/xarray.c   			offset = node->offset + 1;
node              540 lib/xarray.c   			node->count = 0;
node              541 lib/xarray.c   			node->nr_values = 0;
node              542 lib/xarray.c   			xas_update(xas, node);
node              543 lib/xarray.c   			xa_node_free(node);
node              544 lib/xarray.c   			if (node == top)
node              546 lib/xarray.c   			node = parent;
node              558 lib/xarray.c   	struct xa_node *node = NULL;
node              569 lib/xarray.c   		node = xa_to_node(head);
node              570 lib/xarray.c   		shift = node->shift + XA_CHUNK_SHIFT;
node              577 lib/xarray.c   		XA_NODE_BUG_ON(node, shift > BITS_PER_LONG);
node              578 lib/xarray.c   		node = xas_alloc(xas, shift);
node              579 lib/xarray.c   		if (!node)
node              582 lib/xarray.c   		node->count = 1;
node              584 lib/xarray.c   			node->nr_values = 1;
node              585 lib/xarray.c   		RCU_INIT_POINTER(node->slots[0], head);
node              590 lib/xarray.c   				node_mark_all(node, XA_FREE_MARK);
node              592 lib/xarray.c   					node_clear_mark(node, 0, XA_FREE_MARK);
node              596 lib/xarray.c   				node_set_mark(node, 0, mark);
node              609 lib/xarray.c   			rcu_assign_pointer(xa_to_node(head)->parent, node);
node              611 lib/xarray.c   		head = xa_mk_node(node);
node              613 lib/xarray.c   		xas_update(xas, node);
node              618 lib/xarray.c   	xas->xa_node = node;
node              640 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              644 lib/xarray.c   	if (xas_top(node)) {
node              658 lib/xarray.c   	} else if (node) {
node              661 lib/xarray.c   		shift = node->shift;
node              662 lib/xarray.c   		entry = xa_entry_locked(xa, node, offset);
node              663 lib/xarray.c   		slot = &node->slots[offset];
node              673 lib/xarray.c   			node = xas_alloc(xas, shift);
node              674 lib/xarray.c   			if (!node)
node              677 lib/xarray.c   				node_mark_all(node, XA_FREE_MARK);
node              678 lib/xarray.c   			rcu_assign_pointer(*slot, xa_mk_node(node));
node              680 lib/xarray.c   			node = xa_to_node(entry);
node              684 lib/xarray.c   		entry = xas_descend(xas, node);
node              685 lib/xarray.c   		slot = &node->slots[xas->xa_offset];
node              721 lib/xarray.c   			struct xa_node *node = xas->xa_node;
node              722 lib/xarray.c   			xas->xa_node = xa_parent_locked(xas->xa, node);
node              723 lib/xarray.c   			xas->xa_offset = node->offset - 1;
node              724 lib/xarray.c   			if (node->offset != 0)
node              741 lib/xarray.c   static void update_node(struct xa_state *xas, struct xa_node *node,
node              744 lib/xarray.c   	if (!node || (!count && !values))
node              747 lib/xarray.c   	node->count += count;
node              748 lib/xarray.c   	node->nr_values += values;
node              749 lib/xarray.c   	XA_NODE_BUG_ON(node, node->count > XA_CHUNK_SIZE);
node              750 lib/xarray.c   	XA_NODE_BUG_ON(node, node->nr_values > XA_CHUNK_SIZE);
node              751 lib/xarray.c   	xas_update(xas, node);
node              771 lib/xarray.c   	struct xa_node *node;
node              788 lib/xarray.c   	node = xas->xa_node;
node              789 lib/xarray.c   	if (node && (xas->xa_shift < node->shift))
node              797 lib/xarray.c   	if (node) {
node              798 lib/xarray.c   		slot = &node->slots[offset];
node              814 lib/xarray.c   		if (xa_is_node(next) && (!node || node->shift))
node              816 lib/xarray.c   		if (!node)
node              829 lib/xarray.c   		next = xa_entry_locked(xas->xa, node, ++offset);
node              838 lib/xarray.c   	update_node(xas, node, count, values);
node              872 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              878 lib/xarray.c   	while (node) {
node              879 lib/xarray.c   		if (node_set_mark(node, offset, mark))
node              881 lib/xarray.c   		offset = node->offset;
node              882 lib/xarray.c   		node = xa_parent_locked(xas->xa, node);
node              901 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              907 lib/xarray.c   	while (node) {
node              908 lib/xarray.c   		if (!node_clear_mark(node, offset, mark))
node              910 lib/xarray.c   		if (node_any_mark(node, mark))
node              913 lib/xarray.c   		offset = node->offset;
node              914 lib/xarray.c   		node = xa_parent_locked(xas->xa, node);
node              966 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node              972 lib/xarray.c   	if (node) {
node              975 lib/xarray.c   			if (!xa_is_sibling(xa_entry(xas->xa, node, offset)))
node              978 lib/xarray.c   		xas->xa_index += (offset - xas->xa_offset) << node->shift;
node             1253 lib/xarray.c   			struct xa_node *node = xa_to_node(curr);
node             1254 lib/xarray.c   			curr = xas_descend(xas, node);
node             1838 lib/xarray.c   	struct xa_node *node = xas->xa_node;
node             1841 lib/xarray.c   	if (!node)
node             1843 lib/xarray.c   	mask = (XA_CHUNK_SIZE << node->shift) - 1;
node             1845 lib/xarray.c   		((unsigned long)xas->xa_offset << node->shift);
node             2007 lib/xarray.c   void xa_dump_node(const struct xa_node *node)
node             2011 lib/xarray.c   	if (!node)
node             2013 lib/xarray.c   	if ((unsigned long)node & 3) {
node             2014 lib/xarray.c   		pr_cont("node %px\n", node);
node             2020 lib/xarray.c   		node, node->parent ? "offset" : "max", node->offset,
node             2021 lib/xarray.c   		node->parent, node->shift, node->count, node->nr_values,
node             2022 lib/xarray.c   		node->array, node->private_list.prev, node->private_list.next);
node             2025 lib/xarray.c   			pr_cont(" %lx", node->marks[i][j]);
node             2051 lib/xarray.c   			struct xa_node *node = xa_to_node(entry);
node             2052 lib/xarray.c   			xa_dump_node(node);
node             2054 lib/xarray.c   				xa_dump_entry(node->slots[i],
node             2055 lib/xarray.c   				      index + (i << node->shift), node->shift);
node              548 lib/zlib_deflate/deftree.c     int node;          /* new node being created */
node              571 lib/zlib_deflate/deftree.c         node = s->heap[++(s->heap_len)] = (max_code < 2 ? ++max_code : 0);
node              572 lib/zlib_deflate/deftree.c         tree[node].Freq = 1;
node              573 lib/zlib_deflate/deftree.c         s->depth[node] = 0;
node              574 lib/zlib_deflate/deftree.c         s->opt_len--; if (stree) s->static_len -= stree[node].Len;
node              587 lib/zlib_deflate/deftree.c     node = elems;              /* next internal node of the tree */
node              596 lib/zlib_deflate/deftree.c         tree[node].Freq = tree[n].Freq + tree[m].Freq;
node              597 lib/zlib_deflate/deftree.c         s->depth[node] = (uch) (max(s->depth[n], s->depth[m]) + 1);
node              598 lib/zlib_deflate/deftree.c         tree[n].Dad = tree[m].Dad = (ush)node;
node              602 lib/zlib_deflate/deftree.c                     node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq);
node              606 lib/zlib_deflate/deftree.c         s->heap[SMALLEST] = node++;
node              409 mm/backing-dev.c 	struct rb_node **node, *parent;
node              414 mm/backing-dev.c 	node = &bdi->cgwb_congested_tree.rb_node;
node              417 mm/backing-dev.c 	while (*node != NULL) {
node              418 mm/backing-dev.c 		parent = *node;
node              422 mm/backing-dev.c 			node = &parent->rb_left;
node              424 mm/backing-dev.c 			node = &parent->rb_right;
node              432 mm/backing-dev.c 		rb_link_node(&congested->rb_node, parent, node);
node               19 mm/cma_debug.c 	struct hlist_node node;
node               74 mm/cma_debug.c 	hlist_add_head(&mem->node, &cma->mem_head);
node               84 mm/cma_debug.c 		mem = hlist_entry(cma->mem_head.first, struct cma_mem, node);
node               85 mm/cma_debug.c 		hlist_del_init(&mem->node);
node             2484 mm/compaction.c int compaction_register_node(struct node *node)
node             2486 mm/compaction.c 	return device_create_file(&node->dev, &dev_attr_compact);
node             2489 mm/compaction.c void compaction_unregister_node(struct node *node)
node             2491 mm/compaction.c 	return device_remove_file(&node->dev, &dev_attr_compact);
node              901 mm/hugetlb.c   	int node = NUMA_NO_NODE;
node              916 mm/hugetlb.c   		if (zone_to_nid(zone) == node)
node              918 mm/hugetlb.c   		node = zone_to_nid(zone);
node              920 mm/hugetlb.c   		page = dequeue_huge_page_node_exact(h, node);
node             1037 mm/hugetlb.c   #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask)		\
node             1040 mm/hugetlb.c   		((node = hstate_next_node_to_alloc(hs, mask)) || 1);	\
node             1043 mm/hugetlb.c   #define for_each_node_mask_to_free(hs, nr_nodes, node, mask)		\
node             1046 mm/hugetlb.c   		((node = hstate_next_node_to_free(hs, mask)) || 1);	\
node             1336 mm/hugetlb.c   	struct llist_node *node;
node             1339 mm/hugetlb.c   	node = llist_del_all(&hpage_freelist);
node             1341 mm/hugetlb.c   	while (node) {
node             1342 mm/hugetlb.c   		page = container_of((struct address_space **)node,
node             1344 mm/hugetlb.c   		node = node->next;
node             1535 mm/hugetlb.c   	int nr_nodes, node;
node             1538 mm/hugetlb.c   	for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
node             1539 mm/hugetlb.c   		page = alloc_fresh_huge_page(h, gfp_mask, node, nodes_allowed,
node             1562 mm/hugetlb.c   	int nr_nodes, node;
node             1565 mm/hugetlb.c   	for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) {
node             1570 mm/hugetlb.c   		if ((!acct_surplus || h->surplus_huge_pages_node[node]) &&
node             1571 mm/hugetlb.c   		    !list_empty(&h->hugepage_freelists[node])) {
node             1573 mm/hugetlb.c   				list_entry(h->hugepage_freelists[node].next,
node             1577 mm/hugetlb.c   			h->free_huge_pages_node[node]--;
node             1580 mm/hugetlb.c   				h->surplus_huge_pages_node[node]--;
node             1802 mm/hugetlb.c   	int node;
node             1805 mm/hugetlb.c   	node = huge_node(vma, address, gfp_mask, &mpol, &nodemask);
node             1806 mm/hugetlb.c   	page = alloc_huge_page_nodemask(h, node, nodemask);
node             2222 mm/hugetlb.c   	int nr_nodes, node;
node             2224 mm/hugetlb.c   	for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) {
node             2229 mm/hugetlb.c   				0, MEMBLOCK_ALLOC_ACCESSIBLE, node);
node             2399 mm/hugetlb.c   	int nr_nodes, node;
node             2404 mm/hugetlb.c   		for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
node             2405 mm/hugetlb.c   			if (h->surplus_huge_pages_node[node])
node             2409 mm/hugetlb.c   		for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) {
node             2410 mm/hugetlb.c   			if (h->surplus_huge_pages_node[node] <
node             2411 mm/hugetlb.c   					h->nr_huge_pages_node[node])
node             2419 mm/hugetlb.c   	h->surplus_huge_pages_node[node] += delta;
node             2856 mm/hugetlb.c   static void hugetlb_unregister_node(struct node *node)
node             2859 mm/hugetlb.c   	struct node_hstate *nhs = &node_hstates[node->dev.id];
node             2881 mm/hugetlb.c   static void hugetlb_register_node(struct node *node)
node             2884 mm/hugetlb.c   	struct node_hstate *nhs = &node_hstates[node->dev.id];
node             2891 mm/hugetlb.c   							&node->dev.kobj);
node             2901 mm/hugetlb.c   				h->name, node->dev.id);
node             2902 mm/hugetlb.c   			hugetlb_unregister_node(node);
node             2918 mm/hugetlb.c   		struct node *node = node_devices[nid];
node             2919 mm/hugetlb.c   		if (node->dev.id == nid)
node             2920 mm/hugetlb.c   			hugetlb_register_node(node);
node             3075 mm/hugetlb.c   	int node;
node             3078 mm/hugetlb.c   	for_each_node_mask(node, cpuset_current_mems_allowed)
node             3079 mm/hugetlb.c   		nr += array[node];
node              574 mm/internal.h  extern struct page *alloc_new_node_page(struct page *page, unsigned long node);
node               28 mm/interval_tree.c void vma_interval_tree_insert_after(struct vm_area_struct *node,
node               34 mm/interval_tree.c 	unsigned long last = vma_last_pgoff(node);
node               36 mm/interval_tree.c 	VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node);
node               55 mm/interval_tree.c 	node->shared.rb_subtree_last = last;
node               56 mm/interval_tree.c 	rb_link_node(&node->shared.rb, &parent->shared.rb, link);
node               57 mm/interval_tree.c 	rb_insert_augmented(&node->shared.rb, &root->rb_root,
node               75 mm/interval_tree.c void anon_vma_interval_tree_insert(struct anon_vma_chain *node,
node               79 mm/interval_tree.c 	node->cached_vma_start = avc_start_pgoff(node);
node               80 mm/interval_tree.c 	node->cached_vma_last = avc_last_pgoff(node);
node               82 mm/interval_tree.c 	__anon_vma_interval_tree_insert(node, root);
node               85 mm/interval_tree.c void anon_vma_interval_tree_remove(struct anon_vma_chain *node,
node               88 mm/interval_tree.c 	__anon_vma_interval_tree_remove(node, root);
node               99 mm/interval_tree.c anon_vma_interval_tree_iter_next(struct anon_vma_chain *node,
node              102 mm/interval_tree.c 	return __anon_vma_interval_tree_iter_next(node, first, last);
node              106 mm/interval_tree.c void anon_vma_interval_tree_verify(struct anon_vma_chain *node)
node              108 mm/interval_tree.c 	WARN_ON_ONCE(node->cached_vma_start != avc_start_pgoff(node));
node              109 mm/interval_tree.c 	WARN_ON_ONCE(node->cached_vma_last != avc_last_pgoff(node));
node               84 mm/kasan/init.c static __init void *early_alloc(size_t size, int node)
node               87 mm/kasan/init.c 					   MEMBLOCK_ALLOC_ACCESSIBLE, node);
node               91 mm/kasan/init.c 		      __func__, size, size, node, (u64)__pa(MAX_DMA_ADDRESS));
node              781 mm/khugepaged.c khugepaged_alloc_page(struct page **hpage, gfp_t gfp, int node)
node              785 mm/khugepaged.c 	*hpage = __alloc_pages_node(node, gfp, HPAGE_PMD_ORDER);
node              845 mm/khugepaged.c khugepaged_alloc_page(struct page **hpage, gfp_t gfp, int node)
node              949 mm/khugepaged.c 				   int node, int referenced)
node              974 mm/khugepaged.c 	new_page = khugepaged_alloc_page(hpage, gfp, node);
node             1126 mm/khugepaged.c 	int node = NUMA_NO_NODE, unmapped = 0;
node             1184 mm/khugepaged.c 		node = page_to_nid(page);
node             1185 mm/khugepaged.c 		if (khugepaged_scan_abort(node)) {
node             1189 mm/khugepaged.c 		khugepaged_node_load[node]++;
node             1230 mm/khugepaged.c 		node = khugepaged_find_target_node();
node             1232 mm/khugepaged.c 		collapse_huge_page(mm, address, hpage, node, referenced);
node             1494 mm/khugepaged.c 		struct page **hpage, int node)
node             1512 mm/khugepaged.c 	new_page = khugepaged_alloc_page(hpage, gfp, node);
node             1822 mm/khugepaged.c 	int node = NUMA_NO_NODE;
node             1846 mm/khugepaged.c 		node = page_to_nid(page);
node             1847 mm/khugepaged.c 		if (khugepaged_scan_abort(node)) {
node             1851 mm/khugepaged.c 		khugepaged_node_load[node]++;
node             1883 mm/khugepaged.c 			node = khugepaged_find_target_node();
node             1884 mm/khugepaged.c 			collapse_file(mm, file, start, hpage, node);
node              121 mm/kmemleak.c  	struct hlist_node node;
node              475 mm/kmemleak.c  	hlist_for_each_entry_safe(area, tmp, &object->area_list, node) {
node              476 mm/kmemleak.c  		hlist_del(&area->node);
node              817 mm/kmemleak.c  	INIT_HLIST_NODE(&area->node);
node              821 mm/kmemleak.c  	hlist_add_head(&area->node, &object->area_list);
node             1352 mm/kmemleak.c  		hlist_for_each_entry(area, &object->area_list, node)
node              157 mm/ksm.c       		struct rb_node node;	/* when node of stable tree */
node              207 mm/ksm.c       		struct rb_node node;	/* when node of unstable tree */
node              371 mm/ksm.c       		rb_erase(&dup->node, root_stable_tree + NUMA(dup->nid));
node              609 mm/ksm.c       		rb_replace_node(&dup->node, &chain->node, root);
node              626 mm/ksm.c       	rb_erase(&chain->node, root);
node              810 mm/ksm.c       			rb_erase(&rmap_item->node,
node              947 mm/ksm.c       						struct stable_node, node);
node             1428 mm/ksm.c       			rb_replace_node(&stable_node->node, &found->node,
node             1578 mm/ksm.c       		stable_node = rb_entry(*new, struct stable_node, node);
node             1703 mm/ksm.c       	rb_link_node(&page_node->node, parent, new);
node             1704 mm/ksm.c       	rb_insert_color(&page_node->node, root);
node             1729 mm/ksm.c       			rb_replace_node(&stable_node_dup->node,
node             1730 mm/ksm.c       					&page_node->node,
node             1737 mm/ksm.c       			rb_erase(&stable_node_dup->node, root);
node             1825 mm/ksm.c       		stable_node = rb_entry(*new, struct stable_node, node);
node             1890 mm/ksm.c       		rb_link_node(&stable_node_dup->node, parent, new);
node             1891 mm/ksm.c       		rb_insert_color(&stable_node_dup->node, root);
node             1942 mm/ksm.c       		tree_rmap_item = rb_entry(*new, struct rmap_item, node);
node             1982 mm/ksm.c       	rb_link_node(&rmap_item->node, parent, new);
node             1983 mm/ksm.c       	rb_insert_color(&rmap_item->node, root);
node             2769 mm/ksm.c       	struct rb_node *node;
node             2773 mm/ksm.c       		node = rb_first(root_stable_tree + nid);
node             2774 mm/ksm.c       		while (node) {
node             2775 mm/ksm.c       			stable_node = rb_entry(node, struct stable_node, node);
node             2780 mm/ksm.c       				node = rb_first(root_stable_tree + nid);
node             2782 mm/ksm.c       				node = rb_next(node);
node              128 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              152 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              187 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              204 mm/list_lru.c  	nlru = &lru->node[nid];
node              272 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              288 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              309 mm/list_lru.c  			struct list_lru_node *nlru = &lru->node[nid];
node              458 mm/list_lru.c  		if (memcg_init_list_lru_node(&lru->node[i]))
node              464 mm/list_lru.c  		if (!lru->node[i].memcg_lrus)
node              466 mm/list_lru.c  		memcg_destroy_list_lru_node(&lru->node[i]);
node              479 mm/list_lru.c  		memcg_destroy_list_lru_node(&lru->node[i]);
node              491 mm/list_lru.c  		if (memcg_update_list_lru_node(&lru->node[i],
node              498 mm/list_lru.c  		if (!lru->node[i].memcg_lrus)
node              501 mm/list_lru.c  		memcg_cancel_update_list_lru_node(&lru->node[i],
node              516 mm/list_lru.c  		memcg_cancel_update_list_lru_node(&lru->node[i],
node              544 mm/list_lru.c  	struct list_lru_node *nlru = &lru->node[nid];
node              614 mm/list_lru.c  	lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL);
node              615 mm/list_lru.c  	if (!lru->node)
node              619 mm/list_lru.c  		spin_lock_init(&lru->node[i].lock);
node              621 mm/list_lru.c  			lockdep_set_class(&lru->node[i].lock, key);
node              622 mm/list_lru.c  		init_one_lru(&lru->node[i].lru);
node              627 mm/list_lru.c  		kfree(lru->node);
node              629 mm/list_lru.c  		lru->node = NULL;
node              643 mm/list_lru.c  	if (!lru->node)
node              651 mm/list_lru.c  	kfree(lru->node);
node              652 mm/list_lru.c  	lru->node = NULL;
node             1685 mm/memcontrol.c 	int node;
node             1688 mm/memcontrol.c 	node = memcg->last_scanned_node;
node             1690 mm/memcontrol.c 	node = next_node_in(node, memcg->scan_nodes);
node             1696 mm/memcontrol.c 	if (unlikely(node == MAX_NUMNODES))
node             1697 mm/memcontrol.c 		node = numa_node_id();
node             1699 mm/memcontrol.c 	memcg->last_scanned_node = node;
node             1700 mm/memcontrol.c 	return node;
node             3477 mm/memcontrol.c 	int node, cpu, i;
node             3487 mm/memcontrol.c 	for_each_node(node) {
node             3488 mm/memcontrol.c 		struct mem_cgroup_per_node *pn = memcg->nodeinfo[node];
node             3499 mm/memcontrol.c 		for (pi = pn; pi; pi = parent_nodeinfo(pi, node))
node             5023 mm/memcontrol.c static int alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node)
node             5026 mm/memcontrol.c 	int tmp = node;
node             5035 mm/memcontrol.c 	if (!node_state(node, N_NORMAL_MEMORY))
node             5059 mm/memcontrol.c 	memcg->nodeinfo[node] = pn;
node             5063 mm/memcontrol.c static void free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node)
node             5065 mm/memcontrol.c 	struct mem_cgroup_per_node *pn = memcg->nodeinfo[node];
node             5077 mm/memcontrol.c 	int node;
node             5079 mm/memcontrol.c 	for_each_node(node)
node             5080 mm/memcontrol.c 		free_mem_cgroup_per_node_info(memcg, node);
node             5102 mm/memcontrol.c 	int node;
node             5129 mm/memcontrol.c 	for_each_node(node)
node             5130 mm/memcontrol.c 		if (alloc_mem_cgroup_per_node_info(memcg, node))
node             6970 mm/memcontrol.c 	int cpu, node;
node             6990 mm/memcontrol.c 	for_each_node(node) {
node             6994 mm/memcontrol.c 				    node_online(node) ? node : NUMA_NO_NODE);
node             6999 mm/memcontrol.c 		soft_limit_tree.rb_tree_per_node[node] = rtpn;
node              228 mm/memory_hotplug.c 	int node = pgdat->node_id;
node              235 mm/memory_hotplug.c 		get_page_bootmem(node, page, NODE_INFO);
node              248 mm/memory_hotplug.c 		if (pfn_valid(pfn) && (early_pfn_to_nid(pfn) == node))
node              663 mm/memory_hotplug.c static void node_states_set_node(int node, struct memory_notify *arg)
node              666 mm/memory_hotplug.c 		node_set_state(node, N_NORMAL_MEMORY);
node              669 mm/memory_hotplug.c 		node_set_state(node, N_HIGH_MEMORY);
node              672 mm/memory_hotplug.c 		node_set_state(node, N_MEMORY);
node             1475 mm/memory_hotplug.c static void node_states_clear_node(int node, struct memory_notify *arg)
node             1478 mm/memory_hotplug.c 		node_clear_state(node, N_NORMAL_MEMORY);
node             1481 mm/memory_hotplug.c 		node_clear_state(node, N_HIGH_MEMORY);
node             1484 mm/memory_hotplug.c 		node_clear_state(node, N_MEMORY);
node             1492 mm/memory_hotplug.c 	int ret, node, nr_isolate_pageblock;
node             1511 mm/memory_hotplug.c 	node = zone_to_nid(zone);
node             1601 mm/memory_hotplug.c 	node_states_clear_node(node, &arg);
node             1603 mm/memory_hotplug.c 		kswapd_stop(node);
node             1604 mm/memory_hotplug.c 		kcompactd_stop(node);
node              133 mm/mempolicy.c 	int node;
node              138 mm/mempolicy.c 	node = numa_node_id();
node              139 mm/mempolicy.c 	if (node != NUMA_NO_NODE) {
node              140 mm/mempolicy.c 		pol = &preferred_node_policy[node];
node              324 mm/mempolicy.c 		int node = first_node(pol->w.user_nodemask);
node              326 mm/mempolicy.c 		if (node_isset(node, *nodes)) {
node              327 mm/mempolicy.c 			pol->v.preferred_node = node;
node             1000 mm/mempolicy.c struct page *alloc_new_node_page(struct page *page, unsigned long node)
node             1004 mm/mempolicy.c 					node);
node             1008 mm/mempolicy.c 		thp = alloc_pages_node(node,
node             1016 mm/mempolicy.c 		return __alloc_pages_node(node, GFP_HIGHUSER_MOVABLE |
node             1844 mm/mempolicy.c 	int node = numa_mem_id();
node             1847 mm/mempolicy.c 		return node;
node             1851 mm/mempolicy.c 		return node;
node             1872 mm/mempolicy.c 		zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK];
node             1875 mm/mempolicy.c 		return z->zone ? zone_to_nid(z->zone) : node;
node             2099 mm/mempolicy.c 		unsigned long addr, int node, bool hugepage)
node             2118 mm/mempolicy.c 		int hpage_node = node;
node             2154 mm/mempolicy.c 	preferred_nid = policy_node(gfp, pol, node);
node             2475 mm/mempolicy.c static void sp_node_init(struct sp_node *node, unsigned long start,
node             2478 mm/mempolicy.c 	node->start = start;
node             2479 mm/mempolicy.c 	node->end = end;
node             2480 mm/mempolicy.c 	node->policy = pol;
node             1502 mm/migrate.c   		struct list_head *pagelist, int node)
node             1509 mm/migrate.c   	err = migrate_pages(pagelist, alloc_new_node_page, NULL, node,
node             1526 mm/migrate.c   		int node, struct list_head *pagelist, bool migrate_all)
node             1552 mm/migrate.c   	if (page_to_nid(page) == node)
node             1610 mm/migrate.c   		int node;
node             1615 mm/migrate.c   		if (get_user(node, nodes + i))
node             1620 mm/migrate.c   		if (node < 0 || node >= MAX_NUMNODES)
node             1622 mm/migrate.c   		if (!node_state(node, N_MEMORY))
node             1626 mm/migrate.c   		if (!node_isset(node, task_nodes))
node             1630 mm/migrate.c   			current_node = node;
node             1632 mm/migrate.c   		} else if (node != current_node) {
node             1651 mm/migrate.c   			current_node = node;
node             1970 mm/migrate.c   			   int node)
node             1972 mm/migrate.c   	pg_data_t *pgdat = NODE_DATA(node);
node             1998 mm/migrate.c   				     NULL, node, MIGRATE_ASYNC,
node             2028 mm/migrate.c   				struct page *page, int node)
node             2031 mm/migrate.c   	pg_data_t *pgdat = NODE_DATA(node);
node             2037 mm/migrate.c   	new_page = alloc_pages_node(node,
node              153 mm/nommu.c     void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags)
node              256 mm/nommu.c     void *vmalloc_node(unsigned long size, int node)
node              274 mm/nommu.c     void *vzalloc_node(unsigned long size, int node)
node              343 mm/nommu.c     void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot)
node              307 mm/page-writeback.c 	int node;
node              311 mm/page-writeback.c 	for_each_node_state(node, N_HIGH_MEMORY) {
node              319 mm/page-writeback.c 			z = &NODE_DATA(node)->node_zones[i];
node             1472 mm/page_alloc.c static inline bool __meminit early_pfn_in_nid(unsigned long pfn, int node)
node             1477 mm/page_alloc.c 	if (nid >= 0 && nid != node)
node             1483 mm/page_alloc.c static inline bool __meminit early_pfn_in_nid(unsigned long pfn, int node)
node             5535 mm/page_alloc.c static int find_next_best_node(int node, nodemask_t *used_node_mask)
node             5543 mm/page_alloc.c 	if (!node_isset(node, *used_node_mask)) {
node             5544 mm/page_alloc.c 		node_set(node, *used_node_mask);
node             5545 mm/page_alloc.c 		return node;
node             5555 mm/page_alloc.c 		val = node_distance(node, n);
node             5558 mm/page_alloc.c 		val += (n < node);
node             5598 mm/page_alloc.c 		pg_data_t *node = NODE_DATA(node_order[i]);
node             5600 mm/page_alloc.c 		nr_zones = build_zonerefs_node(node, zonerefs);
node             5632 mm/page_alloc.c 	int node, load, nr_nodes = 0;
node             5643 mm/page_alloc.c 	while ((node = find_next_best_node(local_node, &used_mask)) >= 0) {
node             5649 mm/page_alloc.c 		if (node_distance(local_node, node) !=
node             5651 mm/page_alloc.c 			node_load[node] = load;
node             5653 mm/page_alloc.c 		node_order[nr_nodes++] = node;
node             5654 mm/page_alloc.c 		prev_node = node;
node             5669 mm/page_alloc.c int local_memory_node(int node)
node             5673 mm/page_alloc.c 	z = first_zones_zonelist(node_zonelist(node, GFP_KERNEL),
node             5686 mm/page_alloc.c 	int node, local_node;
node             5704 mm/page_alloc.c 	for (node = local_node + 1; node < MAX_NUMNODES; node++) {
node             5705 mm/page_alloc.c 		if (!node_online(node))
node             5707 mm/page_alloc.c 		nr_zones = build_zonerefs_node(NODE_DATA(node), zonerefs);
node             5710 mm/page_alloc.c 	for (node = 0; node < local_node; node++) {
node             5711 mm/page_alloc.c 		if (!node_online(node))
node             5713 mm/page_alloc.c 		nr_zones = build_zonerefs_node(NODE_DATA(node), zonerefs);
node              212 mm/slab.c      			int node, struct list_head *list);
node              492 mm/slab.c      	int node = __this_cpu_read(slab_reap_node);
node              494 mm/slab.c      	node = next_node_in(node, node_online_map);
node              495 mm/slab.c      	__this_cpu_write(slab_reap_node, node);
node              532 mm/slab.c      static struct array_cache *alloc_arraycache(int node, int entries,
node              538 mm/slab.c      	ac = kmalloc_node(memsize, gfp, node);
node              596 mm/slab.c      static inline struct alien_cache **alloc_alien_cache(int node,
node              633 mm/slab.c      static struct alien_cache *__alloc_alien_cache(int node, int entries,
node              639 mm/slab.c      	alc = kmalloc_node(memsize, gfp, node);
node              648 mm/slab.c      static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
node              655 mm/slab.c      	alc_ptr = kcalloc_node(nr_node_ids, sizeof(void *), gfp, node);
node              660 mm/slab.c      		if (i == node || !node_online(i))
node              662 mm/slab.c      		alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp);
node              685 mm/slab.c      				struct array_cache *ac, int node,
node              688 mm/slab.c      	struct kmem_cache_node *n = get_node(cachep, node);
node              700 mm/slab.c      		free_block(cachep, ac->entry, ac->avail, node, list);
node              711 mm/slab.c      	int node = __this_cpu_read(slab_reap_node);
node              714 mm/slab.c      		struct alien_cache *alc = n->alien[node];
node              722 mm/slab.c      				__drain_alien_cache(cachep, ac, node, &list);
node              753 mm/slab.c      				int node, int page_node)
node              760 mm/slab.c      	n = get_node(cachep, node);
node              786 mm/slab.c      	int node = numa_mem_id();
node              791 mm/slab.c      	if (likely(node == page_node))
node              794 mm/slab.c      	return __cache_free_alien(cachep, objp, node, page_node);
node              807 mm/slab.c      static int init_cache_node(struct kmem_cache *cachep, int node, gfp_t gfp)
node              816 mm/slab.c      	n = get_node(cachep, node);
node              819 mm/slab.c      		n->free_limit = (1 + nr_cpus_node(node)) * cachep->batchcount +
node              826 mm/slab.c      	n = kmalloc_node(sizeof(struct kmem_cache_node), gfp, node);
node              835 mm/slab.c      		(1 + nr_cpus_node(node)) * cachep->batchcount + cachep->num;
node              842 mm/slab.c      	cachep->node[node] = n;
node              857 mm/slab.c      static int init_cache_node_node(int node)
node              863 mm/slab.c      		ret = init_cache_node(cachep, node, GFP_KERNEL);
node              873 mm/slab.c      				int node, gfp_t gfp, bool force_change)
node              883 mm/slab.c      		new_alien = alloc_alien_cache(node, cachep->limit, gfp);
node              889 mm/slab.c      		new_shared = alloc_arraycache(node,
node              895 mm/slab.c      	ret = init_cache_node(cachep, node, gfp);
node              899 mm/slab.c      	n = get_node(cachep, node);
node              903 mm/slab.c      				n->shared->avail, node, &list);
node              944 mm/slab.c      	int node = cpu_to_mem(cpu);
node              945 mm/slab.c      	const struct cpumask *mask = cpumask_of_node(node);
node              953 mm/slab.c      		n = get_node(cachep, node);
node              964 mm/slab.c      		free_block(cachep, nc->entry, nc->avail, node, &list);
node              975 mm/slab.c      				   shared->avail, node, &list);
node              999 mm/slab.c      		n = get_node(cachep, node);
node             1009 mm/slab.c      	int node = cpu_to_mem(cpu);
node             1018 mm/slab.c      	err = init_cache_node_node(node);
node             1027 mm/slab.c      		err = setup_kmem_cache_node(cachep, node, GFP_KERNEL, false);
node             1095 mm/slab.c      static int __meminit drain_cache_node_node(int node)
node             1103 mm/slab.c      		n = get_node(cachep, node);
node             1169 mm/slab.c      	cachep->node[nodeid] = ptr;
node             1178 mm/slab.c      	int node;
node             1180 mm/slab.c      	for_each_online_node(node) {
node             1181 mm/slab.c      		cachep->node[node] = &init_kmem_cache_node[index + node];
node             1182 mm/slab.c      		cachep->node[node]->next_reap = jiffies +
node             1238 mm/slab.c      		offsetof(struct kmem_cache, node) +
node             1322 mm/slab.c      	int node;
node             1334 mm/slab.c      	for_each_kmem_cache_node(cachep, node, n) {
node             1344 mm/slab.c      			node, total_slabs - free_slabs, total_slabs,
node             1758 mm/slab.c      		int node;
node             1760 mm/slab.c      		for_each_online_node(node) {
node             1761 mm/slab.c      			cachep->node[node] = kmalloc_node(
node             1762 mm/slab.c      				sizeof(struct kmem_cache_node), gfp, node);
node             1763 mm/slab.c      			BUG_ON(!cachep->node[node]);
node             1764 mm/slab.c      			kmem_cache_node_init(cachep->node[node]);
node             1768 mm/slab.c      	cachep->node[numa_mem_id()]->next_reap =
node             2103 mm/slab.c      static void check_spinlock_acquired_node(struct kmem_cache *cachep, int node)
node             2107 mm/slab.c      	assert_spin_locked(&get_node(cachep, node)->list_lock);
node             2120 mm/slab.c      				int node, bool free_all, struct list_head *list)
node             2131 mm/slab.c      	free_block(cachep, ac->entry, tofree, node, list);
node             2140 mm/slab.c      	int node = numa_mem_id();
node             2146 mm/slab.c      	n = get_node(cachep, node);
node             2148 mm/slab.c      	free_block(cachep, ac->entry, ac->avail, node, &list);
node             2157 mm/slab.c      	int node;
node             2162 mm/slab.c      	for_each_kmem_cache_node(cachep, node, n)
node             2166 mm/slab.c      	for_each_kmem_cache_node(cachep, node, n) {
node             2168 mm/slab.c      		drain_array_locked(cachep, n->shared, node, true, &list);
node             2217 mm/slab.c      	int node;
node             2220 mm/slab.c      	for_each_kmem_cache_node(s, node, n)
node             2230 mm/slab.c      	int node;
node             2236 mm/slab.c      	for_each_kmem_cache_node(cachep, node, n) {
node             2275 mm/slab.c      		cachep->node[i] = NULL;
node             2910 mm/slab.c      	int node;
node             2915 mm/slab.c      	node = numa_mem_id();
node             2927 mm/slab.c      	n = get_node(cachep, node);
node             2971 mm/slab.c      		page = cache_grow_begin(cachep, gfp_exact_node(flags), node);
node             3328 mm/slab.c      			int nr_objects, int node, struct list_head *list)
node             3331 mm/slab.c      	struct kmem_cache_node *n = get_node(cachep, node);
node             3344 mm/slab.c      		check_spinlock_acquired_node(cachep, node);
node             3375 mm/slab.c      	int node = numa_mem_id();
node             3381 mm/slab.c      	n = get_node(cachep, node);
node             3396 mm/slab.c      	free_block(cachep, ac->entry, batchcount, node, &list);
node             3604 mm/slab.c      __do_kmalloc_node(size_t size, gfp_t flags, int node, unsigned long caller)
node             3614 mm/slab.c      	ret = kmem_cache_alloc_node_trace(cachep, flags, node, size);
node             3620 mm/slab.c      void *__kmalloc_node(size_t size, gfp_t flags, int node)
node             3622 mm/slab.c      	return __do_kmalloc_node(size, flags, node, _RET_IP_);
node             3627 mm/slab.c      		int node, unsigned long caller)
node             3629 mm/slab.c      	return __do_kmalloc_node(size, flags, node, caller);
node             3767 mm/slab.c      	int node;
node             3770 mm/slab.c      	for_each_online_node(node) {
node             3771 mm/slab.c      		ret = setup_kmem_cache_node(cachep, node, gfp, true);
node             3782 mm/slab.c      		node--;
node             3783 mm/slab.c      		while (node >= 0) {
node             3784 mm/slab.c      			n = get_node(cachep, node);
node             3789 mm/slab.c      				cachep->node[node] = NULL;
node             3791 mm/slab.c      			node--;
node             3827 mm/slab.c      		int node;
node             3831 mm/slab.c      		node = cpu_to_mem(cpu);
node             3832 mm/slab.c      		n = get_node(cachep, node);
node             3834 mm/slab.c      		free_block(cachep, ac->entry, ac->avail, node, &list);
node             3945 mm/slab.c      			 struct array_cache *ac, int node)
node             3961 mm/slab.c      	drain_array_locked(cachep, ac, node, false, &list);
node             3983 mm/slab.c      	int node = numa_mem_id();
node             3998 mm/slab.c      		n = get_node(searchp, node);
node             4002 mm/slab.c      		drain_array(searchp, n, cpu_cache_get(searchp), node);
node             4013 mm/slab.c      		drain_array(searchp, n, n->shared, node);
node             4041 mm/slab.c      	int node;
node             4044 mm/slab.c      	for_each_kmem_cache_node(cachep, node, n) {
node              628 mm/slab.h      static inline struct kmem_cache_node *get_node(struct kmem_cache *s, int node)
node              630 mm/slab.h      	return s->node[node];
node              191 mm/slob.c      static void *slob_new_pages(gfp_t gfp, int order, int node)
node              196 mm/slob.c      	if (node != NUMA_NO_NODE)
node              197 mm/slob.c      		page = __alloc_pages_node(node, gfp, order);
node              301 mm/slob.c      static void *slob_alloc(size_t size, gfp_t gfp, int align, int node,
node              326 mm/slob.c      		if (node != NUMA_NO_NODE && page_to_nid(sp) != node)
node              358 mm/slob.c      		b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node);
node              469 mm/slob.c      __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller)
node              493 mm/slob.c      		m = slob_alloc(size + minalign, gfp, align, node, minalign);
node              501 mm/slob.c      				   size, size + minalign, gfp, node);
node              507 mm/slob.c      		ret = slob_new_pages(gfp, order, node);
node              510 mm/slob.c      				   size, PAGE_SIZE << order, gfp, node);
node              530 mm/slob.c      					int node, unsigned long caller)
node              532 mm/slob.c      	return __do_kmalloc_node(size, gfp, node, caller);
node              592 mm/slob.c      static void *slob_alloc_node(struct kmem_cache *c, gfp_t flags, int node)
node              602 mm/slob.c      		b = slob_alloc(c->size, flags, c->align, node, 0);
node              605 mm/slob.c      					    flags, node);
node              607 mm/slob.c      		b = slob_new_pages(flags, get_order(c->size), node);
node              610 mm/slob.c      					    flags, node);
node              629 mm/slob.c      void *__kmalloc_node(size_t size, gfp_t gfp, int node)
node              631 mm/slob.c      	return __do_kmalloc_node(size, gfp, node, _RET_IP_);
node              635 mm/slob.c      void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node)
node              637 mm/slob.c      	return slob_alloc_node(cachep, gfp, node);
node             1031 mm/slub.c      static inline unsigned long slabs_node(struct kmem_cache *s, int node)
node             1033 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node             1043 mm/slub.c      static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects)
node             1045 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node             1058 mm/slub.c      static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects)
node             1060 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node             1373 mm/slub.c      static inline unsigned long slabs_node(struct kmem_cache *s, int node)
node             1377 mm/slub.c      static inline void inc_slabs_node(struct kmem_cache *s, int node,
node             1379 mm/slub.c      static inline void dec_slabs_node(struct kmem_cache *s, int node,
node             1489 mm/slub.c      		gfp_t flags, int node, struct kmem_cache_order_objects oo)
node             1494 mm/slub.c      	if (node == NUMA_NO_NODE)
node             1497 mm/slub.c      		page = __alloc_pages_node(node, flags, order);
node             1616 mm/slub.c      static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node)
node             1640 mm/slub.c      	page = alloc_slab_page(s, alloc_gfp, node, oo);
node             1648 mm/slub.c      		page = alloc_slab_page(s, alloc_gfp, node, oo);
node             1696 mm/slub.c      static struct page *new_slab(struct kmem_cache *s, gfp_t flags, int node)
node             1707 mm/slub.c      		flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node);
node             1948 mm/slub.c      static void *get_partial(struct kmem_cache *s, gfp_t flags, int node,
node             1952 mm/slub.c      	int searchnode = node;
node             1954 mm/slub.c      	if (node == NUMA_NO_NODE)
node             1958 mm/slub.c      	if (object || node != NUMA_NO_NODE)
node             2367 mm/slub.c      static inline int node_match(struct page *page, int node)
node             2370 mm/slub.c      	if (node != NUMA_NO_NODE && page_to_nid(page) != node)
node             2410 mm/slub.c      	int node;
node             2426 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             2436 mm/slub.c      			node, nr_slabs, nr_objs, nr_free);
node             2442 mm/slub.c      			int node, struct kmem_cache_cpu **pc)
node             2450 mm/slub.c      	freelist = get_partial(s, flags, node, c);
node             2455 mm/slub.c      	page = new_slab(s, flags, node);
node             2537 mm/slub.c      static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
node             2549 mm/slub.c      		if (unlikely(node != NUMA_NO_NODE &&
node             2550 mm/slub.c      			     !node_state(node, N_NORMAL_MEMORY)))
node             2551 mm/slub.c      			node = NUMA_NO_NODE;
node             2556 mm/slub.c      	if (unlikely(!node_match(page, node))) {
node             2561 mm/slub.c      		if (!node_state(node, N_NORMAL_MEMORY)) {
node             2562 mm/slub.c      			node = NUMA_NO_NODE;
node             2616 mm/slub.c      	freelist = new_slab_objects(s, gfpflags, node, &c);
node             2619 mm/slub.c      		slab_out_of_memory(s, gfpflags, node);
node             2640 mm/slub.c      static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
node             2656 mm/slub.c      	p = ___slab_alloc(s, gfpflags, node, addr, c);
node             2683 mm/slub.c      		gfp_t gfpflags, int node, unsigned long addr)
node             2729 mm/slub.c      	if (unlikely(!object || !node_match(page, node))) {
node             2730 mm/slub.c      		object = __slab_alloc(s, gfpflags, node, addr, c);
node             2800 mm/slub.c      void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node)
node             2802 mm/slub.c      	void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_);
node             2805 mm/slub.c      				    s->object_size, s->size, gfpflags, node);
node             2814 mm/slub.c      				    int node, size_t size)
node             2816 mm/slub.c      	void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_);
node             2819 mm/slub.c      			   size, s->size, gfpflags, node);
node             3385 mm/slub.c      static void early_kmem_cache_node_alloc(int node)
node             3392 mm/slub.c      	page = new_slab(kmem_cache_node, GFP_NOWAIT, node);
node             3395 mm/slub.c      	if (page_to_nid(page) != node) {
node             3396 mm/slub.c      		pr_err("SLUB: Unable to allocate memory from node %d\n", node);
node             3411 mm/slub.c      	kmem_cache_node->node[node] = n;
node             3413 mm/slub.c      	inc_slabs_node(kmem_cache_node, node, page->objects);
node             3424 mm/slub.c      	int node;
node             3427 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             3428 mm/slub.c      		s->node[node] = NULL;
node             3442 mm/slub.c      	int node;
node             3444 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             3448 mm/slub.c      			early_kmem_cache_node_alloc(node);
node             3452 mm/slub.c      						GFP_KERNEL, node);
node             3460 mm/slub.c      		s->node[node] = n;
node             3744 mm/slub.c      	int node;
node             3747 mm/slub.c      	for_each_kmem_cache_node(s, node, n)
node             3748 mm/slub.c      		if (n->nr_partial || slabs_node(s, node))
node             3758 mm/slub.c      	int node;
node             3763 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             3765 mm/slub.c      		if (n->nr_partial || slabs_node(s, node))
node             3828 mm/slub.c      static void *kmalloc_large_node(size_t size, gfp_t flags, int node)
node             3835 mm/slub.c      	page = alloc_pages_node(node, flags, order);
node             3845 mm/slub.c      void *__kmalloc_node(size_t size, gfp_t flags, int node)
node             3851 mm/slub.c      		ret = kmalloc_large_node(size, flags, node);
node             3855 mm/slub.c      				   flags, node);
node             3865 mm/slub.c      	ret = slab_alloc_node(s, flags, node, _RET_IP_);
node             3867 mm/slub.c      	trace_kmalloc_node(_RET_IP_, ret, size, s->size, flags, node);
node             3992 mm/slub.c      	int node;
node             4003 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             4045 mm/slub.c      		if (slabs_node(s, node))
node             4122 mm/slub.c      			s->node[offline_node] = NULL;
node             4162 mm/slub.c      		s->node[nid] = n;
node             4213 mm/slub.c      	int node;
node             4225 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             4262 mm/slub.c      			offsetof(struct kmem_cache, node) +
node             4363 mm/slub.c      					int node, unsigned long caller)
node             4369 mm/slub.c      		ret = kmalloc_large_node(size, gfpflags, node);
node             4373 mm/slub.c      				   gfpflags, node);
node             4383 mm/slub.c      	ret = slab_alloc_node(s, gfpflags, node, caller);
node             4386 mm/slub.c      	trace_kmalloc_node(caller, ret, size, s->size, gfpflags, node);
node             4475 mm/slub.c      	int node;
node             4484 mm/slub.c      	for_each_kmem_cache_node(s, node, n)
node             4636 mm/slub.c      	int node;
node             4648 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node             4810 mm/slub.c      	int node;
node             4824 mm/slub.c      			int node;
node             4831 mm/slub.c      			node = page_to_nid(page);
node             4840 mm/slub.c      			nodes[node] += x;
node             4844 mm/slub.c      				node = page_to_nid(page);
node             4852 mm/slub.c      				nodes[node] += x;
node             4872 mm/slub.c      		for_each_kmem_cache_node(s, node, n) {
node             4882 mm/slub.c      			nodes[node] += x;
node             4890 mm/slub.c      		for_each_kmem_cache_node(s, node, n) {
node             4898 mm/slub.c      			nodes[node] += x;
node             4903 mm/slub.c      	for (node = 0; node < nr_node_ids; node++)
node             4904 mm/slub.c      		if (nodes[node])
node             4906 mm/slub.c      					node, nodes[node]);
node             4915 mm/slub.c      	int node;
node             4918 mm/slub.c      	for_each_kmem_cache_node(s, node, n)
node             5923 mm/slub.c      	int node;
node             5926 mm/slub.c      	for_each_kmem_cache_node(s, node, n) {
node               40 mm/sparse-vmemmap.c static void * __ref __earlyonly_bootmem_alloc(int node,
node               46 mm/sparse-vmemmap.c 					       MEMBLOCK_ALLOC_ACCESSIBLE, node);
node               49 mm/sparse-vmemmap.c void * __meminit vmemmap_alloc_block(unsigned long size, int node)
node               58 mm/sparse-vmemmap.c 		page = alloc_pages_node(node, gfp_mask, order);
node               69 mm/sparse-vmemmap.c 		return __earlyonly_bootmem_alloc(node, size, size,
node               74 mm/sparse-vmemmap.c void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node)
node               79 mm/sparse-vmemmap.c 		ptr = vmemmap_alloc_block(size, node);
node              132 mm/sparse-vmemmap.c void __meminit vmemmap_verify(pte_t *pte, int node,
node              138 mm/sparse-vmemmap.c 	if (node_distance(actual_node, node) > LOCAL_DISTANCE)
node              143 mm/sparse-vmemmap.c pte_t * __meminit vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node)
node              148 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block_buf(PAGE_SIZE, node);
node              157 mm/sparse-vmemmap.c static void * __meminit vmemmap_alloc_block_zero(unsigned long size, int node)
node              159 mm/sparse-vmemmap.c 	void *p = vmemmap_alloc_block(size, node);
node              168 mm/sparse-vmemmap.c pmd_t * __meminit vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node)
node              172 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node);
node              180 mm/sparse-vmemmap.c pud_t * __meminit vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node)
node              184 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node);
node              192 mm/sparse-vmemmap.c p4d_t * __meminit vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node)
node              196 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node);
node              204 mm/sparse-vmemmap.c pgd_t * __meminit vmemmap_pgd_populate(unsigned long addr, int node)
node              208 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block_zero(PAGE_SIZE, node);
node              217 mm/sparse-vmemmap.c 					 unsigned long end, int node)
node              227 mm/sparse-vmemmap.c 		pgd = vmemmap_pgd_populate(addr, node);
node              230 mm/sparse-vmemmap.c 		p4d = vmemmap_p4d_populate(pgd, addr, node);
node              233 mm/sparse-vmemmap.c 		pud = vmemmap_pud_populate(p4d, addr, node);
node              236 mm/sparse-vmemmap.c 		pmd = vmemmap_pmd_populate(pud, addr, node);
node              239 mm/sparse-vmemmap.c 		pte = vmemmap_pte_populate(pmd, addr, node);
node              242 mm/sparse-vmemmap.c 		vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
node              998 mm/swapfile.c  	int node;
node             1018 mm/swapfile.c  	node = numa_node_id();
node             1019 mm/swapfile.c  	plist_for_each_entry_safe(si, next, &swap_avail_heads[node], avail_lists[node]) {
node             1021 mm/swapfile.c  		plist_requeue(&si->avail_lists[node], &swap_avail_heads[node]);
node             1026 mm/swapfile.c  			if (plist_node_empty(&si->avail_lists[node])) {
node             1065 mm/swapfile.c  		if (plist_node_empty(&next->avail_lists[node]))
node             3740 mm/swapfile.c  void mem_cgroup_throttle_swaprate(struct mem_cgroup *memcg, int node,
node             3758 mm/swapfile.c  	plist_for_each_entry_safe(si, next, &swap_avail_heads[node],
node             3759 mm/swapfile.c  				  avail_lists[node]) {
node              538 mm/util.c      void *kvmalloc_node(size_t size, gfp_t flags, int node)
node              548 mm/util.c      		return kmalloc_node(size, flags, node);
node              564 mm/util.c      	ret = kmalloc_node(size, kmalloc_flags, node);
node              573 mm/util.c      	return __vmalloc_node_flags_caller(size, node, flags,
node              381 mm/vmalloc.c   get_subtree_max_size(struct rb_node *node)
node              385 mm/vmalloc.c   	va = rb_entry_safe(node, struct vmap_area, rb_node);
node              560 mm/vmalloc.c   	struct rb_node *node;
node              569 mm/vmalloc.c   	node = n;
node              571 mm/vmalloc.c   	while (node) {
node              572 mm/vmalloc.c   		va = rb_entry(node, struct vmap_area, rb_node);
node              574 mm/vmalloc.c   		if (get_subtree_max_size(node->rb_left) == size) {
node              575 mm/vmalloc.c   			node = node->rb_left;
node              582 mm/vmalloc.c   			node = node->rb_right;
node              627 mm/vmalloc.c   	struct rb_node *node = &va->rb_node;
node              630 mm/vmalloc.c   	while (node) {
node              631 mm/vmalloc.c   		va = rb_entry(node, struct vmap_area, rb_node);
node              644 mm/vmalloc.c   		node = rb_parent(&va->rb_node);
node              793 mm/vmalloc.c   	struct rb_node *node;
node              797 mm/vmalloc.c   	node = free_vmap_area_root.rb_node;
node              802 mm/vmalloc.c   	while (node) {
node              803 mm/vmalloc.c   		va = rb_entry(node, struct vmap_area, rb_node);
node              805 mm/vmalloc.c   		if (get_subtree_max_size(node->rb_left) >= length &&
node              807 mm/vmalloc.c   			node = node->rb_left;
node              817 mm/vmalloc.c   			if (get_subtree_max_size(node->rb_right) >= length) {
node              818 mm/vmalloc.c   				node = node->rb_right;
node              827 mm/vmalloc.c   			while ((node = rb_parent(node))) {
node              828 mm/vmalloc.c   				va = rb_entry(node, struct vmap_area, rb_node);
node              832 mm/vmalloc.c   				if (get_subtree_max_size(node->rb_right) >= length &&
node              834 mm/vmalloc.c   					node = node->rb_right;
node             1053 mm/vmalloc.c   				int node, gfp_t gfp_mask)
node             1069 mm/vmalloc.c   			gfp_mask & GFP_RECLAIM_MASK, node);
node             1095 mm/vmalloc.c   		pva = kmem_cache_alloc_node(vmap_area_cachep, GFP_KERNEL, node);
node             1463 mm/vmalloc.c   	int node, err;
node             1466 mm/vmalloc.c   	node = numa_node_id();
node             1469 mm/vmalloc.c   			gfp_mask & GFP_RECLAIM_MASK, node);
node             1475 mm/vmalloc.c   					node, gfp_mask);
node             1778 mm/vmalloc.c   void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot)
node             1792 mm/vmalloc.c   				VMALLOC_START, VMALLOC_END, node, GFP_KERNEL);
node             2043 mm/vmalloc.c   		unsigned long end, int node, gfp_t gfp_mask, const void *caller)
node             2057 mm/vmalloc.c   	area = kzalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node);
node             2064 mm/vmalloc.c   	va = alloc_vmap_area(size, align, start, end, node, gfp_mask);
node             2397 mm/vmalloc.c   			    int node, const void *caller);
node             2399 mm/vmalloc.c   				 pgprot_t prot, int node)
node             2415 mm/vmalloc.c   				PAGE_KERNEL, node, area->caller);
node             2417 mm/vmalloc.c   		pages = kmalloc_node(array_size, nested_gfp, node);
node             2432 mm/vmalloc.c   		if (node == NUMA_NO_NODE)
node             2435 mm/vmalloc.c   			page = alloc_pages_node(node, alloc_mask|highmem_mask, 0);
node             2481 mm/vmalloc.c   			pgprot_t prot, unsigned long vm_flags, int node,
node             2493 mm/vmalloc.c   				vm_flags, start, end, node, gfp_mask, caller);
node             2497 mm/vmalloc.c   	addr = __vmalloc_area_node(area, gfp_mask, prot, node);
node             2550 mm/vmalloc.c   			    int node, const void *caller)
node             2553 mm/vmalloc.c   				gfp_mask, prot, 0, node, caller);
node             2564 mm/vmalloc.c   					int node, gfp_t flags)
node             2567 mm/vmalloc.c   					node, __builtin_return_address(0));
node             2571 mm/vmalloc.c   void *__vmalloc_node_flags_caller(unsigned long size, int node, gfp_t flags,
node             2574 mm/vmalloc.c   	return __vmalloc_node(size, 1, flags, PAGE_KERNEL, node, caller);
node             2647 mm/vmalloc.c   void *vmalloc_node(unsigned long size, int node)
node             2650 mm/vmalloc.c   					node, __builtin_return_address(0));
node             2668 mm/vmalloc.c   void *vzalloc_node(unsigned long size, int node)
node             2670 mm/vmalloc.c   	return __vmalloc_node_flags(size, node,
node              155 mm/vmpressure.c 	struct list_head node;
node              166 mm/vmpressure.c 	list_for_each_entry(ev, &vmpr->events, node) {
node              406 mm/vmpressure.c 	list_add(&ev->node, &vmpr->events);
node              432 mm/vmpressure.c 	list_for_each_entry(ev, &vmpr->events, node) {
node              435 mm/vmpressure.c 		list_del(&ev->node);
node              959 mm/vmstat.c    unsigned long sum_zone_node_page_state(int node,
node              962 mm/vmstat.c    	struct zone *zones = NODE_DATA(node)->node_zones;
node              976 mm/vmstat.c    unsigned long sum_zone_numa_state(int node,
node              979 mm/vmstat.c    	struct zone *zones = NODE_DATA(node)->node_zones;
node             1303 mm/vmstat.c    	loff_t node = *pos;
node             1306 mm/vmstat.c    	     pgdat && node;
node             1308 mm/vmstat.c    		--node;
node             1925 mm/vmstat.c    	int node;
node             1927 mm/vmstat.c    	for_each_online_node(node) {
node             1928 mm/vmstat.c    		if (cpumask_weight(cpumask_of_node(node)) > 0)
node             1929 mm/vmstat.c    			node_set_state(node, N_CPU);
node             1949 mm/vmstat.c    	int node;
node             1951 mm/vmstat.c    	node = cpu_to_node(cpu);
node             1954 mm/vmstat.c    	node_cpus = cpumask_of_node(node);
node             1958 mm/vmstat.c    	node_clear_state(node, N_CPU);
node              368 mm/workingset.c void workingset_update_node(struct xa_node *node)
node              380 mm/workingset.c 	if (node->count && node->count == node->nr_values) {
node              381 mm/workingset.c 		if (list_empty(&node->private_list)) {
node              382 mm/workingset.c 			list_lru_add(&shadow_nodes, &node->private_list);
node              383 mm/workingset.c 			__inc_lruvec_slab_state(node, WORKINGSET_NODES);
node              386 mm/workingset.c 		if (!list_empty(&node->private_list)) {
node              387 mm/workingset.c 			list_lru_del(&shadow_nodes, &node->private_list);
node              388 mm/workingset.c 			__dec_lruvec_slab_state(node, WORKINGSET_NODES);
node              454 mm/workingset.c 	struct xa_node *node = container_of(item, struct xa_node, private_list);
node              455 mm/workingset.c 	XA_STATE(xas, node->array, 0);
node              471 mm/workingset.c 	mapping = container_of(node->array, struct address_space, i_pages);
node              481 mm/workingset.c 	__dec_lruvec_slab_state(node, WORKINGSET_NODES);
node              490 mm/workingset.c 	if (WARN_ON_ONCE(!node->nr_values))
node              492 mm/workingset.c 	if (WARN_ON_ONCE(node->count != node->nr_values))
node              494 mm/workingset.c 	mapping->nrexceptional -= node->nr_values;
node              495 mm/workingset.c 	xas.xa_node = xa_parent_locked(&mapping->i_pages, node);
node              496 mm/workingset.c 	xas.xa_offset = node->offset;
node              497 mm/workingset.c 	xas.xa_shift = node->shift + XA_CHUNK_SHIFT;
node              504 mm/workingset.c 	__inc_lruvec_slab_state(node, WORKINGSET_NODERECLAIM);
node              127 mm/zswap.c     	struct hlist_node node;
node              269 mm/zswap.c     	struct rb_node *node = root->rb_node;
node              272 mm/zswap.c     	while (node) {
node              273 mm/zswap.c     		entry = rb_entry(node, struct zswap_entry, rbnode);
node              275 mm/zswap.c     			node = node->rb_left;
node              277 mm/zswap.c     			node = node->rb_right;
node              398 mm/zswap.c     static int zswap_cpu_comp_prepare(unsigned int cpu, struct hlist_node *node)
node              400 mm/zswap.c     	struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node);
node              416 mm/zswap.c     static int zswap_cpu_comp_dead(unsigned int cpu, struct hlist_node *node)
node              418 mm/zswap.c     	struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node);
node              544 mm/zswap.c     				       &pool->node);
node              611 mm/zswap.c     	cpuhp_state_remove_instance(CPUHP_MM_ZSWP_POOL_PREPARE, &pool->node);
node               26 net/6lowpan/nhc.c 						   node);
node               50 net/6lowpan/nhc.c 	rb_link_node(&nhc->node, parent, new);
node               51 net/6lowpan/nhc.c 	rb_insert_color(&nhc->node, &rb_root);
node               58 net/6lowpan/nhc.c 	rb_erase(&nhc->node, &rb_root);
node               63 net/6lowpan/nhc.c 	struct rb_node *node = rb_root.rb_node;
node               66 net/6lowpan/nhc.c 	while (node) {
node               67 net/6lowpan/nhc.c 		struct lowpan_nhc *nhc = rb_entry(node, struct lowpan_nhc,
node               68 net/6lowpan/nhc.c 						  node);
node               82 net/6lowpan/nhc.c 			node = node->rb_left;
node               84 net/6lowpan/nhc.c 			node = node->rb_right;
node               68 net/6lowpan/nhc.h 	struct rb_node	node;
node              155 net/802/garp.c 		attr = rb_entry(parent, struct garp_attr, node);
node              176 net/802/garp.c 		attr = rb_entry(parent, struct garp_attr, node);
node              195 net/802/garp.c 	rb_link_node(&attr->node, parent, p);
node              196 net/802/garp.c 	rb_insert_color(&attr->node, &app->gid);
node              202 net/802/garp.c 	rb_erase(&attr->node, &app->gid);
node              382 net/802/garp.c 	struct rb_node *node, *next;
node              385 net/802/garp.c 	for (node = rb_first(&app->gid);
node              386 net/802/garp.c 	     next = node ? rb_next(node) : NULL, node != NULL;
node              387 net/802/garp.c 	     node = next) {
node              388 net/802/garp.c 		attr = rb_entry(node, struct garp_attr, node);
node              244 net/802/mrp.c  		attr = rb_entry(parent, struct mrp_attr, node);
node              265 net/802/mrp.c  		attr = rb_entry(parent, struct mrp_attr, node);
node              284 net/802/mrp.c  	rb_link_node(&attr->node, parent, p);
node              285 net/802/mrp.c  	rb_insert_color(&attr->node, &app->mad);
node              291 net/802/mrp.c  	rb_erase(&attr->node, &app->mad);
node              567 net/802/mrp.c  	struct rb_node *node, *next;
node              570 net/802/mrp.c  	for (node = rb_first(&app->mad);
node              571 net/802/mrp.c  	     next = node ? rb_next(node) : NULL, node != NULL;
node              572 net/802/mrp.c  	     node = next) {
node              573 net/802/mrp.c  		attr = rb_entry(node, struct mrp_attr, node);
node               33 net/802/psnap.c 	list_for_each_entry_rcu(p, &snap_list, node) {
node              141 net/802/psnap.c 		list_add_rcu(&proto->node, &snap_list);
node              155 net/802/psnap.c 	list_del_rcu(&proto->node);
node              375 net/appletalk/ddp.c static struct atalk_iface *atalk_find_anynet(int node, struct net_device *dev)
node              382 net/appletalk/ddp.c 	if (node != ATADDR_BCAST &&
node              383 net/appletalk/ddp.c 	    iface->address.s_node != node &&
node              384 net/appletalk/ddp.c 	    node != ATADDR_ANYNODE)
node              394 net/appletalk/ddp.c static struct atalk_iface *atalk_find_interface(__be16 net, int node)
node              400 net/appletalk/ddp.c 		if ((node == ATADDR_BCAST ||
node              401 net/appletalk/ddp.c 		     node == ATADDR_ANYNODE ||
node              402 net/appletalk/ddp.c 		     iface->address.s_node == node) &&
node              408 net/appletalk/ddp.c 		if (node == ATADDR_ANYNODE && net != ATADDR_ANYNET &&
node              825 net/atm/lec.c  	struct hlist_node *node;
node              835 net/atm/lec.c  	struct hlist_node *e = state->node;
node              848 net/atm/lec.c  	state->node = e;
node              942 net/atm/lec.c  	state->node = SEQ_START_TOKEN;
node              979 net/atm/lec.c  		struct lec_arp_table *entry = hlist_entry(state->node,
node              102 net/batman-adv/bridge_loop_avoidance.c static bool batadv_compare_backbone_gw(const struct hlist_node *node,
node              105 net/batman-adv/bridge_loop_avoidance.c 	const void *data1 = container_of(node, struct batadv_bla_backbone_gw,
node              126 net/batman-adv/bridge_loop_avoidance.c static bool batadv_compare_claim(const struct hlist_node *node,
node              129 net/batman-adv/bridge_loop_avoidance.c 	const void *data1 = container_of(node, struct batadv_bla_claim,
node              215 net/batman-adv/distributed-arp-table.c static bool batadv_compare_dat(const struct hlist_node *node, const void *data2)
node              217 net/batman-adv/distributed-arp-table.c 	const void *data1 = container_of(node, struct batadv_dat_entry,
node               42 net/batman-adv/fragmentation.c 	struct hlist_node *node;
node               44 net/batman-adv/fragmentation.c 	hlist_for_each_entry_safe(entry, node, head, list) {
node               84 net/batman-adv/hash.h 	struct hlist_node *node;
node               96 net/batman-adv/hash.h 	hlist_for_each(node, head) {
node               97 net/batman-adv/hash.h 		if (!compare(node, data))
node              135 net/batman-adv/hash.h 	struct hlist_node *node;
node              143 net/batman-adv/hash.h 	hlist_for_each(node, head) {
node              144 net/batman-adv/hash.h 		if (!compare(node, data))
node              147 net/batman-adv/hash.h 		data_save = node;
node              148 net/batman-adv/hash.h 		hlist_del_rcu(node);
node             1753 net/batman-adv/multicast.c 	struct hlist_node *node = &orig->mcast_want_all_unsnoopables_node;
node             1765 net/batman-adv/multicast.c 		WARN_ON(!hlist_unhashed(node));
node             1767 net/batman-adv/multicast.c 		hlist_add_head_rcu(node, head);
node             1776 net/batman-adv/multicast.c 		WARN_ON(hlist_unhashed(node));
node             1778 net/batman-adv/multicast.c 		hlist_del_init_rcu(node);
node             1798 net/batman-adv/multicast.c 	struct hlist_node *node = &orig->mcast_want_all_ipv4_node;
node             1810 net/batman-adv/multicast.c 		WARN_ON(!hlist_unhashed(node));
node             1812 net/batman-adv/multicast.c 		hlist_add_head_rcu(node, head);
node             1821 net/batman-adv/multicast.c 		WARN_ON(hlist_unhashed(node));
node             1823 net/batman-adv/multicast.c 		hlist_del_init_rcu(node);
node             1843 net/batman-adv/multicast.c 	struct hlist_node *node = &orig->mcast_want_all_ipv6_node;
node             1855 net/batman-adv/multicast.c 		WARN_ON(!hlist_unhashed(node));
node             1857 net/batman-adv/multicast.c 		hlist_add_head_rcu(node, head);
node             1866 net/batman-adv/multicast.c 		WARN_ON(hlist_unhashed(node));
node             1868 net/batman-adv/multicast.c 		hlist_del_init_rcu(node);
node             1888 net/batman-adv/multicast.c 	struct hlist_node *node = &orig->mcast_want_all_rtr4_node;
node             1900 net/batman-adv/multicast.c 		WARN_ON(!hlist_unhashed(node));
node             1902 net/batman-adv/multicast.c 		hlist_add_head_rcu(node, head);
node             1911 net/batman-adv/multicast.c 		WARN_ON(hlist_unhashed(node));
node             1913 net/batman-adv/multicast.c 		hlist_del_init_rcu(node);
node             1933 net/batman-adv/multicast.c 	struct hlist_node *node = &orig->mcast_want_all_rtr6_node;
node             1945 net/batman-adv/multicast.c 		WARN_ON(!hlist_unhashed(node));
node             1947 net/batman-adv/multicast.c 		hlist_add_head_rcu(node, head);
node             1956 net/batman-adv/multicast.c 		WARN_ON(hlist_unhashed(node));
node             1958 net/batman-adv/multicast.c 		hlist_del_init_rcu(node);
node              513 net/batman-adv/network-coding.c static bool batadv_nc_hash_compare(const struct hlist_node *node,
node              518 net/batman-adv/network-coding.c 	nc_path1 = container_of(node, struct batadv_nc_path, hash_entry);
node               96 net/batman-adv/originator.c bool batadv_compare_orig(const struct hlist_node *node, const void *data2)
node               98 net/batman-adv/originator.c 	const void *data1 = container_of(node, struct batadv_orig_node,
node               20 net/batman-adv/originator.h bool batadv_compare_orig(const struct hlist_node *node, const void *data2);
node               87 net/batman-adv/translation-table.c static bool batadv_compare_tt(const struct hlist_node *node, const void *data2)
node               89 net/batman-adv/translation-table.c 	const void *data1 = container_of(node, struct batadv_tt_common_entry,
node             2824 net/batman-adv/translation-table.c 	struct batadv_tt_req_node *node;
node             2829 net/batman-adv/translation-table.c 	hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) {
node             2830 net/batman-adv/translation-table.c 		hlist_del_init(&node->list);
node             2831 net/batman-adv/translation-table.c 		batadv_tt_req_node_put(node);
node             2860 net/batman-adv/translation-table.c 	struct batadv_tt_req_node *node;
node             2864 net/batman-adv/translation-table.c 	hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) {
node             2865 net/batman-adv/translation-table.c 		if (batadv_has_timed_out(node->issued_at,
node             2867 net/batman-adv/translation-table.c 			hlist_del_init(&node->list);
node             2868 net/batman-adv/translation-table.c 			batadv_tt_req_node_put(node);
node             3607 net/batman-adv/translation-table.c 	struct batadv_tt_req_node *node;
node             3646 net/batman-adv/translation-table.c 	hlist_for_each_entry_safe(node, safe, &bat_priv->tt.req_list, list) {
node             3647 net/batman-adv/translation-table.c 		if (!batadv_compare_eth(node->addr, resp_src))
node             3649 net/batman-adv/translation-table.c 		hlist_del_init(&node->list);
node             3650 net/batman-adv/translation-table.c 		batadv_tt_req_node_put(node);
node             3661 net/batman-adv/translation-table.c 	struct batadv_tt_roam_node *node, *safe;
node             3665 net/batman-adv/translation-table.c 	list_for_each_entry_safe(node, safe, &bat_priv->tt.roam_list, list) {
node             3666 net/batman-adv/translation-table.c 		list_del(&node->list);
node             3667 net/batman-adv/translation-table.c 		kmem_cache_free(batadv_tt_roam_cache, node);
node             3675 net/batman-adv/translation-table.c 	struct batadv_tt_roam_node *node, *safe;
node             3678 net/batman-adv/translation-table.c 	list_for_each_entry_safe(node, safe, &bat_priv->tt.roam_list, list) {
node             3679 net/batman-adv/translation-table.c 		if (!batadv_has_timed_out(node->first_time,
node             3683 net/batman-adv/translation-table.c 		list_del(&node->list);
node             3684 net/batman-adv/translation-table.c 		kmem_cache_free(batadv_tt_roam_cache, node);
node               30 net/caif/cfcnfg.c 	struct list_head node;
node              138 net/caif/cfcnfg.c 	list_for_each_entry_rcu(phy, &cnfg->phys, node)
node              154 net/caif/cfcnfg.c 	list_for_each_entry_rcu(phy, &cnfg->phys, node) {
node              162 net/caif/cfcnfg.c 	list_for_each_entry_rcu(phy, &cnfg->phys, node)
node              173 net/caif/cfcnfg.c 	list_for_each_entry_rcu(phy, &cnfg->phys, node)
node              511 net/caif/cfcnfg.c 	list_add_rcu(&phyinfo->node, &cnfg->phys);
node              576 net/caif/cfcnfg.c 	list_del_rcu(&phyinfo->node);
node              582 net/caif/cfcnfg.c 		list_add_rcu(&phyinfo->node, &cnfg->phys);
node               70 net/caif/cfmuxl.c 	list_add_rcu(&dn->node, &muxl->frml_list);
node               78 net/caif/cfmuxl.c 	list_for_each_entry_rcu(lyr, list, node) {
node               96 net/caif/cfmuxl.c 		list_del_rcu(&old->node);
node               98 net/caif/cfmuxl.c 	list_add_rcu(&up->node, &muxl->srvl_list);
node              116 net/caif/cfmuxl.c 	list_del_rcu(&dn->node);
node              168 net/caif/cfmuxl.c 	list_del_rcu(&up->node);
node              253 net/caif/cfmuxl.c 	list_for_each_entry_rcu(layer, &muxl->srvl_list, node) {
node              125 net/ceph/auth_x.c 		th = rb_entry(parent, struct ceph_x_ticket_handler, node);
node              139 net/ceph/auth_x.c 	rb_link_node(&th->node, parent, p);
node              140 net/ceph/auth_x.c 	rb_insert_color(&th->node, &xi->ticket_handlers);
node              150 net/ceph/auth_x.c 	rb_erase(&th->node, &xi->ticket_handlers);
node              773 net/ceph/auth_x.c 			rb_entry(p, struct ceph_x_ticket_handler, node);
node               16 net/ceph/auth_x.h 	struct rb_node node;
node               71 net/ceph/debugfs.c 			rb_entry(n, struct ceph_pg_pool_info, node);
node               92 net/ceph/debugfs.c 			rb_entry(n, struct ceph_pg_mapping, node);
node              103 net/ceph/debugfs.c 			rb_entry(n, struct ceph_pg_mapping, node);
node              110 net/ceph/debugfs.c 			rb_entry(n, struct ceph_pg_mapping, node);
node              121 net/ceph/debugfs.c 			rb_entry(n, struct ceph_pg_mapping, node);
node              160 net/ceph/debugfs.c 		req = rb_entry(rp, struct ceph_mon_generic_request, node);
node              262 net/ceph/debugfs.c 		    rb_entry(n, struct ceph_osd_linger_request, node);
node              506 net/ceph/mon_client.c DEFINE_RB_FUNCS(generic_request, struct ceph_mon_generic_request, tid, node)
node              515 net/ceph/mon_client.c 	WARN_ON(!RB_EMPTY_NODE(&req->node));
node              547 net/ceph/mon_client.c 	RB_CLEAR_NODE(&req->node);
node              961 net/ceph/mon_client.c 		req = rb_entry(p, struct ceph_mon_generic_request, node);
node             1328 net/ceph/osd_client.c 		    rb_entry(n, struct ceph_osd_linger_request, node);
node             1461 net/ceph/osd_client.c 		    rb_entry(n, struct ceph_pg_pool_info, node);
node             1622 net/ceph/osd_client.c 	RB_CLEAR_NODE(&spg->node);
node             1629 net/ceph/osd_client.c 	WARN_ON(!RB_EMPTY_NODE(&spg->node));
node             1642 net/ceph/osd_client.c 		 RB_BYPTR, const struct ceph_spg *, node)
node             1874 net/ceph/osd_client.c 			     struct ceph_spg_mapping, node);
node             2660 net/ceph/osd_client.c 	WARN_ON(!RB_EMPTY_NODE(&lreq->node));
node             2699 net/ceph/osd_client.c 	RB_CLEAR_NODE(&lreq->node);
node             2714 net/ceph/osd_client.c DEFINE_RB_INSDEL_FUNCS(linger, struct ceph_osd_linger_request, linger_id, node)
node             3312 net/ceph/osd_client.c 			    rb_entry(p, struct ceph_osd_linger_request, node);
node             3710 net/ceph/osd_client.c 		    rb_entry(n, struct ceph_pg_pool_info, node);
node             3763 net/ceph/osd_client.c 		    rb_entry(n, struct ceph_osd_linger_request, node);
node             3857 net/ceph/osd_client.c 			    rb_entry(n, struct ceph_pg_pool_info, node);
node             4089 net/ceph/osd_client.c 		    rb_entry(n, struct ceph_osd_linger_request, node);
node              149 net/ceph/osdmap.c 	RB_CLEAR_NODE(&arg_map->node);
node              158 net/ceph/osdmap.c 		WARN_ON(!RB_EMPTY_NODE(&arg_map->node));
node              174 net/ceph/osdmap.c 		node);
node              181 net/ceph/osdmap.c 			     struct crush_choose_arg_map, node);
node              618 net/ceph/osdmap.c 	RB_CLEAR_NODE(&pg->node);
node              624 net/ceph/osdmap.c 	WARN_ON(!RB_EMPTY_NODE(&pg->node));
node              634 net/ceph/osdmap.c 		 RB_BYPTR, const struct ceph_pg *, node)
node              647 net/ceph/osdmap.c 		pi = rb_entry(parent, struct ceph_pg_pool_info, node);
node              656 net/ceph/osdmap.c 	rb_link_node(&new->node, parent, p);
node              657 net/ceph/osdmap.c 	rb_insert_color(&new->node, root);
node              667 net/ceph/osdmap.c 		pi = rb_entry(n, struct ceph_pg_pool_info, node);
node              705 net/ceph/osdmap.c 			rb_entry(rbp, struct ceph_pg_pool_info, node);
node              724 net/ceph/osdmap.c 	rb_erase(&pi->node, root);
node              954 net/ceph/osdmap.c 				 struct ceph_pg_mapping, node);
node              961 net/ceph/osdmap.c 				 struct ceph_pg_mapping, node);
node              968 net/ceph/osdmap.c 				 struct ceph_pg_mapping, node);
node              969 net/ceph/osdmap.c 		rb_erase(&pg->node, &map->pg_upmap);
node              975 net/ceph/osdmap.c 				 struct ceph_pg_mapping, node);
node              976 net/ceph/osdmap.c 		rb_erase(&pg->node, &map->pg_upmap_items);
node              982 net/ceph/osdmap.c 				 struct ceph_pg_pool_info, node);
node               21 net/ceph/string_table.c 		exist = rb_entry(*p, struct ceph_string, node);
node               32 net/ceph/string_table.c 		rb_erase(&exist->node, &string_tree);
node               33 net/ceph/string_table.c 		RB_CLEAR_NODE(&exist->node);
node               56 net/ceph/string_table.c 		exist = rb_entry(*p, struct ceph_string, node);
node               68 net/ceph/string_table.c 		rb_link_node(&cs->node, parent, p);
node               69 net/ceph/string_table.c 		rb_insert_color(&cs->node, &string_tree);
node               71 net/ceph/string_table.c 		rb_erase(&exist->node, &string_tree);
node               72 net/ceph/string_table.c 		RB_CLEAR_NODE(&exist->node);
node               93 net/ceph/string_table.c 	if (!RB_EMPTY_NODE(&cs->node)) {
node               94 net/ceph/string_table.c 		rb_erase(&cs->node, &string_tree);
node               95 net/ceph/string_table.c 		RB_CLEAR_NODE(&cs->node);
node              409 net/core/pktgen.c 	int node;               /* Memory node */
node              637 net/core/pktgen.c 	if (pkt_dev->node >= 0)
node              638 net/core/pktgen.c 		seq_printf(seq, "     node: %d\n", pkt_dev->node);
node             1160 net/core/pktgen.c 			pkt_dev->node = value;
node             1161 net/core/pktgen.c 			sprintf(pg_result, "OK: node=%d", pkt_dev->node);
node             2643 net/core/pktgen.c 				int node = numa_node_id();
node             2645 net/core/pktgen.c 				if (pkt_dev->node >= 0 && (pkt_dev->flags & F_NODE))
node             2646 net/core/pktgen.c 					node = pkt_dev->node;
node             2647 net/core/pktgen.c 				pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
node             2701 net/core/pktgen.c 		int node = pkt_dev->node >= 0 ? pkt_dev->node : numa_node_id();
node             2703 net/core/pktgen.c 		skb = __alloc_skb(NET_SKB_PAD + size, GFP_NOWAIT, 0, node);
node             3594 net/core/pktgen.c 	int node = cpu_to_node(t->cpu);
node             3604 net/core/pktgen.c 	pkt_dev = kzalloc_node(sizeof(struct pktgen_dev), GFP_KERNEL, node);
node             3611 net/core/pktgen.c 				      node);
node             3633 net/core/pktgen.c 	pkt_dev->node = NUMA_NO_NODE;
node              128 net/core/skbuff.c #define kmalloc_reserve(size, gfp, node, pfmemalloc) \
node              129 net/core/skbuff.c 	 __kmalloc_reserve(size, gfp, node, _RET_IP_, pfmemalloc)
node              131 net/core/skbuff.c static void *__kmalloc_reserve(size_t size, gfp_t flags, int node,
node              143 net/core/skbuff.c 					node);
node              149 net/core/skbuff.c 	obj = kmalloc_node_track_caller(size, flags, node);
node              182 net/core/skbuff.c 			    int flags, int node)
node              197 net/core/skbuff.c 	skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node);
node              209 net/core/skbuff.c 	data = kmalloc_reserve(size, gfp_mask, node, &pfmemalloc);
node              494 net/core/skmsg.c struct sk_psock *sk_psock_init(struct sock *sk, int node)
node              498 net/core/skmsg.c 					      node);
node             3418 net/core/sock.c 	list_add(&prot->node, &proto_list);
node             3441 net/core/sock.c 	list_del(&prot->node);
node             3561 net/core/sock.c 		proto_seq_printf(seq, list_entry(v, struct proto, node));
node              522 net/core/sock_map.c 	struct hlist_node node;
node              557 net/core/sock_map.c 	hlist_for_each_entry_rcu(elem, head, node) {
node              607 net/core/sock_map.c 		hlist_del_rcu(&elem->node);
node              628 net/core/sock_map.c 		hlist_del_rcu(&elem->node);
node              715 net/core/sock_map.c 	hlist_add_head_rcu(&elem_new->node, &bucket->head);
node              717 net/core/sock_map.c 		hlist_del_rcu(&elem->node);
node              780 net/core/sock_map.c 	elem_next = hlist_entry_safe(rcu_dereference_raw(hlist_next_rcu(&elem->node)),
node              781 net/core/sock_map.c 				     struct bpf_htab_elem, node);
node              793 net/core/sock_map.c 					     struct bpf_htab_elem, node);
node              865 net/core/sock_map.c 	struct hlist_node *node;
node              875 net/core/sock_map.c 		hlist_for_each_entry_safe(elem, node, &bucket->head, node) {
node              876 net/core/sock_map.c 			hlist_del_rcu(&elem->node);
node               57 net/core/xdp.c 	.head_offset = offsetof(struct xdp_mem_allocator, node),
node               88 net/core/xdp.c 	if (!rhashtable_remove_fast(mem_id_ht, &xa->node, mem_id_rht_params))
node              131 net/core/xdp.c 	if (!rhashtable_remove_fast(mem_id_ht, &xa->node, mem_id_rht_params))
node              339 net/core/xdp.c 	ptr = rhashtable_insert_slow(mem_id_ht, &id, &xdp_alloc->node);
node              357 net/dccp/ackvec.c 	list_add_tail(&new->node, head);
node              366 net/dccp/ackvec.c 	list_for_each_entry_safe(cur, next, parsed_chunks, node)
node              131 net/dccp/ackvec.h 	struct list_head node;
node              584 net/dccp/ccids/ccid2.c 	list_for_each_entry(avp, &hc->tx_av_chunks, node) {
node              295 net/dccp/feat.c 	list_for_each_entry(___entry, fn_list, node)	\
node              431 net/dccp/feat.c 	list_for_each_entry(entry, fn_list, node) {
node              453 net/dccp/feat.c 	list_for_each_entry(entry, head, node)
node              458 net/dccp/feat.c 			head = &entry->node;
node              466 net/dccp/feat.c 		list_add_tail(&entry->node, head);
node              535 net/dccp/feat.c 	list_del(&entry->node);
node              543 net/dccp/feat.c 	list_for_each_entry_safe(entry, next, fn_list, node)
node              555 net/dccp/feat.c 	list_for_each_entry(entry, from, node) {
node              559 net/dccp/feat.c 		list_add_tail(&new->node, to);
node              638 net/dccp/feat.c 	list_for_each_entry_safe_reverse(pos, next, fn, node) {
node              983 net/dccp/feat.c 	list_for_each_entry(entry, fn, node)
node             1506 net/dccp/feat.c 	list_for_each_entry(cur, fn_list, node) {
node             1544 net/dccp/feat.c 	list_for_each_entry_safe(cur, next, fn_list, node)
node               76 net/dccp/feat.h 	struct list_head	node;
node              657 net/decnet/af_decnet.c 	unsigned short node, area;
node              659 net/decnet/af_decnet.c 	node = addr & 0x03ff;
node              661 net/decnet/af_decnet.c 	sprintf(buf, "%hd.%hd", area, node);
node               92 net/decnet/sysctl_net_decnet.c 	__u16 area, node;
node              111 net/decnet/sysctl_net_decnet.c 	node = *str++ - '0';
node              113 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              114 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              117 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              118 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              121 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              122 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              125 net/decnet/sysctl_net_decnet.c 	if ((node > 1023) || (area > 63))
node              131 net/decnet/sysctl_net_decnet.c 	*addr = cpu_to_le16((area << 10) | node);
node              144 net/dsa/dsa2.c 		link_dp = dsa_tree_find_port_by_node(dst, it.node);
node              146 net/dsa/dsa2.c 			of_node_put(it.node);
node               36 net/hsr/hsr_debugfs.c 	struct hsr_node *node;
node               42 net/hsr/hsr_debugfs.c 	list_for_each_entry_rcu(node, &priv->node_db, mac_list) {
node               44 net/hsr/hsr_debugfs.c 		if (hsr_addr_is_self(priv, node->macaddress_A))
node               46 net/hsr/hsr_debugfs.c 		print_mac_address(sfp, &node->macaddress_A[0]);
node               48 net/hsr/hsr_debugfs.c 		print_mac_address(sfp, &node->macaddress_B[0]);
node               49 net/hsr/hsr_debugfs.c 		seq_printf(sfp, "0x%lx, ", node->time_in[HSR_PT_SLAVE_A]);
node               50 net/hsr/hsr_debugfs.c 		seq_printf(sfp, "0x%lx ", node->time_in[HSR_PT_SLAVE_B]);
node               51 net/hsr/hsr_debugfs.c 		seq_printf(sfp, "0x%x\n", node->addr_B_port);
node               43 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node               45 net/hsr/hsr_framereg.c 	node = list_first_or_null_rcu(&hsr->self_node_db, struct hsr_node,
node               47 net/hsr/hsr_framereg.c 	if (!node) {
node               52 net/hsr/hsr_framereg.c 	if (ether_addr_equal(addr, node->macaddress_A))
node               54 net/hsr/hsr_framereg.c 	if (ether_addr_equal(addr, node->macaddress_B))
node               65 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node               67 net/hsr/hsr_framereg.c 	list_for_each_entry_rcu(node, node_db, mac_list) {
node               68 net/hsr/hsr_framereg.c 		if (ether_addr_equal(node->macaddress_A, addr))
node               69 net/hsr/hsr_framereg.c 			return node;
node               83 net/hsr/hsr_framereg.c 	struct hsr_node *node, *oldnode;
node               85 net/hsr/hsr_framereg.c 	node = kmalloc(sizeof(*node), GFP_KERNEL);
node               86 net/hsr/hsr_framereg.c 	if (!node)
node               89 net/hsr/hsr_framereg.c 	ether_addr_copy(node->macaddress_A, addr_a);
node               90 net/hsr/hsr_framereg.c 	ether_addr_copy(node->macaddress_B, addr_b);
node               96 net/hsr/hsr_framereg.c 		list_replace_rcu(&oldnode->mac_list, &node->mac_list);
node              100 net/hsr/hsr_framereg.c 		list_add_tail_rcu(&node->mac_list, self_node_db);
node              110 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              113 net/hsr/hsr_framereg.c 	node = list_first_or_null_rcu(self_node_db, struct hsr_node, mac_list);
node              114 net/hsr/hsr_framereg.c 	if (node) {
node              115 net/hsr/hsr_framereg.c 		list_del_rcu(&node->mac_list);
node              116 net/hsr/hsr_framereg.c 		kfree_rcu(node, rcu_head);
node              123 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              126 net/hsr/hsr_framereg.c 	list_for_each_entry_safe(node, tmp, node_db, mac_list)
node              127 net/hsr/hsr_framereg.c 		kfree(node);
node              139 net/hsr/hsr_framereg.c 	struct hsr_node *new_node, *node;
node              159 net/hsr/hsr_framereg.c 	list_for_each_entry_rcu(node, node_db, mac_list) {
node              160 net/hsr/hsr_framereg.c 		if (ether_addr_equal(node->macaddress_A, addr))
node              162 net/hsr/hsr_framereg.c 		if (ether_addr_equal(node->macaddress_B, addr))
node              171 net/hsr/hsr_framereg.c 	return node;
node              181 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              190 net/hsr/hsr_framereg.c 	list_for_each_entry_rcu(node, node_db, mac_list) {
node              191 net/hsr/hsr_framereg.c 		if (ether_addr_equal(node->macaddress_A, ethhdr->h_source))
node              192 net/hsr/hsr_framereg.c 			return node;
node              193 net/hsr/hsr_framereg.c 		if (ether_addr_equal(node->macaddress_B, ethhdr->h_source))
node              194 net/hsr/hsr_framereg.c 			return node;
node              286 net/hsr/hsr_framereg.c void hsr_addr_subst_source(struct hsr_node *node, struct sk_buff *skb)
node              293 net/hsr/hsr_framereg.c 	memcpy(&eth_hdr(skb)->h_source, node->macaddress_A, ETH_ALEN);
node              330 net/hsr/hsr_framereg.c void hsr_register_frame_in(struct hsr_node *node, struct hsr_port *port,
node              337 net/hsr/hsr_framereg.c 	if (seq_nr_before(sequence_nr, node->seq_out[port->type]))
node              340 net/hsr/hsr_framereg.c 	node->time_in[port->type] = jiffies;
node              341 net/hsr/hsr_framereg.c 	node->time_in_stale[port->type] = false;
node              352 net/hsr/hsr_framereg.c int hsr_register_frame_out(struct hsr_port *port, struct hsr_node *node,
node              355 net/hsr/hsr_framereg.c 	if (seq_nr_before_or_eq(sequence_nr, node->seq_out[port->type]))
node              358 net/hsr/hsr_framereg.c 	node->seq_out[port->type] = sequence_nr;
node              363 net/hsr/hsr_framereg.c 				      struct hsr_node *node)
node              365 net/hsr/hsr_framereg.c 	if (node->time_in_stale[HSR_PT_SLAVE_A])
node              367 net/hsr/hsr_framereg.c 	if (node->time_in_stale[HSR_PT_SLAVE_B])
node              370 net/hsr/hsr_framereg.c 	if (time_after(node->time_in[HSR_PT_SLAVE_B],
node              371 net/hsr/hsr_framereg.c 		       node->time_in[HSR_PT_SLAVE_A] +
node              374 net/hsr/hsr_framereg.c 	if (time_after(node->time_in[HSR_PT_SLAVE_A],
node              375 net/hsr/hsr_framereg.c 		       node->time_in[HSR_PT_SLAVE_B] +
node              388 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              395 net/hsr/hsr_framereg.c 	list_for_each_entry_safe(node, tmp, &hsr->node_db, mac_list) {
node              401 net/hsr/hsr_framereg.c 		if (hsr_addr_is_self(hsr, node->macaddress_A))
node              405 net/hsr/hsr_framereg.c 		time_a = node->time_in[HSR_PT_SLAVE_A];
node              406 net/hsr/hsr_framereg.c 		time_b = node->time_in[HSR_PT_SLAVE_B];
node              410 net/hsr/hsr_framereg.c 			node->time_in_stale[HSR_PT_SLAVE_A] = true;
node              412 net/hsr/hsr_framereg.c 			node->time_in_stale[HSR_PT_SLAVE_B] = true;
node              419 net/hsr/hsr_framereg.c 		if (node->time_in_stale[HSR_PT_SLAVE_A] ||
node              420 net/hsr/hsr_framereg.c 		    (!node->time_in_stale[HSR_PT_SLAVE_B] &&
node              428 net/hsr/hsr_framereg.c 			port = get_late_port(hsr, node);
node              430 net/hsr/hsr_framereg.c 				hsr_nl_ringerror(hsr, node->macaddress_A, port);
node              437 net/hsr/hsr_framereg.c 			hsr_nl_nodedown(hsr, node->macaddress_A);
node              438 net/hsr/hsr_framereg.c 			list_del_rcu(&node->mac_list);
node              440 net/hsr/hsr_framereg.c 			kfree_rcu(node, rcu_head);
node              453 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              456 net/hsr/hsr_framereg.c 		node = list_first_or_null_rcu(&hsr->node_db,
node              458 net/hsr/hsr_framereg.c 		if (node)
node              459 net/hsr/hsr_framereg.c 			ether_addr_copy(addr, node->macaddress_A);
node              460 net/hsr/hsr_framereg.c 		return node;
node              463 net/hsr/hsr_framereg.c 	node = _pos;
node              464 net/hsr/hsr_framereg.c 	list_for_each_entry_continue_rcu(node, &hsr->node_db, mac_list) {
node              465 net/hsr/hsr_framereg.c 		ether_addr_copy(addr, node->macaddress_A);
node              466 net/hsr/hsr_framereg.c 		return node;
node              481 net/hsr/hsr_framereg.c 	struct hsr_node *node;
node              485 net/hsr/hsr_framereg.c 	node = find_node_by_addr_A(&hsr->node_db, addr);
node              486 net/hsr/hsr_framereg.c 	if (!node)
node              489 net/hsr/hsr_framereg.c 	ether_addr_copy(addr_b, node->macaddress_B);
node              491 net/hsr/hsr_framereg.c 	tdiff = jiffies - node->time_in[HSR_PT_SLAVE_A];
node              492 net/hsr/hsr_framereg.c 	if (node->time_in_stale[HSR_PT_SLAVE_A])
node              501 net/hsr/hsr_framereg.c 	tdiff = jiffies - node->time_in[HSR_PT_SLAVE_B];
node              502 net/hsr/hsr_framereg.c 	if (node->time_in_stale[HSR_PT_SLAVE_B])
node              512 net/hsr/hsr_framereg.c 	*if1_seq = node->seq_out[HSR_PT_SLAVE_B];
node              513 net/hsr/hsr_framereg.c 	*if2_seq = node->seq_out[HSR_PT_SLAVE_A];
node              515 net/hsr/hsr_framereg.c 	if (node->addr_B_port != HSR_PT_NONE) {
node              516 net/hsr/hsr_framereg.c 		port = hsr_port_get_hsr(hsr, node->addr_B_port);
node               23 net/hsr/hsr_framereg.h void hsr_addr_subst_source(struct hsr_node *node, struct sk_buff *skb);
node               27 net/hsr/hsr_framereg.h void hsr_register_frame_in(struct hsr_node *node, struct hsr_port *port,
node               29 net/hsr/hsr_framereg.h int hsr_register_frame_out(struct hsr_port *port, struct hsr_node *node,
node              508 net/ieee802154/6lowpan/reassembly.c 	.head_offset		= offsetof(struct inet_frag_queue, node),
node              913 net/ipv4/inet_diag.c 			struct hlist_nulls_node *node;
node              918 net/ipv4/inet_diag.c 			sk_nulls_for_each(sk, node, &ilb->nulls_head) {
node              962 net/ipv4/inet_diag.c 		struct hlist_nulls_node *node;
node              977 net/ipv4/inet_diag.c 		sk_nulls_for_each(sk, node, &head->chain) {
node              210 net/ipv4/inet_fragment.c 			rhashtable_remove_fast(&fqdir->rhashtable, &fq->node,
node              311 net/ipv4/inet_fragment.c 						 &q->node, f->rhash_params);
node               76 net/ipv4/inet_hashtables.c 		hlist_add_head(&tb->node, &head->chain);
node               87 net/ipv4/inet_hashtables.c 		__hlist_del(&tb->node);
node              355 net/ipv4/inet_hashtables.c 	const struct hlist_nulls_node *node;
node              364 net/ipv4/inet_hashtables.c 	sk_nulls_for_each_rcu(sk, node, &head->chain) {
node              385 net/ipv4/inet_hashtables.c 	if (get_nulls_value(node) != slot)
node              413 net/ipv4/inet_hashtables.c 	const struct hlist_nulls_node *node;
node              418 net/ipv4/inet_hashtables.c 	sk_nulls_for_each(sk2, node, &head->chain) {
node              519 net/ipv4/inet_hashtables.c 	const struct hlist_nulls_node *node;
node              523 net/ipv4/inet_hashtables.c 	sk_nulls_for_each_rcu(sk2, node, &ilb->nulls_head) {
node              261 net/ipv4/inet_timewait_sock.c 	struct hlist_nulls_node *node;
node              270 net/ipv4/inet_timewait_sock.c 		sk_nulls_for_each_rcu(sk, node, &head->chain) {
node              297 net/ipv4/inet_timewait_sock.c 		if (get_nulls_value(node) != slot)
node              730 net/ipv4/ip_fragment.c 	.head_offset		= offsetof(struct inet_frag_queue, node),
node             1088 net/ipv4/nexthop.c 	struct rb_node *node;
node             1091 net/ipv4/nexthop.c 	while ((node = rb_first(root))) {
node             1092 net/ipv4/nexthop.c 		nh = rb_entry(node, struct nexthop, rb_node);
node             1721 net/ipv4/nexthop.c 	struct rb_node *node;
node             1731 net/ipv4/nexthop.c 	for (node = rb_first(root); node; node = rb_next(node)) {
node             1737 net/ipv4/nexthop.c 		nh = rb_entry(node, struct nexthop, rb_node);
node               79 net/ipv4/ping.c 	struct hlist_nulls_node *node;
node               95 net/ipv4/ping.c 			ping_portaddr_for_each_entry(sk2, node, hlist) {
node              112 net/ipv4/ping.c 		ping_portaddr_for_each_entry(sk2, node, hlist) {
node             1019 net/ipv4/ping.c 		struct hlist_nulls_node *node;
node             1027 net/ipv4/ping.c 		sk_nulls_for_each(sk, node, hslot) {
node               63 net/ipv4/tcp_diag.c 	hlist_for_each_entry_rcu(key, &md5sig->head, node)
node               75 net/ipv4/tcp_diag.c 	hlist_for_each_entry_rcu(key, &md5sig->head, node) {
node              158 net/ipv4/tcp_diag.c 			hlist_for_each_entry_rcu(key, &md5sig->head, node)
node             5077 net/ipv4/tcp_input.c 	struct rb_node *node, *prev;
node             5085 net/ipv4/tcp_input.c 	node = &tp->ooo_last_skb->rbnode;
node             5087 net/ipv4/tcp_input.c 		prev = rb_prev(node);
node             5088 net/ipv4/tcp_input.c 		rb_erase(node, &tp->out_of_order_queue);
node             5089 net/ipv4/tcp_input.c 		goal -= rb_to_skb(node)->truesize;
node             5090 net/ipv4/tcp_input.c 		tcp_drop(sk, rb_to_skb(node));
node             5098 net/ipv4/tcp_input.c 		node = prev;
node             5099 net/ipv4/tcp_input.c 	} while (node);
node             1005 net/ipv4/tcp_ipv4.c 	hlist_for_each_entry_rcu(key, &md5sig->head, node) {
node             1048 net/ipv4/tcp_ipv4.c 	hlist_for_each_entry_rcu(key, &md5sig->head, node) {
node             1113 net/ipv4/tcp_ipv4.c 	hlist_add_head_rcu(&key->node, &md5sig->head);
node             1126 net/ipv4/tcp_ipv4.c 	hlist_del_rcu(&key->node);
node             1142 net/ipv4/tcp_ipv4.c 	hlist_for_each_entry_safe(key, n, &md5sig->head, node) {
node             1143 net/ipv4/tcp_ipv4.c 		hlist_del_rcu(&key->node);
node             2152 net/ipv4/tcp_ipv4.c 	struct hlist_nulls_node *node;
node             2169 net/ipv4/tcp_ipv4.c 	sk_nulls_for_each_from(sk, node) {
node             2217 net/ipv4/tcp_ipv4.c 		struct hlist_nulls_node *node;
node             2225 net/ipv4/tcp_ipv4.c 		sk_nulls_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) {
node             2243 net/ipv4/tcp_ipv4.c 	struct hlist_nulls_node *node;
node             2252 net/ipv4/tcp_ipv4.c 	sk_nulls_for_each_from(sk, node) {
node             2156 net/ipv4/udp.c 	struct hlist_node *node;
node             2168 net/ipv4/udp.c 	sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) {
node               20 net/ipv6/ila/ila_xlat.c 	struct rhash_head node;
node               87 net/ipv6/ila/ila_xlat.c 	.head_offset = offsetof(struct ila_map, node),
node              241 net/ipv6/ila/ila_xlat.c 						    &ila->node, rht_params);
node              267 net/ipv6/ila/ila_xlat.c 						      &head->node,
node              268 net/ipv6/ila/ila_xlat.c 						      &ila->node, rht_params);
node              322 net/ipv6/ila/ila_xlat.c 					&ilan->xlat.rhash_table, &ila->node,
node              323 net/ipv6/ila/ila_xlat.c 					&head->node, rht_params);
node              330 net/ipv6/ila/ila_xlat.c 						&ila->node, rht_params);
node              408 net/ipv6/ila/ila_xlat.c 					     &ila->node, rht_params);
node               58 net/ipv6/inet6_hashtables.c 	const struct hlist_nulls_node *node;
node               69 net/ipv6/inet6_hashtables.c 	sk_nulls_for_each_rcu(sk, node, &head->chain) {
node               83 net/ipv6/inet6_hashtables.c 	if (get_nulls_value(node) != slot)
node              217 net/ipv6/inet6_hashtables.c 	const struct hlist_nulls_node *node;
node              222 net/ipv6/inet6_hashtables.c 	sk_nulls_for_each(sk2, node, &head->chain) {
node              545 net/ipv6/ip6_fib.c 			w->node = w->root;
node             1796 net/ipv6/ip6_fib.c 				if (w->node == fn) {
node             1798 net/ipv6/ip6_fib.c 					w->node = pn;
node             1802 net/ipv6/ip6_fib.c 				if (w->node == fn) {
node             1803 net/ipv6/ip6_fib.c 					w->node = child;
node             1961 net/ipv6/ip6_fib.c 		fn = w->node;
node             1969 net/ipv6/ip6_fib.c 				w->node = FIB6_SUBTREE(fn);
node             1978 net/ipv6/ip6_fib.c 				w->node = left;
node             1987 net/ipv6/ip6_fib.c 				w->node = right;
node             2019 net/ipv6/ip6_fib.c 			w->node = pn;
node             2033 net/ipv6/ip6_fib.c 				w->leaf = rcu_dereference_protected(w->node->leaf, 1);
node             2048 net/ipv6/ip6_fib.c 	w->node = w->root;
node             2068 net/ipv6/ip6_fib.c 	    w->node->fn_sernum != c->sernum)
node             2069 net/ipv6/ip6_fib.c 		w->node->fn_sernum = c->sernum;
node             2431 net/ipv6/ip6_fib.c 	iter->w.node = iter->w.root;
node             2442 net/ipv6/ip6_fib.c 	struct hlist_node *node;
node             2446 net/ipv6/ip6_fib.c 		node = rcu_dereference_bh(hlist_next_rcu(&tbl->tb6_hlist));
node             2449 net/ipv6/ip6_fib.c 		node = NULL;
node             2452 net/ipv6/ip6_fib.c 	while (!node && h < FIB6_TABLE_HASHSZ) {
node             2453 net/ipv6/ip6_fib.c 		node = rcu_dereference_bh(
node             2456 net/ipv6/ip6_fib.c 	return hlist_entry_safe(node, struct fib6_table, tb6_hlist);
node             2464 net/ipv6/ip6_fib.c 		iter->w.node = iter->w.root;
node             2530 net/ipv6/ip6_fib.c 	return w->node && !(w->state == FWS_U && w->node == w->root);
node              522 net/ipv6/netfilter/nf_conntrack_reasm.c 	.head_offset		= offsetof(struct inet_frag_queue, node),
node              541 net/ipv6/reassembly.c 	.head_offset		= offsetof(struct inet_frag_queue, node),
node               68 net/ipv6/seg6_hmac.c 	.head_offset		= offsetof(struct seg6_hmac_info, node),
node              298 net/ipv6/seg6_hmac.c 	err = rhashtable_lookup_insert_fast(&sdata->hmac_infos, &hinfo->node,
node              315 net/ipv6/seg6_hmac.c 	err = rhashtable_remove_fast(&sdata->hmac_infos, &hinfo->node,
node              755 net/ipv6/udp.c 	struct hlist_node *node;
node              767 net/ipv6/udp.c 	sk_for_each_entry_offset_rcu(sk, node, &hslot->head, offset) {
node               66 net/lapb/lapb_iface.c 	if (lapb->node.next) {
node               67 net/lapb/lapb_iface.c 		list_del(&lapb->node);
node               77 net/lapb/lapb_iface.c 	list_add(&lapb->node, &lapb_list);
node               87 net/lapb/lapb_iface.c 		lapb = list_entry(entry, struct lapb_cb, node);
node              482 net/llc/llc_conn.c 	struct hlist_nulls_node *node;
node              488 net/llc/llc_conn.c 	sk_nulls_for_each_rcu(rc, node, laddr_hb) {
node              507 net/llc/llc_conn.c 	if (unlikely(get_nulls_value(node) != slot))
node              541 net/llc/llc_conn.c 	struct hlist_nulls_node *node;
node              547 net/llc/llc_conn.c 	sk_nulls_for_each_rcu(rc, node, laddr_hb) {
node              566 net/llc/llc_conn.c 	if (unlikely(get_nulls_value(node) != slot))
node               53 net/llc/llc_core.c 	list_for_each_entry(sap, &llc_sap_list, node)
node              107 net/llc/llc_core.c 	list_add_tail_rcu(&sap->node, &llc_sap_list);
node              127 net/llc/llc_core.c 	list_del_rcu(&sap->node);
node               40 net/llc/llc_proc.c 	list_for_each_entry_rcu(sap, &llc_sap_list, node) {
node               44 net/llc/llc_proc.c 			struct hlist_nulls_node *node;
node               46 net/llc/llc_proc.c 			sk_nulls_for_each(sk, node, head) {
node               69 net/llc/llc_proc.c 	struct hlist_nulls_node *node;
node               73 net/llc/llc_proc.c 		sk_nulls_for_each(sk, node, &sap->sk_laddr_hash[bucket])
node              103 net/llc/llc_proc.c 	list_for_each_entry_continue_rcu(sap, &llc_sap_list, node) {
node              316 net/llc/llc_sap.c 	struct hlist_nulls_node *node;
node              322 net/llc/llc_sap.c 	sk_nulls_for_each_rcu(rc, node, laddr_hb) {
node              341 net/llc/llc_sap.c 	if (unlikely(get_nulls_value(node) != slot))
node              214 net/ncsi/internal.h 	struct list_head            node;
node              225 net/ncsi/internal.h 	struct list_head     node;        /* Form list of packages  */
node              309 net/ncsi/internal.h 	struct list_head    node;            /* Form NCSI device list      */
node              340 net/ncsi/internal.h 	list_for_each_entry_rcu(ndp, &ncsi_dev_list, node)
node              342 net/ncsi/internal.h 	list_for_each_entry_rcu(np, &ndp->packages, node)
node              344 net/ncsi/internal.h 	list_for_each_entry_rcu(nc, &np->channels, node)
node              230 net/ncsi/ncsi-manage.c 	list_add_tail_rcu(&nc->node, &np->channels);
node              254 net/ncsi/ncsi-manage.c 	list_del_rcu(&nc->node);
node              298 net/ncsi/ncsi-manage.c 	list_add_tail_rcu(&np->node, &ndp->packages);
node              312 net/ncsi/ncsi-manage.c 	list_for_each_entry_safe(nc, tmp, &np->channels, node)
node              317 net/ncsi/ncsi-manage.c 	list_del_rcu(&np->node);
node             1661 net/ncsi/ncsi-manage.c 	list_add_tail_rcu(&ndp->node, &ncsi_dev_list);
node             1822 net/ncsi/ncsi-manage.c 	list_for_each_entry_safe(np, tmp, &ndp->packages, node)
node             1826 net/ncsi/ncsi-manage.c 	list_del_rcu(&ndp->node);
node               42 net/netfilter/nf_conncount.c 	struct list_head		node;
node               50 net/netfilter/nf_conncount.c 	struct rb_node node;
node               91 net/netfilter/nf_conncount.c 	list_del(&conn->node);
node              136 net/netfilter/nf_conncount.c 	list_for_each_entry_safe(conn, conn_n, &list->head, node) {
node              191 net/netfilter/nf_conncount.c 	list_add_tail(&conn->node, &list->head);
node              234 net/netfilter/nf_conncount.c 	list_for_each_entry_safe(conn, conn_n, &list->head, node) {
node              286 net/netfilter/nf_conncount.c 			rb_erase(&rbconn->node, root);
node              322 net/netfilter/nf_conncount.c 		rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node);
node              373 net/netfilter/nf_conncount.c 	list_add(&conn->node, &rbconn->list.head);
node              377 net/netfilter/nf_conncount.c 	rb_link_node_rcu(&rbconn->node, parent, rbnode);
node              378 net/netfilter/nf_conncount.c 	rb_insert_color(&rbconn->node, root);
node              404 net/netfilter/nf_conncount.c 		rbconn = rb_entry(parent, struct nf_conncount_rb, node);
node              449 net/netfilter/nf_conncount.c 	struct rb_node *node;
node              457 net/netfilter/nf_conncount.c 	for (node = rb_first(root); node != NULL; node = rb_next(node)) {
node              458 net/netfilter/nf_conncount.c 		rbconn = rb_entry(node, struct nf_conncount_rb, node);
node              472 net/netfilter/nf_conncount.c 	node = rb_first(root);
node              473 net/netfilter/nf_conncount.c 	while (node != NULL) {
node              474 net/netfilter/nf_conncount.c 		rbconn = rb_entry(node, struct nf_conncount_rb, node);
node              475 net/netfilter/nf_conncount.c 		node = rb_next(node);
node              554 net/netfilter/nf_conncount.c 	list_for_each_entry_safe(conn, conn_n, &list->head, node)
node              562 net/netfilter/nf_conncount.c 	struct rb_node *node;
node              564 net/netfilter/nf_conncount.c 	while ((node = rb_first(r)) != NULL) {
node              565 net/netfilter/nf_conncount.c 		rbconn = rb_entry(node, struct nf_conncount_rb, node);
node              567 net/netfilter/nf_conncount.c 		rb_erase(node, r);
node              194 net/netfilter/nf_flow_table_core.c 	.head_offset		= offsetof(struct flow_offload_tuple_rhash, node),
node              208 net/netfilter/nf_flow_table_core.c 				     &flow->tuplehash[0].node,
node              214 net/netfilter/nf_flow_table_core.c 				     &flow->tuplehash[1].node,
node              218 net/netfilter/nf_flow_table_core.c 				       &flow->tuplehash[0].node,
node              238 net/netfilter/nf_flow_table_core.c 			       &flow->tuplehash[FLOW_OFFLOAD_DIR_ORIGINAL].node,
node              241 net/netfilter/nf_flow_table_core.c 			       &flow->tuplehash[FLOW_OFFLOAD_DIR_REPLY].node,
node               30 net/netfilter/nft_set_hash.c 	struct rhash_head		node;
node               70 net/netfilter/nft_set_hash.c 	.head_offset		= offsetof(struct nft_rhash_elem, node),
node              137 net/netfilter/nft_set_hash.c 	prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node,
node              171 net/netfilter/nft_set_hash.c 	prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node,
node              234 net/netfilter/nft_set_hash.c 	rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params);
node              252 net/netfilter/nft_set_hash.c 	return rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params) == 0;
node              333 net/netfilter/nft_set_hash.c 		rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params);
node              418 net/netfilter/nft_set_hash.c 	struct hlist_node		node;
node              432 net/netfilter/nft_set_hash.c 	hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
node              452 net/netfilter/nft_set_hash.c 	hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
node              472 net/netfilter/nft_set_hash.c 	hlist_for_each_entry_rcu(he, &priv->table[hash], node) {
node              510 net/netfilter/nft_set_hash.c 	hlist_for_each_entry(he, &priv->table[hash], node) {
node              518 net/netfilter/nft_set_hash.c 	hlist_add_head_rcu(&this->node, &priv->table[hash]);
node              549 net/netfilter/nft_set_hash.c 	hlist_for_each_entry(he, &priv->table[hash], node) {
node              566 net/netfilter/nft_set_hash.c 	hlist_del_rcu(&he->node);
node              578 net/netfilter/nft_set_hash.c 		hlist_for_each_entry_rcu(he, &priv->table[i], node) {
node              622 net/netfilter/nft_set_hash.c 		hlist_for_each_entry_safe(he, next, &priv->table[i], node) {
node              623 net/netfilter/nft_set_hash.c 			hlist_del_rcu(&he->node);
node               26 net/netfilter/nft_set_rbtree.c 	struct rb_node		node;
node               63 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(parent, struct nft_rbtree_elem, node);
node              144 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(parent, struct nft_rbtree_elem, node);
node              231 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(parent, struct nft_rbtree_elem, node);
node              254 net/netfilter/nft_set_rbtree.c 	rb_link_node_rcu(&new->node, parent, p);
node              255 net/netfilter/nft_set_rbtree.c 	rb_insert_color(&new->node, &priv->root);
node              285 net/netfilter/nft_set_rbtree.c 	rb_erase(&rbe->node, &priv->root);
node              324 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(parent, struct nft_rbtree_elem, node);
node              359 net/netfilter/nft_set_rbtree.c 	struct rb_node *node;
node              362 net/netfilter/nft_set_rbtree.c 	for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) {
node              363 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(node, struct nft_rbtree_elem, node);
node              390 net/netfilter/nft_set_rbtree.c 	struct rb_node *node;
node              398 net/netfilter/nft_set_rbtree.c 	for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) {
node              399 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(node, struct nft_rbtree_elem, node);
node              411 net/netfilter/nft_set_rbtree.c 			rb_erase(&rbe_prev->node, &priv->root);
node              425 net/netfilter/nft_set_rbtree.c 			rb_erase(&rbe_end->node, &priv->root);
node              428 net/netfilter/nft_set_rbtree.c 		node = rb_next(node);
node              429 net/netfilter/nft_set_rbtree.c 		if (!node)
node              433 net/netfilter/nft_set_rbtree.c 		rb_erase(&rbe_prev->node, &priv->root);
node              471 net/netfilter/nft_set_rbtree.c 	struct rb_node *node;
node              475 net/netfilter/nft_set_rbtree.c 	while ((node = priv->root.rb_node) != NULL) {
node              476 net/netfilter/nft_set_rbtree.c 		rb_erase(node, &priv->root);
node              477 net/netfilter/nft_set_rbtree.c 		rbe = rb_entry(node, struct nft_rbtree_elem, node);
node               91 net/netfilter/xt_hashlimit.c 	struct hlist_node node;
node              117 net/netfilter/xt_hashlimit.c 	struct hlist_node node;		/* global list of all htables */
node              206 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry_rcu(ent, &ht->hash[hash], node)
node              252 net/netfilter/xt_hashlimit.c 		hlist_add_head_rcu(&ent->node, &ht->hash[hash_dst(ht, dst)]);
node              269 net/netfilter/xt_hashlimit.c 	hlist_del_rcu(&ent->node);
node              356 net/netfilter/xt_hashlimit.c 	hlist_add_head(&hinfo->node, &hashlimit_net->htables);
node              370 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry_safe(dh, n, &ht->hash[i], node) {
node              412 net/netfilter/xt_hashlimit.c 	hlist_for_each_entry(hinfo, &hashlimit_net->htables, node) {
node              425 net/netfilter/xt_hashlimit.c 		hlist_del(&hinfo->node);
node             1177 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry(ent, &htable->hash[*bucket], node)
node             1191 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry(ent, &htable->hash[*bucket], node)
node             1205 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry(ent, &htable->hash[*bucket], node)
node             1260 net/netfilter/xt_hashlimit.c 	hlist_for_each_entry(hinfo, &hashlimit_net->htables, node)
node              513 net/netlink/af_netlink.c 					    &nlk_sk(sk)->node,
node              598 net/netlink/af_netlink.c 	if (!rhashtable_remove_fast(&table->hash, &nlk_sk(sk)->node,
node             2744 net/netlink/af_netlink.c 	.head_offset = offsetof(struct netlink_sock, node),
node               47 net/netlink/af_netlink.h 	struct rhash_head	node;
node               48 net/nfc/llcp.h 	struct hlist_node node;
node              128 net/nfc/llcp_commands.c 	INIT_HLIST_NODE(&sdres->node);
node              169 net/nfc/llcp_commands.c 	INIT_HLIST_NODE(&sdreq->node);
node              185 net/nfc/llcp_commands.c 	hlist_for_each_entry_safe(sdp, n, head, node) {
node              186 net/nfc/llcp_commands.c 		hlist_del(&sdp->node);
node              563 net/nfc/llcp_commands.c 	hlist_for_each_entry_safe(sdp, n, tlv_list, node) {
node              566 net/nfc/llcp_commands.c 		hlist_del(&sdp->node);
node              593 net/nfc/llcp_commands.c 	hlist_for_each_entry_safe(sdreq, n, tlv_list, node) {
node              598 net/nfc/llcp_commands.c 		hlist_del(&sdreq->node);
node              600 net/nfc/llcp_commands.c 		hlist_add_head(&sdreq->node, &local->pending_sdreqs);
node              255 net/nfc/llcp_core.c 	hlist_for_each_entry_safe(sdp, n, &local->pending_sdreqs, node) {
node              261 net/nfc/llcp_core.c 		hlist_del(&sdp->node);
node              263 net/nfc/llcp_core.c 		hlist_add_head(&sdp->node, &nl_sdres_list);
node             1319 net/nfc/llcp_core.c 			hlist_add_head(&sdp->node, &llc_sdres_list);
node             1327 net/nfc/llcp_core.c 			hlist_for_each_entry(sdp, &local->pending_sdreqs, node) {
node             1336 net/nfc/llcp_core.c 				hlist_del(&sdp->node);
node             1338 net/nfc/llcp_core.c 				hlist_add_head(&sdp->node, &nl_sdres_list);
node              395 net/nfc/netlink.c 	hlist_for_each_entry_safe(sdres, n, sdres_list, node) {
node              412 net/nfc/netlink.c 		hlist_del(&sdres->node);
node             1204 net/nfc/netlink.c 		hlist_add_head(&sdreq->node, &sdreq_list);
node              208 net/openvswitch/flow.h 		struct hlist_node node[2];
node              220 net/openvswitch/flow_table.c 		hlist_for_each_entry_safe(flow, n, head, flow_table.node[ver]) {
node              221 net/openvswitch/flow_table.c 			hlist_del_rcu(&flow->flow_table.node[ver]);
node              223 net/openvswitch/flow_table.c 				hlist_del_rcu(&flow->ufid_table.node[ufid_ver]);
node              261 net/openvswitch/flow_table.c 		hlist_for_each_entry_rcu(flow, head, flow_table.node[ver]) {
node              288 net/openvswitch/flow_table.c 	hlist_add_head_rcu(&flow->flow_table.node[ti->node_ver], head);
node              297 net/openvswitch/flow_table.c 	hlist_add_head_rcu(&flow->ufid_table.node[ti->node_ver], head);
node              316 net/openvswitch/flow_table.c 					     ufid_table.node[old_ver])
node              320 net/openvswitch/flow_table.c 					     flow_table.node[old_ver])
node              438 net/openvswitch/flow_table.c 	hlist_for_each_entry_rcu(flow, head, flow_table.node[ti->node_ver]) {
node              521 net/openvswitch/flow_table.c 	hlist_for_each_entry_rcu(flow, head, ufid_table.node[ti->node_ver]) {
node              571 net/openvswitch/flow_table.c 	hlist_del_rcu(&flow->flow_table.node[ti->node_ver]);
node              574 net/openvswitch/flow_table.c 		hlist_del_rcu(&flow->ufid_table.node[ufid_ti->node_ver]);
node              131 net/qrtr/qrtr.c static int qrtr_local_enqueue(struct qrtr_node *node, struct sk_buff *skb,
node              134 net/qrtr/qrtr.c static int qrtr_bcast_enqueue(struct qrtr_node *node, struct sk_buff *skb,
node              145 net/qrtr/qrtr.c 	struct qrtr_node *node = container_of(kref, struct qrtr_node, ref);
node              147 net/qrtr/qrtr.c 	if (node->nid != QRTR_EP_NID_AUTO)
node              148 net/qrtr/qrtr.c 		radix_tree_delete(&qrtr_nodes, node->nid);
node              150 net/qrtr/qrtr.c 	list_del(&node->item);
node              153 net/qrtr/qrtr.c 	cancel_work_sync(&node->work);
node              154 net/qrtr/qrtr.c 	skb_queue_purge(&node->rx_queue);
node              155 net/qrtr/qrtr.c 	kfree(node);
node              159 net/qrtr/qrtr.c static struct qrtr_node *qrtr_node_acquire(struct qrtr_node *node)
node              161 net/qrtr/qrtr.c 	if (node)
node              162 net/qrtr/qrtr.c 		kref_get(&node->ref);
node              163 net/qrtr/qrtr.c 	return node;
node              167 net/qrtr/qrtr.c static void qrtr_node_release(struct qrtr_node *node)
node              169 net/qrtr/qrtr.c 	if (!node)
node              171 net/qrtr/qrtr.c 	kref_put_mutex(&node->ref, __qrtr_node_release, &qrtr_node_lock);
node              175 net/qrtr/qrtr.c static int qrtr_node_enqueue(struct qrtr_node *node, struct sk_buff *skb,
node              189 net/qrtr/qrtr.c 		hdr->dst_node_id = cpu_to_le32(node->nid);
node              201 net/qrtr/qrtr.c 	mutex_lock(&node->ep_lock);
node              202 net/qrtr/qrtr.c 	if (node->ep)
node              203 net/qrtr/qrtr.c 		rc = node->ep->xmit(node->ep, skb);
node              206 net/qrtr/qrtr.c 	mutex_unlock(&node->ep_lock);
node              217 net/qrtr/qrtr.c 	struct qrtr_node *node;
node              220 net/qrtr/qrtr.c 	node = radix_tree_lookup(&qrtr_nodes, nid);
node              221 net/qrtr/qrtr.c 	node = qrtr_node_acquire(node);
node              224 net/qrtr/qrtr.c 	return node;
node              232 net/qrtr/qrtr.c static void qrtr_node_assign(struct qrtr_node *node, unsigned int nid)
node              234 net/qrtr/qrtr.c 	if (node->nid != QRTR_EP_NID_AUTO || nid == QRTR_EP_NID_AUTO)
node              238 net/qrtr/qrtr.c 	radix_tree_insert(&qrtr_nodes, nid, node);
node              239 net/qrtr/qrtr.c 	node->nid = nid;
node              253 net/qrtr/qrtr.c 	struct qrtr_node *node = ep->node;
node              319 net/qrtr/qrtr.c 	skb_queue_tail(&node->rx_queue, skb);
node              320 net/qrtr/qrtr.c 	schedule_work(&node->work);
node              364 net/qrtr/qrtr.c 	struct qrtr_node *node = container_of(work, struct qrtr_node, work);
node              370 net/qrtr/qrtr.c 	while ((skb = skb_dequeue(&node->rx_queue)) != NULL) {
node              382 net/qrtr/qrtr.c 		qrtr_node_assign(node, cb->src_node);
node              400 net/qrtr/qrtr.c 			pkt->client.node = cpu_to_le32(dst.sq_node);
node              403 net/qrtr/qrtr.c 			if (qrtr_node_enqueue(node, skb, QRTR_TYPE_RESUME_TX,
node              420 net/qrtr/qrtr.c 	struct qrtr_node *node;
node              425 net/qrtr/qrtr.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              426 net/qrtr/qrtr.c 	if (!node)
node              429 net/qrtr/qrtr.c 	INIT_WORK(&node->work, qrtr_node_rx_work);
node              430 net/qrtr/qrtr.c 	kref_init(&node->ref);
node              431 net/qrtr/qrtr.c 	mutex_init(&node->ep_lock);
node              432 net/qrtr/qrtr.c 	skb_queue_head_init(&node->rx_queue);
node              433 net/qrtr/qrtr.c 	node->nid = QRTR_EP_NID_AUTO;
node              434 net/qrtr/qrtr.c 	node->ep = ep;
node              436 net/qrtr/qrtr.c 	qrtr_node_assign(node, nid);
node              439 net/qrtr/qrtr.c 	list_add(&node->item, &qrtr_all_nodes);
node              441 net/qrtr/qrtr.c 	ep->node = node;
node              453 net/qrtr/qrtr.c 	struct qrtr_node *node = ep->node;
node              454 net/qrtr/qrtr.c 	struct sockaddr_qrtr src = {AF_QIPCRTR, node->nid, QRTR_PORT_CTRL};
node              459 net/qrtr/qrtr.c 	mutex_lock(&node->ep_lock);
node              460 net/qrtr/qrtr.c 	node->ep = NULL;
node              461 net/qrtr/qrtr.c 	mutex_unlock(&node->ep_lock);
node              470 net/qrtr/qrtr.c 	qrtr_node_release(node);
node              471 net/qrtr/qrtr.c 	ep->node = NULL;
node              516 net/qrtr/qrtr.c 		pkt->client.node = cpu_to_le32(ipc->us.sq_node);
node              669 net/qrtr/qrtr.c static int qrtr_local_enqueue(struct qrtr_node *node, struct sk_buff *skb,
node              698 net/qrtr/qrtr.c static int qrtr_bcast_enqueue(struct qrtr_node *node, struct sk_buff *skb,
node              705 net/qrtr/qrtr.c 	list_for_each_entry(node, &qrtr_all_nodes, item) {
node              710 net/qrtr/qrtr.c 		qrtr_node_enqueue(node, skbn, type, from, to);
node              727 net/qrtr/qrtr.c 	struct qrtr_node *node;
node              764 net/qrtr/qrtr.c 	node = NULL;
node              775 net/qrtr/qrtr.c 		node = qrtr_node_lookup(addr->sq_node);
node              776 net/qrtr/qrtr.c 		if (!node) {
node              809 net/qrtr/qrtr.c 	rc = enqueue_fn(node, skb, type, &ipc->us, addr);
node              814 net/qrtr/qrtr.c 	qrtr_node_release(node);
node               23 net/qrtr/qrtr.h 	struct qrtr_node *node;
node              401 net/rds/cong.c 	struct rb_node *node;
node              405 net/rds/cong.c 	while ((node = rb_first(&rds_cong_tree))) {
node              406 net/rds/cong.c 		map = rb_entry(node, struct rds_cong_map, m_rb_node);
node              289 net/rds/ib_rdma.c 	struct llist_node *node;
node              293 net/rds/ib_rdma.c 	node = llist_del_all(llist);
node              294 net/rds/ib_rdma.c 	while (node) {
node              295 net/rds/ib_rdma.c 		next = node->next;
node              296 net/rds/ib_rdma.c 		ibmr = llist_entry(node, struct rds_ib_mr, llnode);
node              298 net/rds/ib_rdma.c 		node = next;
node              131 net/rds/rdma.c 	struct rb_node *node;
node              136 net/rds/rdma.c 	while ((node = rb_first(&rs->rs_rdma_keys))) {
node              137 net/rds/rdma.c 		mr = rb_entry(node, struct rds_mr, r_rb_node);
node               60 net/rfkill/core.c 	struct list_head	node;
node              180 net/rfkill/core.c 	list_for_each_entry(rfkill, &rfkill_list, node) {
node              398 net/rfkill/core.c 	list_for_each_entry(rfkill, &rfkill_list, node) {
node              449 net/rfkill/core.c 	list_for_each_entry(rfkill, &rfkill_list, node)
node              945 net/rfkill/core.c 	INIT_LIST_HEAD(&rfkill->node);
node             1024 net/rfkill/core.c 	list_add_tail(&rfkill->node, &rfkill_list);
node             1065 net/rfkill/core.c 	list_del_init(&rfkill->node);
node             1088 net/rfkill/core.c 	list_del_init(&rfkill->node);
node             1124 net/rfkill/core.c 	list_for_each_entry(rfkill, &rfkill_list, node) {
node             1234 net/rfkill/core.c 		list_for_each_entry(rfkill, &rfkill_list, node)
node             1241 net/rfkill/core.c 		list_for_each_entry(rfkill, &rfkill_list, node)
node              678 net/rose/rose_route.c 	struct rose_node *node;
node              683 net/rose/rose_route.c 	for (node = rose_node_list; node != NULL; node = node->next) {
node              684 net/rose/rose_route.c 		if (rosecmpm(addr, &node->address, node->mask) == 0) {
node              685 net/rose/rose_route.c 			for (i = 0; i < node->count; i++) {
node              686 net/rose/rose_route.c 				if (node->neighbour[i]->restarted) {
node              687 net/rose/rose_route.c 					res = node->neighbour[i];
node              694 net/rose/rose_route.c 		for (node = rose_node_list; node != NULL; node = node->next) {
node              695 net/rose/rose_route.c 			if (rosecmpm(addr, &node->address, node->mask) == 0) {
node              696 net/rose/rose_route.c 				for (i = 0; i < node->count; i++) {
node              697 net/rose/rose_route.c 					if (!rose_ftimer_running(node->neighbour[i])) {
node              698 net/rose/rose_route.c 						res = node->neighbour[i];
node              137 net/sched/sch_htb.c 	struct rb_node		node[TC_HTB_NUMPRIO];	/* node for self or feed tree */
node              282 net/sched/sch_htb.c 		c = rb_entry(parent, struct htb_class, node[prio]);
node              289 net/sched/sch_htb.c 	rb_link_node(&cl->node[prio], parent, p);
node              290 net/sched/sch_htb.c 	rb_insert_color(&cl->node[prio], root);
node              383 net/sched/sch_htb.c 		if (hprio->ptr == cl->node + prio)
node              386 net/sched/sch_htb.c 		htb_safe_rb_erase(cl->node + prio, &hprio->row);
node              447 net/sched/sch_htb.c 			if (p->inner.clprio[prio].ptr == cl->node + prio) {
node              456 net/sched/sch_htb.c 			htb_safe_rb_erase(cl->node + prio,
node              747 net/sched/sch_htb.c 		    rb_entry(n, struct htb_class, node[prio]);
node              807 net/sched/sch_htb.c 			cl = rb_entry(*sp->pptr, struct htb_class, node[prio]);
node             1393 net/sched/sch_htb.c 			RB_CLEAR_NODE(&cl->node[prio]);
node              766 net/sctp/input.c 	hlist_add_head(&epb->node, &head->chain);
node              800 net/sctp/input.c 	hlist_del_init(&epb->node);
node              913 net/sctp/input.c 	.head_offset		= offsetof(struct sctp_transport, node),
node              948 net/sctp/input.c 	rhl_for_each_entry_rcu(transport, tmp, list, node)
node              956 net/sctp/input.c 				  &t->node, sctp_hash_params);
node              968 net/sctp/input.c 	rhltable_remove(&sctp_transport_hashtable, &t->node,
node              989 net/sctp/input.c 	rhl_for_each_entry_rcu(t, tmp, list, node) {
node             1019 net/sctp/input.c 	rhl_for_each_entry_rcu(t, tmp, list, node)
node             8545 net/sctp/socket.c 		hlist_add_head(&pp->node, &head->chain);
node             8554 net/sctp/socket.c 		__hlist_del(&pp->node);
node              462 net/smc/smc_core.c 	struct rb_node *node;
node              473 net/smc/smc_core.c 	node = rb_first(&lgr->conns_all);
node              474 net/smc/smc_core.c 	while (node) {
node              475 net/smc/smc_core.c 		conn = rb_entry(node, struct smc_connection, alert_node);
node              485 net/smc/smc_core.c 		node = rb_first(&lgr->conns_all);
node              261 net/smc/smc_core.h 	struct rb_node *node;
node              263 net/smc/smc_core.h 	node = lgr->conns_all.rb_node;
node              264 net/smc/smc_core.h 	while (node) {
node              265 net/smc/smc_core.h 		struct smc_connection *cur = rb_entry(node,
node              269 net/smc/smc_core.h 			node = node->rb_left;
node              272 net/smc/smc_core.h 				node = node->rb_right;
node              113 net/sunrpc/svc.c 	unsigned int node;
node              123 net/sunrpc/svc.c 	node = first_online_node;
node              124 net/sunrpc/svc.c 	if (nr_cpus_node(node) > 2) {
node              198 net/sunrpc/svc.c 	unsigned int node;
node              205 net/sunrpc/svc.c 	for_each_node_with_cpus(node) {
node              208 net/sunrpc/svc.c 		m->to_pool[node] = pidx;
node              209 net/sunrpc/svc.c 		m->pool_to[pidx] = node;
node              306 net/sunrpc/svc.c 	unsigned int node = m->pool_to[pidx];
node              319 net/sunrpc/svc.c 		set_cpus_allowed_ptr(task, cpumask_of(node));
node              324 net/sunrpc/svc.c 		set_cpus_allowed_ptr(task, cpumask_of_node(node));
node              565 net/sunrpc/svc.c svc_init_buffer(struct svc_rqst *rqstp, unsigned int size, int node)
node              581 net/sunrpc/svc.c 		struct page *p = alloc_pages_node(node, GFP_KERNEL, 0);
node              604 net/sunrpc/svc.c svc_rqst_alloc(struct svc_serv *serv, struct svc_pool *pool, int node)
node              608 net/sunrpc/svc.c 	rqstp = kzalloc_node(sizeof(*rqstp), GFP_KERNEL, node);
node              617 net/sunrpc/svc.c 	rqstp->rq_argp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node);
node              621 net/sunrpc/svc.c 	rqstp->rq_resp = kmalloc_node(serv->sv_xdrsize, GFP_KERNEL, node);
node              625 net/sunrpc/svc.c 	if (!svc_init_buffer(rqstp, serv->sv_max_mesg, node))
node              636 net/sunrpc/svc.c svc_prepare_thread(struct svc_serv *serv, struct svc_pool *pool, int node)
node              640 net/sunrpc/svc.c 	rqstp = svc_rqst_alloc(serv, pool, node);
node              714 net/sunrpc/svc.c 	int node;
node              720 net/sunrpc/svc.c 		node = svc_pool_map_get_node(chosen_pool->sp_id);
node              721 net/sunrpc/svc.c 		rqstp = svc_prepare_thread(serv, chosen_pool, node);
node              727 net/sunrpc/svc.c 					      node, "%s", serv->sv_name);
node              175 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	struct llist_node *node;
node              177 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	while ((node = llist_del_first(&rdma->sc_recv_ctxts))) {
node              178 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 		ctxt = llist_entry(node, struct svc_rdma_recv_ctxt, rc_node);
node              187 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	struct llist_node *node;
node              189 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	node = llist_del_first(&rdma->sc_recv_ctxts);
node              190 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	if (!node)
node              192 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c 	ctxt = llist_entry(node, struct svc_rdma_recv_ctxt, rc_node);
node             1107 net/sunrpc/xprtrdma/verbs.c 	struct llist_node *node;
node             1110 net/sunrpc/xprtrdma/verbs.c 	node = llist_del_first(&buf->rb_free_reps);
node             1111 net/sunrpc/xprtrdma/verbs.c 	if (!node)
node             1113 net/sunrpc/xprtrdma/verbs.c 	return llist_entry(node, struct rpcrdma_rep, rr_node);
node               70 net/tipc/addr.h static inline int tipc_node2scope(u32 node)
node               72 net/tipc/addr.h 	return node ? TIPC_NODE_SCOPE : TIPC_CLUSTER_SCOPE;
node              292 net/tipc/bcast.c 		dnode = dst->node;
node              725 net/tipc/bcast.c void tipc_nlist_add(struct tipc_nlist *nl, u32 node)
node              727 net/tipc/bcast.c 	if (node == nl->self)
node              729 net/tipc/bcast.c 	else if (tipc_dest_push(&nl->list, node, 0))
node              733 net/tipc/bcast.c void tipc_nlist_del(struct tipc_nlist *nl, u32 node)
node              735 net/tipc/bcast.c 	if (node == nl->self)
node              737 net/tipc/bcast.c 	else if (tipc_dest_del(&nl->list, node, 0))
node              777 net/tipc/bcast.c 	u32 node, port;
node              788 net/tipc/bcast.c 	node = msg_orignode(hdr);
node              789 net/tipc/bcast.c 	if (node == tipc_own_addr(net))
node              797 net/tipc/bcast.c 		if (msg_orignode(_hdr) != node)
node              842 net/tipc/bcast.c 		if (msg_orignode(_hdr) != node)
node               64 net/tipc/bcast.h void tipc_nlist_add(struct tipc_nlist *nl, u32 node);
node               65 net/tipc/bcast.h void tipc_nlist_del(struct tipc_nlist *nl, u32 node);
node               68 net/tipc/group.c 	u32 node;
node              232 net/tipc/group.c 						  u32 node, u32 port)
node              235 net/tipc/group.c 	u64 nkey, key = (u64)node << 32 | port;
node              240 net/tipc/group.c 		nkey = (u64)m->node << 32 | m->port;
node              252 net/tipc/group.c 						u32 node, u32 port)
node              256 net/tipc/group.c 	m = tipc_group_find_member(grp, node, port);
node              263 net/tipc/group.c 						u32 node)
node              270 net/tipc/group.c 		if (m->node == node)
node              279 net/tipc/group.c 	u64 nkey, key = (u64)m->node << 32 | m->port;
node              288 net/tipc/group.c 		nkey = (u64)tmp->node << 32 | tmp->port;
node              301 net/tipc/group.c 						    u32 node, u32 port,
node              313 net/tipc/group.c 	m->node = node;
node              319 net/tipc/group.c 	tipc_nlist_add(&grp->dests, m->node);
node              324 net/tipc/group.c void tipc_group_add_member(struct tipc_group *grp, u32 node,
node              327 net/tipc/group.c 	tipc_group_create_member(grp, node, port, instance, MBR_PUBLISHED);
node              345 net/tipc/group.c 	if (!tipc_group_find_node(grp, m->node))
node              346 net/tipc/group.c 		tipc_nlist_del(&grp->dests, m->node);
node              460 net/tipc/group.c 	return tipc_group_cong(grp, m->node, m->port, len, &m);
node              497 net/tipc/group.c 	u32 node, port;
node              504 net/tipc/group.c 	node =  msg_orignode(hdr);
node              510 net/tipc/group.c 	m = tipc_group_find_member(grp, node, port);
node              574 net/tipc/group.c 		tipc_group_update_rcv_win(grp, blks, node, port, xmitq);
node              581 net/tipc/group.c void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node,
node              590 net/tipc/group.c 	m = tipc_group_find_member(grp, node, port);
node              675 net/tipc/group.c 	evt.port.node = m->node;
node              681 net/tipc/group.c 			      GROUP_H_SIZE, sizeof(evt), dnode, m->node,
node              704 net/tipc/group.c 			      m->node, tipc_own_addr(grp->net),
node              738 net/tipc/group.c 	u32 node = msg_orignode(hdr);
node              746 net/tipc/group.c 	if (grp->scope == TIPC_NODE_SCOPE && node != tipc_own_addr(grp->net))
node              749 net/tipc/group.c 	m = tipc_group_find_member(grp, node, port);
node              754 net/tipc/group.c 			m = tipc_group_create_member(grp, node, port,
node              857 net/tipc/group.c 	u32 node = evt->port.node;
node              869 net/tipc/group.c 	if (!grp->loopback && node == self && port == grp->portid)
node              872 net/tipc/group.c 	m = tipc_group_find_member(grp, node, port);
node              878 net/tipc/group.c 			m = tipc_group_create_member(grp, node, port, instance,
node              909 net/tipc/group.c 		if (!tipc_node_is_up(net, node))
node               50 net/tipc/group.h void tipc_group_add_member(struct tipc_group *grp, u32 node,
node               71 net/tipc/group.h void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node,
node               46 net/tipc/name_distr.c 	u32 node;
node              214 net/tipc/name_distr.c 				     publ->node, publ->key);
node              222 net/tipc/name_distr.c 		       publ->type, publ->lower, publ->node, publ->port,
node              240 net/tipc/name_distr.c 		if (e->node != addr)
node              265 net/tipc/name_distr.c 				u32 node, u32 dtype)
node              276 net/tipc/name_distr.c 					     TIPC_CLUSTER_SCOPE, node,
node              279 net/tipc/name_distr.c 			tipc_node_subscribe(net, &p->binding_node, node);
node              284 net/tipc/name_distr.c 					     upper, node, key);
node              286 net/tipc/name_distr.c 			tipc_node_unsubscribe(net, &p->binding_node, node);
node              291 net/tipc/name_distr.c 				    type, lower, node);
node              307 net/tipc/name_distr.c 	u32 node;
node              318 net/tipc/name_distr.c 		node = msg_orignode(msg);
node              320 net/tipc/name_distr.c 			tipc_update_nametbl(net, item, node, mtype);
node              345 net/tipc/name_distr.c 		publ->node = self;
node              347 net/tipc/name_distr.c 		publ->node = self;
node               93 net/tipc/name_table.c 					    u32 scope, u32 node, u32 port,
node              105 net/tipc/name_table.c 	publ->node = node;
node              223 net/tipc/name_table.c 						    u32 node, u32 port,
node              239 net/tipc/name_table.c 		if (p->key == key && (!p->node || p->node == node))
node              244 net/tipc/name_table.c 	p = tipc_publ_create(type, lower, upper, scope, node, port, key);
node              247 net/tipc/name_table.c 	if (in_own_node(net, node))
node              254 net/tipc/name_table.c 					p->port, p->node, p->scope, first);
node              266 net/tipc/name_table.c 						    u32 node, u32 key)
node              271 net/tipc/name_table.c 		if (p->key != key || (node && node != p->node))
node              316 net/tipc/name_table.c 						p->node, p->scope, first);
node              338 net/tipc/name_table.c 					     u32 scope, u32 node,
node              358 net/tipc/name_table.c 				     scope, node, port, key);
node              365 net/tipc/name_table.c 					     u32 node, u32 key)
node              380 net/tipc/name_table.c 	p = tipc_service_remove_publ(sr, node, key);
node              388 net/tipc/name_table.c 					p->port, node, p->scope, last);
node              433 net/tipc/name_table.c 	u32 node = 0;
node              465 net/tipc/name_table.c 	node = p->node;
node              470 net/tipc/name_table.c 	*dnode = node;
node              498 net/tipc/name_table.c 		if (p->port == exclude && p->node == self)
node              500 net/tipc/name_table.c 		tipc_dest_push(dsts, p->node, p->port);
node              571 net/tipc/name_table.c 			tipc_nlist_add(nodes, p->node);
node              600 net/tipc/name_table.c 			tipc_group_add_member(grp, p->node, p->port, p->lower);
node              764 net/tipc/name_table.c 			tipc_service_remove_publ(sr, p->node, p->key);
node              848 net/tipc/name_table.c 		if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_NODE, p->node))
node              983 net/tipc/name_table.c struct tipc_dest *tipc_dest_find(struct list_head *l, u32 node, u32 port)
node              988 net/tipc/name_table.c 		if (dst->node == node && dst->port == port)
node              994 net/tipc/name_table.c bool tipc_dest_push(struct list_head *l, u32 node, u32 port)
node              998 net/tipc/name_table.c 	if (tipc_dest_find(l, node, port))
node             1004 net/tipc/name_table.c 	dst->node = node;
node             1010 net/tipc/name_table.c bool tipc_dest_pop(struct list_head *l, u32 *node, u32 *port)
node             1019 net/tipc/name_table.c 	if (node)
node             1020 net/tipc/name_table.c 		*node = dst->node;
node             1026 net/tipc/name_table.c bool tipc_dest_del(struct list_head *l, u32 node, u32 port)
node             1030 net/tipc/name_table.c 	dst = tipc_dest_find(l, node, port);
node               79 net/tipc/name_table.h 	u32 node;
node              109 net/tipc/name_table.h u32 tipc_nametbl_translate(struct net *net, u32 type, u32 instance, u32 *node);
node              126 net/tipc/name_table.h 					     u32 node, u32 ref, u32 key);
node              129 net/tipc/name_table.h 					     u32 node, u32 key);
node              138 net/tipc/name_table.h 	u32 node;
node              141 net/tipc/name_table.h struct tipc_dest *tipc_dest_find(struct list_head *l, u32 node, u32 port);
node              142 net/tipc/name_table.h bool tipc_dest_push(struct list_head *l, u32 node, u32 port);
node              143 net/tipc/name_table.h bool tipc_dest_pop(struct list_head *l, u32 *node, u32 *port);
node              144 net/tipc/name_table.h bool tipc_dest_del(struct list_head *l, u32 node, u32 port);
node              871 net/tipc/netlink_compat.c 	u32 node, depth, type, lowbound, upbound;
node              922 net/tipc/netlink_compat.c 	node = nla_get_u32(publ[TIPC_NLA_PUBL_NODE]);
node              923 net/tipc/netlink_compat.c 	sprintf(port_str, "<%u.%u.%u:%u>", tipc_zone(node), tipc_cluster(node),
node              924 net/tipc/netlink_compat.c 		tipc_node(node), nla_get_u32(publ[TIPC_NLA_PUBL_REF]));
node             1028 net/tipc/netlink_compat.c 		u32 node;
node             1038 net/tipc/netlink_compat.c 		node = nla_get_u32(con[TIPC_NLA_CON_NODE]);
node             1040 net/tipc/netlink_compat.c 				 tipc_zone(node),
node             1041 net/tipc/netlink_compat.c 				 tipc_cluster(node),
node             1042 net/tipc/netlink_compat.c 				 tipc_node(node),
node             1086 net/tipc/netlink_compat.c 	struct nlattr *node[TIPC_NLA_NODE_MAX + 1];
node             1092 net/tipc/netlink_compat.c 	err = nla_parse_nested_deprecated(node, TIPC_NLA_NODE_MAX,
node             1097 net/tipc/netlink_compat.c 	node_info.addr = htonl(nla_get_u32(node[TIPC_NLA_NODE_ADDR]));
node             1098 net/tipc/netlink_compat.c 	node_info.up = htonl(nla_get_flag(node[TIPC_NLA_NODE_UP]));
node              161 net/tipc/node.c static void tipc_node_delete(struct tipc_node *node);
node              166 net/tipc/node.c static void tipc_node_put(struct tipc_node *node);
node              168 net/tipc/node.c static void tipc_node_delete_from_list(struct tipc_node *node);
node              246 net/tipc/node.c static void tipc_node_put(struct tipc_node *node)
node              248 net/tipc/node.c 	kref_put(&node->kref, tipc_node_kref_release);
node              251 net/tipc/node.c static void tipc_node_get(struct tipc_node *node)
node              253 net/tipc/node.c 	kref_get(&node->kref);
node              262 net/tipc/node.c 	struct tipc_node *node;
node              266 net/tipc/node.c 	hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) {
node              267 net/tipc/node.c 		if (node->addr != addr)
node              269 net/tipc/node.c 		if (!kref_get_unless_zero(&node->kref))
node              270 net/tipc/node.c 			node = NULL;
node              274 net/tipc/node.c 	return node;
node              466 net/tipc/node.c static void tipc_node_delete_from_list(struct tipc_node *node)
node              468 net/tipc/node.c 	list_del_rcu(&node->list);
node              469 net/tipc/node.c 	hlist_del_rcu(&node->hash);
node              470 net/tipc/node.c 	tipc_node_put(node);
node              473 net/tipc/node.c static void tipc_node_delete(struct tipc_node *node)
node              475 net/tipc/node.c 	trace_tipc_node_delete(node, true, " ");
node              476 net/tipc/node.c 	tipc_node_delete_from_list(node);
node              478 net/tipc/node.c 	del_timer_sync(&node->timer);
node              479 net/tipc/node.c 	tipc_node_put(node);
node              485 net/tipc/node.c 	struct tipc_node *node, *t_node;
node              488 net/tipc/node.c 	list_for_each_entry_safe(node, t_node, &tn->node_list, list)
node              489 net/tipc/node.c 		tipc_node_delete(node);
node              531 net/tipc/node.c 	struct tipc_node *node;
node              538 net/tipc/node.c 	node = tipc_node_find(net, dnode);
node              539 net/tipc/node.c 	if (!node) {
node              552 net/tipc/node.c 	tipc_node_write_lock(node);
node              553 net/tipc/node.c 	list_add_tail(&conn->list, &node->conn_sks);
node              554 net/tipc/node.c 	tipc_node_write_unlock(node);
node              556 net/tipc/node.c 	tipc_node_put(node);
node              562 net/tipc/node.c 	struct tipc_node *node;
node              568 net/tipc/node.c 	node = tipc_node_find(net, dnode);
node              569 net/tipc/node.c 	if (!node)
node              572 net/tipc/node.c 	tipc_node_write_lock(node);
node              573 net/tipc/node.c 	list_for_each_entry_safe(conn, safe, &node->conn_sks, list) {
node              579 net/tipc/node.c 	tipc_node_write_unlock(node);
node              580 net/tipc/node.c 	tipc_node_put(node);
node              583 net/tipc/node.c static void  tipc_node_clear_links(struct tipc_node *node)
node              588 net/tipc/node.c 		struct tipc_link_entry *le = &node->links[i];
node              593 net/tipc/node.c 			node->link_cnt--;
node             1373 net/tipc/node.c 	struct tipc_node *node = tipc_node_find(net, addr);
node             1375 net/tipc/node.c 	if (!node)
node             1381 net/tipc/node.c 	tipc_node_read_lock(node);
node             1382 net/tipc/node.c 	link = node->links[bearer_id].link;
node             1387 net/tipc/node.c 	tipc_node_read_unlock(node);
node             1389 net/tipc/node.c 	tipc_node_put(node);
node             1394 net/tipc/node.c static int __tipc_nl_add_node(struct tipc_nl_msg *msg, struct tipc_node *node)
node             1408 net/tipc/node.c 	if (nla_put_u32(msg->skb, TIPC_NLA_NODE_ADDR, node->addr))
node             1410 net/tipc/node.c 	if (node_is_up(node))
node             1985 net/tipc/node.c 	struct tipc_node *node;
node             1997 net/tipc/node.c 		node = tipc_node_find(net, last_addr);
node             1998 net/tipc/node.c 		if (!node) {
node             2010 net/tipc/node.c 		tipc_node_put(node);
node             2013 net/tipc/node.c 	list_for_each_entry_rcu(node, &tn->node_list, list) {
node             2015 net/tipc/node.c 			if (node->addr == last_addr)
node             2021 net/tipc/node.c 		tipc_node_read_lock(node);
node             2022 net/tipc/node.c 		err = __tipc_nl_add_node(&msg, node);
node             2024 net/tipc/node.c 			last_addr = node->addr;
node             2025 net/tipc/node.c 			tipc_node_read_unlock(node);
node             2029 net/tipc/node.c 		tipc_node_read_unlock(node);
node             2085 net/tipc/node.c 	struct tipc_node *node;
node             2109 net/tipc/node.c 	node = tipc_node_find_by_name(net, name, &bearer_id);
node             2110 net/tipc/node.c 	if (!node)
node             2113 net/tipc/node.c 	tipc_node_read_lock(node);
node             2115 net/tipc/node.c 	link = node->links[bearer_id].link;
node             2152 net/tipc/node.c 	tipc_node_read_unlock(node);
node             2153 net/tipc/node.c 	tipc_bearer_xmit(net, bearer_id, &xmitq, &node->links[bearer_id].maddr);
node             2192 net/tipc/node.c 		struct tipc_node *node;
node             2195 net/tipc/node.c 		node = tipc_node_find_by_name(net, name, &bearer_id);
node             2196 net/tipc/node.c 		if (!node) {
node             2201 net/tipc/node.c 		tipc_node_read_lock(node);
node             2202 net/tipc/node.c 		link = node->links[bearer_id].link;
node             2204 net/tipc/node.c 			tipc_node_read_unlock(node);
node             2210 net/tipc/node.c 		tipc_node_read_unlock(node);
node             2228 net/tipc/node.c 	struct tipc_node *node;
node             2254 net/tipc/node.c 	node = tipc_node_find_by_name(net, link_name, &bearer_id);
node             2255 net/tipc/node.c 	if (!node)
node             2258 net/tipc/node.c 	le = &node->links[bearer_id];
node             2259 net/tipc/node.c 	tipc_node_read_lock(node);
node             2261 net/tipc/node.c 	link = node->links[bearer_id].link;
node             2264 net/tipc/node.c 		tipc_node_read_unlock(node);
node             2269 net/tipc/node.c 	tipc_node_read_unlock(node);
node             2275 net/tipc/node.c 				    struct tipc_node *node, u32 *prev_link)
node             2283 net/tipc/node.c 		if (!node->links[i].link)
node             2287 net/tipc/node.c 					 node->links[i].link, NLM_F_MULTI);
node             2300 net/tipc/node.c 	struct tipc_node *node;
node             2316 net/tipc/node.c 		node = tipc_node_find(net, prev_node);
node             2317 net/tipc/node.c 		if (!node) {
node             2327 net/tipc/node.c 		tipc_node_put(node);
node             2329 net/tipc/node.c 		list_for_each_entry_continue_rcu(node, &tn->node_list,
node             2331 net/tipc/node.c 			tipc_node_read_lock(node);
node             2332 net/tipc/node.c 			err = __tipc_nl_add_node_links(net, &msg, node,
node             2334 net/tipc/node.c 			tipc_node_read_unlock(node);
node             2338 net/tipc/node.c 			prev_node = node->addr;
node             2345 net/tipc/node.c 		list_for_each_entry_rcu(node, &tn->node_list, list) {
node             2346 net/tipc/node.c 			tipc_node_read_lock(node);
node             2347 net/tipc/node.c 			err = __tipc_nl_add_node_links(net, &msg, node,
node             2349 net/tipc/node.c 			tipc_node_read_unlock(node);
node             2353 net/tipc/node.c 			prev_node = node->addr;
node             2533 net/tipc/node.c u32 tipc_node_get_addr(struct tipc_node *node)
node             2535 net/tipc/node.c 	return (node) ? node->addr : 0;
node               74 net/tipc/node.h u32 tipc_node_get_addr(struct tipc_node *node);
node               83 net/tipc/node.h int tipc_node_get_linkname(struct net *net, u32 bearer_id, u32 node,
node              115 net/tipc/socket.c 	struct rhash_head node;
node              689 net/tipc/socket.c 		addr->addr.id.node = tsk_peer_node(tsk);
node              692 net/tipc/socket.c 		addr->addr.id.node = tipc_own_addr(sock_net(sk));
node              897 net/tipc/socket.c 	u32 node, port;
node              900 net/tipc/socket.c 	node = dest->addr.id.node;
node              902 net/tipc/socket.c 	if (!port && !node)
node              907 net/tipc/socket.c 				!tipc_dest_find(&tsk->cong_links, node, 0) &&
node              909 net/tipc/socket.c 				!tipc_group_cong(tsk->group, node, port, blks,
node              917 net/tipc/socket.c 	rc = tipc_send_group_msg(net, tsk, m, mb, node, port, dlen);
node              944 net/tipc/socket.c 	u32 node, port, exclude;
node              967 net/tipc/socket.c 			tipc_dest_pop(&dsts, &node, &port);
node              968 net/tipc/socket.c 			cong = tipc_group_cong(tsk->group, node, port, blks,
node              982 net/tipc/socket.c 		if (likely(!cong && !tipc_dest_find(cong_links, node, 0)))
node              987 net/tipc/socket.c 					!tipc_dest_find(cong_links, node, 0) &&
node              989 net/tipc/socket.c 					!tipc_group_cong(tsk->group, node, port,
node             1002 net/tipc/socket.c 	rc = tipc_send_group_msg(net, tsk, m, mbr, node, port, dlen);
node             1116 net/tipc/socket.c 		tipc_dest_pop(&dsts, &dest->addr.id.node, &dest->addr.id.ref);
node             1367 net/tipc/socket.c 		dnode = dest->addr.id.node;
node             1565 net/tipc/socket.c 	srcaddr->sock.addr.id.node = msg_orignode(hdr);
node             2836 net/tipc/socket.c 		if (!rhashtable_lookup_insert_fast(&tn->sk_rht, &tsk->node,
node             2850 net/tipc/socket.c 	if (!rhashtable_remove_fast(&tn->sk_rht, &tsk->node, tsk_rht_params)) {
node             2858 net/tipc/socket.c 	.head_offset = offsetof(struct tipc_sock, node),
node             3148 net/tipc/socket.c 	tsk1->peer.addr.id.node = onode;
node             3153 net/tipc/socket.c 	tsk2->peer.addr.id.node = onode;
node               43 net/tipc/subscr.c 				u32 event, u32 port, u32 node)
node               53 net/tipc/subscr.c 	tipc_evt_write(evt, port.node, node);
node               77 net/tipc/subscr.c 			     u32 event, u32 port, u32 node,
node               99 net/tipc/subscr.c 			    event, port, node);
node               78 net/tipc/subscr.h 			     u32 event, u32 port, u32 node,
node               50 net/x25/x25_forward.c 		x25_frwd = list_entry(entry, struct x25_forward, node);
node               69 net/x25/x25_forward.c 		list_add(&new_frwd->node, &x25_forward_list);
node              103 net/x25/x25_forward.c 		frwd = list_entry(entry, struct x25_forward, node);
node              140 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
node              142 net/x25/x25_forward.c 			list_del(&fwd->node);
node              158 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
node              160 net/x25/x25_forward.c 			list_del(&fwd->node);
node              267 net/x25/x25_link.c 	list_add(&nb->node, &x25_neigh_list);
node              283 net/x25/x25_link.c 	if (nb->node.next) {
node              284 net/x25/x25_link.c 		list_del(&nb->node);
node              300 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node              321 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node              400 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node               45 net/x25/x25_proc.c 	struct x25_route *rt = list_entry(v, struct x25_route, node);
node              133 net/x25/x25_proc.c 	struct x25_forward *f = list_entry(v, struct x25_forward, node);
node               36 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node               55 net/x25/x25_route.c 	list_add(&rt->node, &x25_route_list);
node               71 net/x25/x25_route.c 	if (rt->node.next) {
node               72 net/x25/x25_route.c 		list_del(&rt->node);
node               87 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              112 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              157 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              217 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              497 net/xdp/xsk.c  	struct xsk_map_node *node;
node              502 net/xdp/xsk.c  	node = list_first_entry_or_null(&xs->map_list, struct xsk_map_node,
node              503 net/xdp/xsk.c  					node);
node              504 net/xdp/xsk.c  	if (node) {
node              505 net/xdp/xsk.c  		WARN_ON(xsk_map_inc(node->map));
node              506 net/xdp/xsk.c  		map = node->map;
node              507 net/xdp/xsk.c  		*map_entry = node->map_entry;
node               59 net/xfrm/xfrm_policy.c 	struct rb_node node;
node              770 net/xfrm/xfrm_policy.c static void xfrm_pol_inexact_node_init(struct xfrm_pol_inexact_node *node,
node              773 net/xfrm/xfrm_policy.c 	node->addr = *addr;
node              774 net/xfrm/xfrm_policy.c 	node->prefixlen = prefixlen;
node              780 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_node *node;
node              782 net/xfrm/xfrm_policy.c 	node = kzalloc(sizeof(*node), GFP_ATOMIC);
node              783 net/xfrm/xfrm_policy.c 	if (node)
node              784 net/xfrm/xfrm_policy.c 		xfrm_pol_inexact_node_init(node, addr, prefixlen);
node              786 net/xfrm/xfrm_policy.c 	return node;
node              893 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_node *node;
node              906 net/xfrm/xfrm_policy.c 		node = rb_entry(*p, struct xfrm_pol_inexact_node, node);
node              908 net/xfrm/xfrm_policy.c 		prefixlen = min(node->prefixlen, n->prefixlen);
node              910 net/xfrm/xfrm_policy.c 		delta = xfrm_policy_addr_delta(&n->addr, &node->addr,
node              917 net/xfrm/xfrm_policy.c 			bool same_prefixlen = node->prefixlen == n->prefixlen;
node              925 net/xfrm/xfrm_policy.c 			node->prefixlen = prefixlen;
node              927 net/xfrm/xfrm_policy.c 			xfrm_policy_inexact_list_reinsert(net, node, family);
node              936 net/xfrm/xfrm_policy.c 			n = node;
node              941 net/xfrm/xfrm_policy.c 	rb_link_node_rcu(&n->node, parent, p);
node              942 net/xfrm/xfrm_policy.c 	rb_insert_color(&n->node, new);
node              951 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_node *node;
node              960 net/xfrm/xfrm_policy.c 		node = rb_entry(rnode, struct xfrm_pol_inexact_node, node);
node              961 net/xfrm/xfrm_policy.c 		rb_erase(&node->node, &v->root);
node              962 net/xfrm/xfrm_policy.c 		xfrm_policy_inexact_node_reinsert(net, node, &n->root,
node              982 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_node *node;
node              989 net/xfrm/xfrm_policy.c 		node = rb_entry(*p, struct xfrm_pol_inexact_node, node);
node              991 net/xfrm/xfrm_policy.c 		delta = xfrm_policy_addr_delta(addr, &node->addr,
node              992 net/xfrm/xfrm_policy.c 					       node->prefixlen,
node              994 net/xfrm/xfrm_policy.c 		if (delta == 0 && prefixlen >= node->prefixlen) {
node              996 net/xfrm/xfrm_policy.c 			return node;
node             1004 net/xfrm/xfrm_policy.c 		if (prefixlen < node->prefixlen) {
node             1005 net/xfrm/xfrm_policy.c 			delta = xfrm_policy_addr_delta(addr, &node->addr,
node             1016 net/xfrm/xfrm_policy.c 			rb_erase(&node->node, root);
node             1019 net/xfrm/xfrm_policy.c 				xfrm_pol_inexact_node_init(node, addr,
node             1021 net/xfrm/xfrm_policy.c 				cached = node;
node             1027 net/xfrm/xfrm_policy.c 				xfrm_policy_inexact_node_merge(net, node,
node             1029 net/xfrm/xfrm_policy.c 				kfree_rcu(node, rcu);
node             1038 net/xfrm/xfrm_policy.c 	node = cached;
node             1039 net/xfrm/xfrm_policy.c 	if (!node) {
node             1040 net/xfrm/xfrm_policy.c 		node = xfrm_pol_inexact_node_alloc(addr, prefixlen);
node             1041 net/xfrm/xfrm_policy.c 		if (!node)
node             1045 net/xfrm/xfrm_policy.c 	rb_link_node_rcu(&node->node, parent, p);
node             1046 net/xfrm/xfrm_policy.c 	rb_insert_color(&node->node, root);
node             1048 net/xfrm/xfrm_policy.c 	return node;
node             1053 net/xfrm/xfrm_policy.c 	struct xfrm_pol_inexact_node *node;
node             1057 net/xfrm/xfrm_policy.c 		node = rb_entry(rn, struct xfrm_pol_inexact_node, node);
node             1059 net/xfrm/xfrm_policy.c 		xfrm_policy_inexact_gc_tree(&node->root, rm);
node             1062 net/xfrm/xfrm_policy.c 		if (!hlist_empty(&node->hhead) || !RB_EMPTY_ROOT(&node->root)) {
node             1067 net/xfrm/xfrm_policy.c 		rb_erase(&node->node, r);
node             1068 net/xfrm/xfrm_policy.c 		kfree_rcu(node, rcu);
node             1917 net/xfrm/xfrm_policy.c 		struct xfrm_pol_inexact_node *node;
node             1920 net/xfrm/xfrm_policy.c 		node = rb_entry(parent, struct xfrm_pol_inexact_node, node);
node             1922 net/xfrm/xfrm_policy.c 		delta = xfrm_policy_addr_delta(addr, &node->addr,
node             1923 net/xfrm/xfrm_policy.c 					       node->prefixlen, family);
node             1932 net/xfrm/xfrm_policy.c 		return node;
node              132 samples/bpf/test_lru_dist.c 	struct pfect_lru_node *node = NULL;
node              136 samples/bpf/test_lru_dist.c 	if (!bpf_map_lookup_elem(lru->map_fd, &key, &node)) {
node              137 samples/bpf/test_lru_dist.c 		if (node) {
node              138 samples/bpf/test_lru_dist.c 			list_move(&node->list, &lru->list);
node              145 samples/bpf/test_lru_dist.c 		node =  &lru->free_nodes[lru->cur_size++];
node              146 samples/bpf/test_lru_dist.c 		INIT_LIST_HEAD(&node->list);
node              150 samples/bpf/test_lru_dist.c 		node = list_last_entry(&lru->list,
node              153 samples/bpf/test_lru_dist.c 		bpf_map_update_elem(lru->map_fd, &node->key, &null_node, BPF_EXIST);
node              156 samples/bpf/test_lru_dist.c 	node->key = key;
node              157 samples/bpf/test_lru_dist.c 	list_move(&node->list, &lru->list);
node              161 samples/bpf/test_lru_dist.c 		assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_EXIST));
node              164 samples/bpf/test_lru_dist.c 		assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_NOEXIST));
node              539 samples/qmi/qmi_sample_client.c 	struct sockaddr_qrtr sq = { AF_QIPCRTR, service->node, service->port };
node              232 samples/v4l/v4l2-pci-skeleton.c 	struct skel_buffer *buf, *node;
node              236 samples/v4l/v4l2-pci-skeleton.c 	list_for_each_entry_safe(buf, node, &skel->buf_list, list) {
node              177 samples/vfio-mdev/mdpy-fb.c 	pci_info(pdev, "fb%d registered\n", info->node);
node               29 scripts/dtc/checks.c typedef void (*check_fn)(struct check *c, struct dt_info *dti, struct node *node);
node               62 scripts/dtc/checks.c 					   struct node *node,
node               76 scripts/dtc/checks.c 	else if (node && node->srcpos)
node               77 scripts/dtc/checks.c 		pos = node->srcpos;
node               92 scripts/dtc/checks.c 	if (node) {
node               94 scripts/dtc/checks.c 			xasprintf_append(&str, "%s:%s: ", node->fullpath, prop->name);
node               96 scripts/dtc/checks.c 			xasprintf_append(&str, "%s: ", node->fullpath);
node              106 scripts/dtc/checks.c 		pos = node->srcpos;
node              119 scripts/dtc/checks.c #define FAIL(c, dti, node, ...)						\
node              123 scripts/dtc/checks.c 		check_msg((c), dti, node, NULL, __VA_ARGS__);		\
node              126 scripts/dtc/checks.c #define FAIL_PROP(c, dti, node, prop, ...)				\
node              130 scripts/dtc/checks.c 		check_msg((c), dti, node, prop, __VA_ARGS__);		\
node              134 scripts/dtc/checks.c static void check_nodes_props(struct check *c, struct dt_info *dti, struct node *node)
node              136 scripts/dtc/checks.c 	struct node *child;
node              138 scripts/dtc/checks.c 	TRACE(c, "%s", node->fullpath);
node              140 scripts/dtc/checks.c 		c->fn(c, dti, node);
node              142 scripts/dtc/checks.c 	for_each_child(node, child)
node              148 scripts/dtc/checks.c 	struct node *dt = dti->dt;
node              192 scripts/dtc/checks.c 				     struct node *node)
node              194 scripts/dtc/checks.c 	FAIL(c, dti, node, "always_fail check");
node              199 scripts/dtc/checks.c 			    struct node *node)
node              204 scripts/dtc/checks.c 	prop = get_property(node, propname);
node              209 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "property is not a string");
node              217 scripts/dtc/checks.c 				 struct node *node)
node              224 scripts/dtc/checks.c 	prop = get_property(node, propname);
node              233 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "property is not a string list");
node              246 scripts/dtc/checks.c 			  struct node *node)
node              251 scripts/dtc/checks.c 	prop = get_property(node, propname);
node              256 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "property is not a single cell");
node              268 scripts/dtc/checks.c 				       struct node *node)
node              270 scripts/dtc/checks.c 	struct node *child, *child2;
node              272 scripts/dtc/checks.c 	for_each_child(node, child)
node              282 scripts/dtc/checks.c 					   struct node *node)
node              286 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              291 scripts/dtc/checks.c 				FAIL_PROP(c, dti, node, prop, "Duplicate property name");
node              304 scripts/dtc/checks.c 				  struct node *node)
node              306 scripts/dtc/checks.c 	int n = strspn(node->name, c->data);
node              308 scripts/dtc/checks.c 	if (n < strlen(node->name))
node              309 scripts/dtc/checks.c 		FAIL(c, dti, node, "Bad character '%c' in node name",
node              310 scripts/dtc/checks.c 		     node->name[n]);
node              315 scripts/dtc/checks.c 					 struct node *node)
node              317 scripts/dtc/checks.c 	int n = strspn(node->name, c->data);
node              319 scripts/dtc/checks.c 	if (n < node->basenamelen)
node              320 scripts/dtc/checks.c 		FAIL(c, dti, node, "Character '%c' not recommended in node name",
node              321 scripts/dtc/checks.c 		     node->name[n]);
node              326 scripts/dtc/checks.c 				   struct node *node)
node              328 scripts/dtc/checks.c 	if (strchr(get_unitname(node), '@'))
node              329 scripts/dtc/checks.c 		FAIL(c, dti, node, "multiple '@' characters in node name");
node              334 scripts/dtc/checks.c 				      struct node *node)
node              336 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node              337 scripts/dtc/checks.c 	struct property *prop = get_property(node, "reg");
node              339 scripts/dtc/checks.c 	if (get_subnode(node, "__overlay__")) {
node              345 scripts/dtc/checks.c 		prop = get_property(node, "ranges");
node              352 scripts/dtc/checks.c 			FAIL(c, dti, node, "node has a reg or ranges property, but no unit name");
node              355 scripts/dtc/checks.c 			FAIL(c, dti, node, "node has a unit name, but no reg property");
node              361 scripts/dtc/checks.c 				      struct node *node)
node              365 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              369 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "Bad character '%c' in property name",
node              377 scripts/dtc/checks.c 					     struct node *node)
node              381 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              401 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "Character '%c' not recommended in property name",
node              408 scripts/dtc/checks.c #define DESCLABEL_ARGS(node,prop,mark)		\
node              412 scripts/dtc/checks.c 	((prop) ? "' in " : ""), (node)->fullpath
node              415 scripts/dtc/checks.c 				  const char *label, struct node *node,
node              418 scripts/dtc/checks.c 	struct node *dt = dti->dt;
node              419 scripts/dtc/checks.c 	struct node *othernode = NULL;
node              434 scripts/dtc/checks.c 	if ((othernode != node) || (otherprop != prop) || (othermark != mark))
node              435 scripts/dtc/checks.c 		FAIL(c, dti, node, "Duplicate label '%s' on " DESCLABEL_FMT
node              437 scripts/dtc/checks.c 		     label, DESCLABEL_ARGS(node, prop, mark),
node              442 scripts/dtc/checks.c 				       struct node *node)
node              447 scripts/dtc/checks.c 	for_each_label(node->labels, l)
node              448 scripts/dtc/checks.c 		check_duplicate_label(c, dti, l->label, node, NULL, NULL);
node              450 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              454 scripts/dtc/checks.c 			check_duplicate_label(c, dti, l->label, node, prop, NULL);
node              457 scripts/dtc/checks.c 			check_duplicate_label(c, dti, m->ref, node, prop, m);
node              463 scripts/dtc/checks.c 				 struct node *node, const char *propname)
node              465 scripts/dtc/checks.c 	struct node *root = dti->dt;
node              470 scripts/dtc/checks.c 	prop = get_property(node, propname);
node              475 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "bad length (%d) %s property",
node              483 scripts/dtc/checks.c 		if (node != get_node_by_ref(root, m->ref))
node              487 scripts/dtc/checks.c 			FAIL(c, dti, node, "%s is a reference to another node",
node              501 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "bad value (0x%x) in %s property",
node              510 scripts/dtc/checks.c 				    struct node *node)
node              512 scripts/dtc/checks.c 	struct node *root = dti->dt;
node              513 scripts/dtc/checks.c 	struct node *other;
node              517 scripts/dtc/checks.c 	assert(!node->phandle);
node              519 scripts/dtc/checks.c 	phandle = check_phandle_prop(c, dti, node, "phandle");
node              521 scripts/dtc/checks.c 	linux_phandle = check_phandle_prop(c, dti, node, "linux,phandle");
node              528 scripts/dtc/checks.c 		FAIL(c, dti, node, "mismatching 'phandle' and 'linux,phandle'"
node              535 scripts/dtc/checks.c 	if (other && (other != node)) {
node              536 scripts/dtc/checks.c 		FAIL(c, dti, node, "duplicated phandle 0x%x (seen before at %s)",
node              541 scripts/dtc/checks.c 	node->phandle = phandle;
node              546 scripts/dtc/checks.c 				  struct node *node)
node              550 scripts/dtc/checks.c 	for (pp = &node->proplist; *pp; pp = &((*pp)->next))
node              559 scripts/dtc/checks.c 	if ((prop->val.len != node->basenamelen+1)
node              560 scripts/dtc/checks.c 	    || (memcmp(prop->val.val, node->name, node->basenamelen) != 0)) {
node              561 scripts/dtc/checks.c 		FAIL(c, dti, node, "\"name\" property is incorrect (\"%s\" instead"
node              580 scripts/dtc/checks.c 				     struct node *node)
node              582 scripts/dtc/checks.c 	struct node *dt = dti->dt;
node              585 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              587 scripts/dtc/checks.c 		struct node *refnode;
node              596 scripts/dtc/checks.c 					FAIL(c, dti, node, "Reference to non-existent node or "
node              615 scripts/dtc/checks.c 				  struct node *node)
node              617 scripts/dtc/checks.c 	struct node *dt = dti->dt;
node              620 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              622 scripts/dtc/checks.c 		struct node *refnode;
node              630 scripts/dtc/checks.c 				FAIL(c, dti, node, "Reference to non-existent node or label \"%s\"\n",
node              646 scripts/dtc/checks.c 				    struct node *node)
node              648 scripts/dtc/checks.c 	if (generate_symbols && node->labels)
node              650 scripts/dtc/checks.c 	if (node->omit_if_unused && !node->is_referenced)
node              651 scripts/dtc/checks.c 		delete_node(node);
node              670 scripts/dtc/checks.c 				       struct node *node)
node              674 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              680 scripts/dtc/checks.c 		check_is_string_list(c, dti, node);
node              686 scripts/dtc/checks.c 				    struct node *node)
node              690 scripts/dtc/checks.c 	if (!streq(node->name, "aliases"))
node              693 scripts/dtc/checks.c 	for_each_property(node, prop) {
node              695 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "aliases property is not a valid node (%s)",
node              700 scripts/dtc/checks.c 			FAIL(c, dti, node, "aliases property name must include only lowercase and '-'");
node              706 scripts/dtc/checks.c 				  struct node *node)
node              710 scripts/dtc/checks.c 	node->addr_cells = -1;
node              711 scripts/dtc/checks.c 	node->size_cells = -1;
node              713 scripts/dtc/checks.c 	prop = get_property(node, "#address-cells");
node              715 scripts/dtc/checks.c 		node->addr_cells = propval_cell(prop);
node              717 scripts/dtc/checks.c 	prop = get_property(node, "#size-cells");
node              719 scripts/dtc/checks.c 		node->size_cells = propval_cell(prop);
node              730 scripts/dtc/checks.c 			     struct node *node)
node              735 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node              739 scripts/dtc/checks.c 	if (!node->parent) {
node              740 scripts/dtc/checks.c 		FAIL(c, dti, node, "Root node has a \"reg\" property");
node              745 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "property is empty");
node              747 scripts/dtc/checks.c 	addr_cells = node_addr_cells(node->parent);
node              748 scripts/dtc/checks.c 	size_cells = node_size_cells(node->parent);
node              752 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "property has invalid length (%d bytes) "
node              759 scripts/dtc/checks.c 				struct node *node)
node              764 scripts/dtc/checks.c 	prop = get_property(node, "ranges");
node              768 scripts/dtc/checks.c 	if (!node->parent) {
node              769 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "Root node has a \"ranges\" property");
node              773 scripts/dtc/checks.c 	p_addr_cells = node_addr_cells(node->parent);
node              774 scripts/dtc/checks.c 	p_size_cells = node_size_cells(node->parent);
node              775 scripts/dtc/checks.c 	c_addr_cells = node_addr_cells(node);
node              776 scripts/dtc/checks.c 	c_size_cells = node_size_cells(node);
node              781 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "empty \"ranges\" property but its "
node              783 scripts/dtc/checks.c 				  c_addr_cells, node->parent->fullpath,
node              786 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "empty \"ranges\" property but its "
node              788 scripts/dtc/checks.c 				  c_size_cells, node->parent->fullpath,
node              791 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "\"ranges\" property has invalid length (%d bytes) "
node              803 scripts/dtc/checks.c static void check_pci_bridge(struct check *c, struct dt_info *dti, struct node *node)
node              808 scripts/dtc/checks.c 	prop = get_property(node, "device_type");
node              812 scripts/dtc/checks.c 	node->bus = &pci_bus;
node              814 scripts/dtc/checks.c 	if (!strprefixeq(node->name, node->basenamelen, "pci") &&
node              815 scripts/dtc/checks.c 	    !strprefixeq(node->name, node->basenamelen, "pcie"))
node              816 scripts/dtc/checks.c 		FAIL(c, dti, node, "node name is not \"pci\" or \"pcie\"");
node              818 scripts/dtc/checks.c 	prop = get_property(node, "ranges");
node              820 scripts/dtc/checks.c 		FAIL(c, dti, node, "missing ranges for PCI bridge (or not a bridge)");
node              822 scripts/dtc/checks.c 	if (node_addr_cells(node) != 3)
node              823 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #address-cells for PCI bridge");
node              824 scripts/dtc/checks.c 	if (node_size_cells(node) != 2)
node              825 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #size-cells for PCI bridge");
node              827 scripts/dtc/checks.c 	prop = get_property(node, "bus-range");
node              832 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "value must be 2 cells");
node              837 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "1st cell must be less than or equal to 2nd cell");
node              839 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "maximum bus number must be less than 256");
node              844 scripts/dtc/checks.c static void check_pci_device_bus_num(struct check *c, struct dt_info *dti, struct node *node)
node              850 scripts/dtc/checks.c 	if (!node->parent || (node->parent->bus != &pci_bus))
node              853 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node              860 scripts/dtc/checks.c 	prop = get_property(node->parent, "bus-range");
node              869 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "PCI bus number %d out of range, expected (%d - %d)",
node              874 scripts/dtc/checks.c static void check_pci_device_reg(struct check *c, struct dt_info *dti, struct node *node)
node              877 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node              882 scripts/dtc/checks.c 	if (!node->parent || (node->parent->bus != &pci_bus))
node              885 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node              887 scripts/dtc/checks.c 		FAIL(c, dti, node, "missing PCI reg property");
node              893 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "PCI reg config space address cells 2 and 3 must be 0");
node              900 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "PCI reg address is not configuration space");
node              902 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "PCI reg config space address register number must be 0");
node              914 scripts/dtc/checks.c 	FAIL(c, dti, node, "PCI unit address format error, expected \"%s\"",
node              923 scripts/dtc/checks.c static bool node_is_compatible(struct node *node, const char *compat)
node              928 scripts/dtc/checks.c 	prop = get_property(node, "compatible");
node              940 scripts/dtc/checks.c static void check_simple_bus_bridge(struct check *c, struct dt_info *dti, struct node *node)
node              942 scripts/dtc/checks.c 	if (node_is_compatible(node, "simple-bus"))
node              943 scripts/dtc/checks.c 		node->bus = &simple_bus;
node              948 scripts/dtc/checks.c static void check_simple_bus_reg(struct check *c, struct dt_info *dti, struct node *node)
node              951 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node              957 scripts/dtc/checks.c 	if (!node->parent || (node->parent->bus != &simple_bus))
node              960 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node              964 scripts/dtc/checks.c 		prop = get_property(node, "ranges");
node              967 scripts/dtc/checks.c 			cells = ((cell_t *)prop->val.val) + node_addr_cells(node);
node              971 scripts/dtc/checks.c 		if (node->parent->parent && !(node->bus == &simple_bus))
node              972 scripts/dtc/checks.c 			FAIL(c, dti, node, "missing or empty reg/ranges property");
node              976 scripts/dtc/checks.c 	size = node_addr_cells(node->parent);
node              982 scripts/dtc/checks.c 		FAIL(c, dti, node, "simple-bus unit address format error, expected \"%s\"",
node              991 scripts/dtc/checks.c static void check_i2c_bus_bridge(struct check *c, struct dt_info *dti, struct node *node)
node              993 scripts/dtc/checks.c 	if (strprefixeq(node->name, node->basenamelen, "i2c-bus") ||
node              994 scripts/dtc/checks.c 	    strprefixeq(node->name, node->basenamelen, "i2c-arb")) {
node              995 scripts/dtc/checks.c 		node->bus = &i2c_bus;
node              996 scripts/dtc/checks.c 	} else if (strprefixeq(node->name, node->basenamelen, "i2c")) {
node              997 scripts/dtc/checks.c 		struct node *child;
node              998 scripts/dtc/checks.c 		for_each_child(node, child) {
node              999 scripts/dtc/checks.c 			if (strprefixeq(child->name, node->basenamelen, "i2c-bus"))
node             1002 scripts/dtc/checks.c 		node->bus = &i2c_bus;
node             1006 scripts/dtc/checks.c 	if (!node->children)
node             1009 scripts/dtc/checks.c 	if (node_addr_cells(node) != 1)
node             1010 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #address-cells for I2C bus");
node             1011 scripts/dtc/checks.c 	if (node_size_cells(node) != 0)
node             1012 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #size-cells for I2C bus");
node             1017 scripts/dtc/checks.c static void check_i2c_bus_reg(struct check *c, struct dt_info *dti, struct node *node)
node             1020 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node             1026 scripts/dtc/checks.c 	if (!node->parent || (node->parent->bus != &i2c_bus))
node             1029 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node             1034 scripts/dtc/checks.c 		FAIL(c, dti, node, "missing or empty reg property");
node             1041 scripts/dtc/checks.c 		FAIL(c, dti, node, "I2C bus unit address format error, expected \"%s\"",
node             1047 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop, "I2C address must be less than 10-bits, got \"0x%x\"",
node             1058 scripts/dtc/checks.c static void check_spi_bus_bridge(struct check *c, struct dt_info *dti, struct node *node)
node             1062 scripts/dtc/checks.c 	if (strprefixeq(node->name, node->basenamelen, "spi")) {
node             1063 scripts/dtc/checks.c 		node->bus = &spi_bus;
node             1066 scripts/dtc/checks.c 		struct node *child;
node             1068 scripts/dtc/checks.c 		if (node_addr_cells(node) != 1 || node_size_cells(node) != 0)
node             1071 scripts/dtc/checks.c 		for_each_child(node, child) {
node             1075 scripts/dtc/checks.c 					node->bus = &spi_bus;
node             1079 scripts/dtc/checks.c 			if (node->bus == &spi_bus)
node             1083 scripts/dtc/checks.c 		if (node->bus == &spi_bus && get_property(node, "reg"))
node             1084 scripts/dtc/checks.c 			FAIL(c, dti, node, "node name for SPI buses should be 'spi'");
node             1086 scripts/dtc/checks.c 	if (node->bus != &spi_bus || !node->children)
node             1089 scripts/dtc/checks.c 	if (get_property(node, "spi-slave"))
node             1091 scripts/dtc/checks.c 	if (node_addr_cells(node) != spi_addr_cells)
node             1092 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #address-cells for SPI bus");
node             1093 scripts/dtc/checks.c 	if (node_size_cells(node) != 0)
node             1094 scripts/dtc/checks.c 		FAIL(c, dti, node, "incorrect #size-cells for SPI bus");
node             1099 scripts/dtc/checks.c static void check_spi_bus_reg(struct check *c, struct dt_info *dti, struct node *node)
node             1102 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node             1107 scripts/dtc/checks.c 	if (!node->parent || (node->parent->bus != &spi_bus))
node             1110 scripts/dtc/checks.c 	if (get_property(node->parent, "spi-slave"))
node             1113 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node             1118 scripts/dtc/checks.c 		FAIL(c, dti, node, "missing or empty reg property");
node             1125 scripts/dtc/checks.c 		FAIL(c, dti, node, "SPI bus unit address format error, expected \"%s\"",
node             1131 scripts/dtc/checks.c 				      struct node *node)
node             1133 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node             1135 scripts/dtc/checks.c 	if (node->parent && node->parent->bus)
node             1142 scripts/dtc/checks.c 		FAIL(c, dti, node, "unit name should not have leading \"0x\"");
node             1147 scripts/dtc/checks.c 		FAIL(c, dti, node, "unit name should not have leading 0s");
node             1156 scripts/dtc/checks.c 					  struct node *node)
node             1160 scripts/dtc/checks.c 	if (!node->parent)
node             1163 scripts/dtc/checks.c 	reg = get_property(node, "reg");
node             1164 scripts/dtc/checks.c 	ranges = get_property(node, "ranges");
node             1169 scripts/dtc/checks.c 	if (node->parent->addr_cells == -1)
node             1170 scripts/dtc/checks.c 		FAIL(c, dti, node, "Relying on default #address-cells value");
node             1172 scripts/dtc/checks.c 	if (node->parent->size_cells == -1)
node             1173 scripts/dtc/checks.c 		FAIL(c, dti, node, "Relying on default #size-cells value");
node             1179 scripts/dtc/checks.c 					      struct node *node)
node             1182 scripts/dtc/checks.c 	struct node *child;
node             1185 scripts/dtc/checks.c 	if (!node->parent || node->addr_cells < 0 || node->size_cells < 0)
node             1188 scripts/dtc/checks.c 	if (get_property(node, "ranges") || !node->children)
node             1191 scripts/dtc/checks.c 	for_each_child(node, child) {
node             1198 scripts/dtc/checks.c 		FAIL(c, dti, node, "unnecessary #address-cells/#size-cells without \"ranges\" or child \"reg\" property");
node             1202 scripts/dtc/checks.c static bool node_is_disabled(struct node *node)
node             1206 scripts/dtc/checks.c 	prop = get_property(node, "status");
node             1218 scripts/dtc/checks.c 						struct node *node,
node             1221 scripts/dtc/checks.c 	struct node *childa;
node             1223 scripts/dtc/checks.c 	if (node->addr_cells < 0 || node->size_cells < 0)
node             1226 scripts/dtc/checks.c 	if (!node->children)
node             1229 scripts/dtc/checks.c 	for_each_child(node, childa) {
node             1230 scripts/dtc/checks.c 		struct node *childb;
node             1239 scripts/dtc/checks.c 		for_each_child(node, childb) {
node             1254 scripts/dtc/checks.c 					      struct node *node)
node             1256 scripts/dtc/checks.c 	check_unique_unit_address_common(c, dti, node, false);
node             1261 scripts/dtc/checks.c 					      struct node *node)
node             1263 scripts/dtc/checks.c 	check_unique_unit_address_common(c, dti, node, true);
node             1270 scripts/dtc/checks.c 						       struct node *node)
node             1272 scripts/dtc/checks.c 	struct node *dt = dti->dt;
node             1273 scripts/dtc/checks.c 	struct node *chosen;
node             1276 scripts/dtc/checks.c 	if (node != dt)
node             1286 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop,
node             1293 scripts/dtc/checks.c 				      struct node *node)
node             1295 scripts/dtc/checks.c 	if (!streq(node->name, "chosen"))
node             1298 scripts/dtc/checks.c 	if (node->parent != dti->dt)
node             1299 scripts/dtc/checks.c 		FAIL(c, dti, node, "chosen node must be at root node");
node             1304 scripts/dtc/checks.c 				       struct node *node)
node             1308 scripts/dtc/checks.c 	if (!streq(node->name, "chosen"))
node             1311 scripts/dtc/checks.c 	prop = get_property(node, "bootargs");
node             1316 scripts/dtc/checks.c 	check_is_string(c, dti, node);
node             1321 scripts/dtc/checks.c 					  struct node *node)
node             1325 scripts/dtc/checks.c 	if (!streq(node->name, "chosen"))
node             1328 scripts/dtc/checks.c 	prop = get_property(node, "stdout-path");
node             1330 scripts/dtc/checks.c 		prop = get_property(node, "linux,stdout-path");
node             1333 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop, "Use 'stdout-path' instead");
node             1337 scripts/dtc/checks.c 	check_is_string(c, dti, node);
node             1349 scripts/dtc/checks.c 				          struct node *node,
node             1353 scripts/dtc/checks.c 	struct node *root = dti->dt;
node             1357 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop,
node             1364 scripts/dtc/checks.c 		struct node *provider_node;
node             1390 scripts/dtc/checks.c 				FAIL_PROP(c, dti, node, prop,
node             1397 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop,
node             1409 scripts/dtc/checks.c 			FAIL(c, dti, node, "Missing property '%s' in node %s or bad phandle (referred from %s[%d])",
node             1417 scripts/dtc/checks.c 			FAIL_PROP(c, dti, node, prop,
node             1426 scripts/dtc/checks.c 				          struct node *node)
node             1431 scripts/dtc/checks.c 	prop = get_property(node, provider->prop_name);
node             1435 scripts/dtc/checks.c 	check_property_phandle_args(c, dti, node, prop, provider);
node             1482 scripts/dtc/checks.c 				          struct node *node)
node             1487 scripts/dtc/checks.c 	if (get_property(node, "gpio-hog"))
node             1490 scripts/dtc/checks.c 	for_each_property(node, prop) {
node             1499 scripts/dtc/checks.c 		check_property_phandle_args(c, dti, node, prop, &provider);
node             1507 scripts/dtc/checks.c 				           struct node *node)
node             1511 scripts/dtc/checks.c 	for_each_property(node, prop) {
node             1521 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop,
node             1528 scripts/dtc/checks.c static bool node_is_interrupt_provider(struct node *node)
node             1532 scripts/dtc/checks.c 	prop = get_property(node, "interrupt-controller");
node             1536 scripts/dtc/checks.c 	prop = get_property(node, "interrupt-map");
node             1544 scripts/dtc/checks.c 				      struct node *node)
node             1546 scripts/dtc/checks.c 	struct node *root = dti->dt;
node             1547 scripts/dtc/checks.c 	struct node *irq_node = NULL, *parent = node;
node             1551 scripts/dtc/checks.c 	irq_prop = get_property(node, "interrupts");
node             1556 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, irq_prop, "size (%d) is invalid, expected multiple of %zu",
node             1560 scripts/dtc/checks.c 		if (parent != node && node_is_interrupt_provider(parent)) {
node             1593 scripts/dtc/checks.c 		FAIL(c, dti, node, "Missing interrupt-parent");
node             1605 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, prop,
node             1621 scripts/dtc/checks.c 			      struct node *node)
node             1623 scripts/dtc/checks.c 	struct node *child;
node             1625 scripts/dtc/checks.c 	for_each_child(node, child) {
node             1630 scripts/dtc/checks.c 		node->bus = &graph_port_bus;
node             1633 scripts/dtc/checks.c 		if (!node->parent->bus &&
node             1634 scripts/dtc/checks.c 		    (streq(node->parent->name, "ports") || get_property(node, "reg")))
node             1635 scripts/dtc/checks.c 			node->parent->bus = &graph_ports_bus;
node             1644 scripts/dtc/checks.c 				      struct node *node)
node             1647 scripts/dtc/checks.c 	struct node *child;
node             1649 scripts/dtc/checks.c 	if (node->bus != &graph_ports_bus && node->bus != &graph_port_bus)
node             1652 scripts/dtc/checks.c 	for_each_child(node, child) {
node             1662 scripts/dtc/checks.c 	if (cnt == 1 && node->addr_cells != -1)
node             1663 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph node has single child node '%s', #address-cells/#size-cells are not necessary",
node             1664 scripts/dtc/checks.c 		     node->children->name);
node             1669 scripts/dtc/checks.c 			    struct node *node)
node             1672 scripts/dtc/checks.c 	const char *unitname = get_unitname(node);
node             1675 scripts/dtc/checks.c 	prop = get_property(node, "reg");
node             1680 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph node malformed 'reg' property");
node             1686 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph node unit address error, expected \"%s\"",
node             1689 scripts/dtc/checks.c 	if (node->parent->addr_cells != 1)
node             1690 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, get_property(node, "#address-cells"),
node             1692 scripts/dtc/checks.c 			  node->parent->addr_cells);
node             1693 scripts/dtc/checks.c 	if (node->parent->size_cells != 0)
node             1694 scripts/dtc/checks.c 		FAIL_PROP(c, dti, node, get_property(node, "#size-cells"),
node             1696 scripts/dtc/checks.c 			  node->parent->size_cells);
node             1700 scripts/dtc/checks.c 			     struct node *node)
node             1702 scripts/dtc/checks.c 	if (node->bus != &graph_port_bus)
node             1705 scripts/dtc/checks.c 	if (!strprefixeq(node->name, node->basenamelen, "port"))
node             1706 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph port node name should be 'port'");
node             1708 scripts/dtc/checks.c 	check_graph_reg(c, dti, node);
node             1712 scripts/dtc/checks.c static struct node *get_remote_endpoint(struct check *c, struct dt_info *dti,
node             1713 scripts/dtc/checks.c 					struct node *endpoint)
node             1716 scripts/dtc/checks.c 	struct node *node;
node             1728 scripts/dtc/checks.c 	node = get_node_by_phandle(dti->dt, phandle);
node             1729 scripts/dtc/checks.c 	if (!node)
node             1732 scripts/dtc/checks.c 	return node;
node             1736 scripts/dtc/checks.c 				 struct node *node)
node             1738 scripts/dtc/checks.c 	struct node *remote_node;
node             1740 scripts/dtc/checks.c 	if (!node->parent || node->parent->bus != &graph_port_bus)
node             1743 scripts/dtc/checks.c 	if (!strprefixeq(node->name, node->basenamelen, "endpoint"))
node             1744 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph endpoint node name should be 'endpoint'");
node             1746 scripts/dtc/checks.c 	check_graph_reg(c, dti, node);
node             1748 scripts/dtc/checks.c 	remote_node = get_remote_endpoint(c, dti, node);
node             1752 scripts/dtc/checks.c 	if (get_remote_endpoint(c, dti, remote_node) != node)
node             1753 scripts/dtc/checks.c 		FAIL(c, dti, node, "graph connection to node '%s' is not bidirectional",
node               37 scripts/dtc/dtc-parser.y 	struct node *node;
node               38 scripts/dtc/dtc-parser.y 	struct node *nodelist;
node               74 scripts/dtc/dtc-parser.y %type <node> devicetree
node               75 scripts/dtc/dtc-parser.y %type <node> nodedef
node               76 scripts/dtc/dtc-parser.y %type <node> subnode
node              175 scripts/dtc/dtc-parser.y 			struct node *target = get_node_by_ref($1, $3);
node              194 scripts/dtc/dtc-parser.y 				struct node *target = get_node_by_ref($1, $2);
node              205 scripts/dtc/dtc-parser.y 			struct node *target = get_node_by_ref($1, $2);
node              224 scripts/dtc/dtc-parser.y 			struct node *target = get_node_by_ref($1, $3);
node              236 scripts/dtc/dtc-parser.y 			struct node *target = get_node_by_ref($1, $3);
node               31 scripts/dtc/dtc.c static void fill_fullpaths(struct node *tree, const char *prefix)
node               33 scripts/dtc/dtc.c 	struct node *child;
node              154 scripts/dtc/dtc.h 	struct node *children;
node              156 scripts/dtc/dtc.h 	struct node *parent;
node              157 scripts/dtc/dtc.h 	struct node *next_sibling;
node              202 scripts/dtc/dtc.h struct node *build_node(struct property *proplist, struct node *children,
node              204 scripts/dtc/dtc.h struct node *build_node_delete(struct srcpos *srcpos);
node              205 scripts/dtc/dtc.h struct node *name_node(struct node *node, char *name);
node              206 scripts/dtc/dtc.h struct node *omit_node_if_unused(struct node *node);
node              207 scripts/dtc/dtc.h struct node *reference_node(struct node *node);
node              208 scripts/dtc/dtc.h struct node *chain_node(struct node *first, struct node *list);
node              209 scripts/dtc/dtc.h struct node *merge_nodes(struct node *old_node, struct node *new_node);
node              210 scripts/dtc/dtc.h struct node *add_orphan_node(struct node *old_node, struct node *new_node, char *ref);
node              212 scripts/dtc/dtc.h void add_property(struct node *node, struct property *prop);
node              213 scripts/dtc/dtc.h void delete_property_by_name(struct node *node, char *name);
node              215 scripts/dtc/dtc.h void add_child(struct node *parent, struct node *child);
node              216 scripts/dtc/dtc.h void delete_node_by_name(struct node *parent, char *name);
node              217 scripts/dtc/dtc.h void delete_node(struct node *node);
node              218 scripts/dtc/dtc.h void append_to_property(struct node *node,
node              222 scripts/dtc/dtc.h const char *get_unitname(struct node *node);
node              223 scripts/dtc/dtc.h struct property *get_property(struct node *node, const char *propname);
node              226 scripts/dtc/dtc.h struct property *get_property_by_label(struct node *tree, const char *label,
node              227 scripts/dtc/dtc.h 				       struct node **node);
node              228 scripts/dtc/dtc.h struct marker *get_marker_label(struct node *tree, const char *label,
node              229 scripts/dtc/dtc.h 				struct node **node, struct property **prop);
node              230 scripts/dtc/dtc.h struct node *get_subnode(struct node *node, const char *nodename);
node              231 scripts/dtc/dtc.h struct node *get_node_by_path(struct node *tree, const char *path);
node              232 scripts/dtc/dtc.h struct node *get_node_by_label(struct node *tree, const char *label);
node              233 scripts/dtc/dtc.h struct node *get_node_by_phandle(struct node *tree, cell_t phandle);
node              234 scripts/dtc/dtc.h struct node *get_node_by_ref(struct node *tree, const char *ref);
node              235 scripts/dtc/dtc.h cell_t get_node_phandle(struct node *root, struct node *node);
node              237 scripts/dtc/dtc.h uint32_t guess_boot_cpuid(struct node *tree);
node              260 scripts/dtc/dtc.h 	struct node *dt;		/* the device tree */
node              270 scripts/dtc/dtc.h 			      struct node *tree, uint32_t boot_cpuid_phys);
node              108 scripts/dtc/fdtget.c static int list_properties(const void *blob, int node)
node              114 scripts/dtc/fdtget.c 	prop = fdt_first_property_offset(blob, node);
node              136 scripts/dtc/fdtget.c static int list_subnodes(const void *blob, int node)
node              145 scripts/dtc/fdtget.c 		tag = fdt_next_tag(blob, node, &nextoffset);
node              148 scripts/dtc/fdtget.c 			pathp = fdt_get_name(blob, node, NULL);
node              177 scripts/dtc/fdtget.c 		node = nextoffset;
node              193 scripts/dtc/fdtget.c 		int node, const char *property)
node              200 scripts/dtc/fdtget.c 		err = list_properties(blob, node);
node              204 scripts/dtc/fdtget.c 		err = list_subnodes(blob, node);
node              209 scripts/dtc/fdtget.c 		value = fdt_getprop(blob, node, property, &len);
node              241 scripts/dtc/fdtget.c 	int i, node;
node              248 scripts/dtc/fdtget.c 		node = fdt_path_offset(blob, arg[i]);
node              249 scripts/dtc/fdtget.c 		if (node < 0) {
node              254 scripts/dtc/fdtget.c 				report_error(arg[i], node);
node              260 scripts/dtc/fdtget.c 		if (show_data_for_item(blob, disp, node, prop))
node              123 scripts/dtc/fdtput.c 	int node;
node              126 scripts/dtc/fdtput.c 	node = fdt_path_offset(blob, node_name);
node              127 scripts/dtc/fdtput.c 	if (node < 0) {
node              128 scripts/dtc/fdtput.c 		report_error(node_name, -1, node);
node              132 scripts/dtc/fdtput.c 	err = fdt_setprop(blob, node, property, buf, len);
node              154 scripts/dtc/fdtput.c 	int node, offset = 0;
node              160 scripts/dtc/fdtput.c 	for (sep = path; *sep; path = sep + 1, offset = node) {
node              166 scripts/dtc/fdtput.c 		node = fdt_subnode_offset_namelen(blob, offset, path,
node              168 scripts/dtc/fdtput.c 		if (node == -FDT_ERR_NOTFOUND) {
node              169 scripts/dtc/fdtput.c 			node = fdt_add_subnode_namelen(blob, offset, path,
node              172 scripts/dtc/fdtput.c 		if (node < 0) {
node              173 scripts/dtc/fdtput.c 			report_error(path, sep - path, node);
node              194 scripts/dtc/fdtput.c 	int node = 0;
node              205 scripts/dtc/fdtput.c 		node = fdt_path_offset(blob, node_name);
node              206 scripts/dtc/fdtput.c 		if (node < 0) {
node              207 scripts/dtc/fdtput.c 			report_error(node_name, -1, node);
node              212 scripts/dtc/fdtput.c 	node = fdt_add_subnode(blob, node, p + 1);
node              213 scripts/dtc/fdtput.c 	if (node < 0) {
node              214 scripts/dtc/fdtput.c 		report_error(p + 1, -1, node);
node              235 scripts/dtc/flattree.c static void flatten_tree(struct node *tree, struct emitter *emit,
node              240 scripts/dtc/flattree.c 	struct node *child;
node              730 scripts/dtc/flattree.c static struct node *unflatten_tree(struct inbuf *dtbuf,
node              734 scripts/dtc/flattree.c 	struct node *node;
node              738 scripts/dtc/flattree.c 	node = build_node(NULL, NULL, NULL);
node              743 scripts/dtc/flattree.c 		node->name = nodename_from_path(parent_flatname, flatname);
node              745 scripts/dtc/flattree.c 		node->name = flatname;
node              749 scripts/dtc/flattree.c 		struct node *child;
node              754 scripts/dtc/flattree.c 			if (node->children)
node              758 scripts/dtc/flattree.c 			add_property(node, prop);
node              763 scripts/dtc/flattree.c 			add_child(node, child);
node              787 scripts/dtc/flattree.c 	if (node->name != flatname) {
node              791 scripts/dtc/flattree.c 	return node;
node              809 scripts/dtc/flattree.c 	struct node *tree;
node               11 scripts/dtc/fstree.c static struct node *read_fstree(const char *dirname)
node               16 scripts/dtc/fstree.c 	struct node *tree;
node               54 scripts/dtc/fstree.c 			struct node *newchild;
node               70 scripts/dtc/fstree.c 	struct node *tree;
node              116 scripts/dtc/libfdt/fdt_overlay.c static int overlay_phandle_add_offset(void *fdt, int node,
node              123 scripts/dtc/libfdt/fdt_overlay.c 	val = fdt_getprop(fdt, node, name, &len);
node              138 scripts/dtc/libfdt/fdt_overlay.c 	return fdt_setprop_inplace_u32(fdt, node, name, adj_val);
node              156 scripts/dtc/libfdt/fdt_overlay.c static int overlay_adjust_node_phandles(void *fdto, int node,
node              162 scripts/dtc/libfdt/fdt_overlay.c 	ret = overlay_phandle_add_offset(fdto, node, "phandle", delta);
node              166 scripts/dtc/libfdt/fdt_overlay.c 	ret = overlay_phandle_add_offset(fdto, node, "linux,phandle", delta);
node              170 scripts/dtc/libfdt/fdt_overlay.c 	fdt_for_each_subnode(child, fdto, node) {
node              557 scripts/dtc/libfdt/fdt_overlay.c 			      void *fdto, int node)
node              562 scripts/dtc/libfdt/fdt_overlay.c 	fdt_for_each_property_offset(property, fdto, node) {
node              580 scripts/dtc/libfdt/fdt_overlay.c 	fdt_for_each_subnode(subnode, fdto, node) {
node              226 scripts/dtc/libfdt/libfdt.h #define fdt_for_each_subnode(node, fdt, parent)		\
node              227 scripts/dtc/libfdt/libfdt.h 	for (node = fdt_first_subnode(fdt, parent);	\
node              228 scripts/dtc/libfdt/libfdt.h 	     node >= 0;					\
node              229 scripts/dtc/libfdt/libfdt.h 	     node = fdt_next_subnode(fdt, node))
node              605 scripts/dtc/libfdt/libfdt.h #define fdt_for_each_property_offset(property, fdt, node)	\
node              606 scripts/dtc/libfdt/libfdt.h 	for (property = fdt_first_property_offset(fdt, node);	\
node               88 scripts/dtc/livetree.c struct node *build_node(struct property *proplist, struct node *children,
node               91 scripts/dtc/livetree.c 	struct node *new = xmalloc(sizeof(*new));
node               92 scripts/dtc/livetree.c 	struct node *child;
node              107 scripts/dtc/livetree.c struct node *build_node_delete(struct srcpos *srcpos)
node              109 scripts/dtc/livetree.c 	struct node *new = xmalloc(sizeof(*new));
node              119 scripts/dtc/livetree.c struct node *name_node(struct node *node, char *name)
node              121 scripts/dtc/livetree.c 	assert(node->name == NULL);
node              123 scripts/dtc/livetree.c 	node->name = name;
node              125 scripts/dtc/livetree.c 	return node;
node              128 scripts/dtc/livetree.c struct node *omit_node_if_unused(struct node *node)
node              130 scripts/dtc/livetree.c 	node->omit_if_unused = 1;
node              132 scripts/dtc/livetree.c 	return node;
node              135 scripts/dtc/livetree.c struct node *reference_node(struct node *node)
node              137 scripts/dtc/livetree.c 	node->is_referenced = 1;
node              139 scripts/dtc/livetree.c 	return node;
node              142 scripts/dtc/livetree.c struct node *merge_nodes(struct node *old_node, struct node *new_node)
node              145 scripts/dtc/livetree.c 	struct node *new_child, *old_child;
node              228 scripts/dtc/livetree.c struct node * add_orphan_node(struct node *dt, struct node *new_node, char *ref)
node              231 scripts/dtc/livetree.c 	struct node *node;
node              251 scripts/dtc/livetree.c 	node = build_node(p, new_node, NULL);
node              252 scripts/dtc/livetree.c 	name_node(node, name);
node              254 scripts/dtc/livetree.c 	add_child(dt, node);
node              258 scripts/dtc/livetree.c struct node *chain_node(struct node *first, struct node *list)
node              266 scripts/dtc/livetree.c void add_property(struct node *node, struct property *prop)
node              272 scripts/dtc/livetree.c 	p = &node->proplist;
node              279 scripts/dtc/livetree.c void delete_property_by_name(struct node *node, char *name)
node              281 scripts/dtc/livetree.c 	struct property *prop = node->proplist;
node              298 scripts/dtc/livetree.c void add_child(struct node *parent, struct node *child)
node              300 scripts/dtc/livetree.c 	struct node **p;
node              312 scripts/dtc/livetree.c void delete_node_by_name(struct node *parent, char *name)
node              314 scripts/dtc/livetree.c 	struct node *node = parent->children;
node              316 scripts/dtc/livetree.c 	while (node) {
node              317 scripts/dtc/livetree.c 		if (streq(node->name, name)) {
node              318 scripts/dtc/livetree.c 			delete_node(node);
node              321 scripts/dtc/livetree.c 		node = node->next_sibling;
node              325 scripts/dtc/livetree.c void delete_node(struct node *node)
node              328 scripts/dtc/livetree.c 	struct node *child;
node              330 scripts/dtc/livetree.c 	node->deleted = 1;
node              331 scripts/dtc/livetree.c 	for_each_child(node, child)
node              333 scripts/dtc/livetree.c 	for_each_property(node, prop)
node              335 scripts/dtc/livetree.c 	delete_labels(&node->labels);
node              338 scripts/dtc/livetree.c void append_to_property(struct node *node,
node              345 scripts/dtc/livetree.c 	p = get_property(node, name);
node              354 scripts/dtc/livetree.c 		add_property(node, p);
node              399 scripts/dtc/livetree.c 			      struct node *tree, uint32_t boot_cpuid_phys)
node              416 scripts/dtc/livetree.c const char *get_unitname(struct node *node)
node              418 scripts/dtc/livetree.c 	if (node->name[node->basenamelen] == '\0')
node              421 scripts/dtc/livetree.c 		return node->name + node->basenamelen + 1;
node              424 scripts/dtc/livetree.c struct property *get_property(struct node *node, const char *propname)
node              428 scripts/dtc/livetree.c 	for_each_property(node, prop)
node              447 scripts/dtc/livetree.c struct property *get_property_by_label(struct node *tree, const char *label,
node              448 scripts/dtc/livetree.c 				       struct node **node)
node              451 scripts/dtc/livetree.c 	struct node *c;
node              453 scripts/dtc/livetree.c 	*node = tree;
node              464 scripts/dtc/livetree.c 		prop = get_property_by_label(c, label, node);
node              469 scripts/dtc/livetree.c 	*node = NULL;
node              473 scripts/dtc/livetree.c struct marker *get_marker_label(struct node *tree, const char *label,
node              474 scripts/dtc/livetree.c 				struct node **node, struct property **prop)
node              478 scripts/dtc/livetree.c 	struct node *c;
node              480 scripts/dtc/livetree.c 	*node = tree;
node              491 scripts/dtc/livetree.c 		m = get_marker_label(c, label, node, prop);
node              497 scripts/dtc/livetree.c 	*node = NULL;
node              501 scripts/dtc/livetree.c struct node *get_subnode(struct node *node, const char *nodename)
node              503 scripts/dtc/livetree.c 	struct node *child;
node              505 scripts/dtc/livetree.c 	for_each_child(node, child)
node              512 scripts/dtc/livetree.c struct node *get_node_by_path(struct node *tree, const char *path)
node              515 scripts/dtc/livetree.c 	struct node *child;
node              539 scripts/dtc/livetree.c struct node *get_node_by_label(struct node *tree, const char *label)
node              541 scripts/dtc/livetree.c 	struct node *child, *node;
node              551 scripts/dtc/livetree.c 		node = get_node_by_label(child, label);
node              552 scripts/dtc/livetree.c 		if (node)
node              553 scripts/dtc/livetree.c 			return node;
node              559 scripts/dtc/livetree.c struct node *get_node_by_phandle(struct node *tree, cell_t phandle)
node              561 scripts/dtc/livetree.c 	struct node *child, *node;
node              575 scripts/dtc/livetree.c 		node = get_node_by_phandle(child, phandle);
node              576 scripts/dtc/livetree.c 		if (node)
node              577 scripts/dtc/livetree.c 			return node;
node              583 scripts/dtc/livetree.c struct node *get_node_by_ref(struct node *tree, const char *ref)
node              593 scripts/dtc/livetree.c cell_t get_node_phandle(struct node *root, struct node *node)
node              598 scripts/dtc/livetree.c 	if ((node->phandle != 0) && (node->phandle != -1))
node              599 scripts/dtc/livetree.c 		return node->phandle;
node              604 scripts/dtc/livetree.c 	node->phandle = phandle;
node              609 scripts/dtc/livetree.c 	if (!get_property(node, "linux,phandle")
node              611 scripts/dtc/livetree.c 		add_property(node, build_property("linux,phandle", d, NULL));
node              613 scripts/dtc/livetree.c 	if (!get_property(node, "phandle")
node              615 scripts/dtc/livetree.c 		add_property(node, build_property("phandle", d, NULL));
node              621 scripts/dtc/livetree.c 	return node->phandle;
node              624 scripts/dtc/livetree.c uint32_t guess_boot_cpuid(struct node *tree)
node              626 scripts/dtc/livetree.c 	struct node *cpus, *bootcpu;
node              706 scripts/dtc/livetree.c static void sort_properties(struct node *node)
node              711 scripts/dtc/livetree.c 	for_each_property_withdel(node, prop)
node              719 scripts/dtc/livetree.c 	for_each_property_withdel(node, prop)
node              724 scripts/dtc/livetree.c 	node->proplist = tbl[0];
node              734 scripts/dtc/livetree.c 	const struct node *a, *b;
node              736 scripts/dtc/livetree.c 	a = *((const struct node * const *)ax);
node              737 scripts/dtc/livetree.c 	b = *((const struct node * const *)bx);
node              742 scripts/dtc/livetree.c static void sort_subnodes(struct node *node)
node              745 scripts/dtc/livetree.c 	struct node *subnode, **tbl;
node              747 scripts/dtc/livetree.c 	for_each_child_withdel(node, subnode)
node              755 scripts/dtc/livetree.c 	for_each_child_withdel(node, subnode)
node              760 scripts/dtc/livetree.c 	node->children = tbl[0];
node              768 scripts/dtc/livetree.c static void sort_node(struct node *node)
node              770 scripts/dtc/livetree.c 	struct node *c;
node              772 scripts/dtc/livetree.c 	sort_properties(node);
node              773 scripts/dtc/livetree.c 	sort_subnodes(node);
node              774 scripts/dtc/livetree.c 	for_each_child_withdel(node, c)
node              785 scripts/dtc/livetree.c static struct node *build_and_name_child_node(struct node *parent, char *name)
node              787 scripts/dtc/livetree.c 	struct node *node;
node              789 scripts/dtc/livetree.c 	node = build_node(NULL, NULL, NULL);
node              790 scripts/dtc/livetree.c 	name_node(node, xstrdup(name));
node              791 scripts/dtc/livetree.c 	add_child(parent, node);
node              793 scripts/dtc/livetree.c 	return node;
node              796 scripts/dtc/livetree.c static struct node *build_root_node(struct node *dt, char *name)
node              798 scripts/dtc/livetree.c 	struct node *an;
node              810 scripts/dtc/livetree.c static bool any_label_tree(struct dt_info *dti, struct node *node)
node              812 scripts/dtc/livetree.c 	struct node *c;
node              814 scripts/dtc/livetree.c 	if (node->labels)
node              817 scripts/dtc/livetree.c 	for_each_child(node, c)
node              825 scripts/dtc/livetree.c 					 struct node *an, struct node *node,
node              828 scripts/dtc/livetree.c 	struct node *dt = dti->dt;
node              829 scripts/dtc/livetree.c 	struct node *c;
node              834 scripts/dtc/livetree.c 	if (node->labels) {
node              837 scripts/dtc/livetree.c 		for_each_label(node->labels, l) {
node              850 scripts/dtc/livetree.c 				data_copy_escape_string(node->fullpath,
node              851 scripts/dtc/livetree.c 						strlen(node->fullpath)),
node              858 scripts/dtc/livetree.c 			(void)get_node_phandle(dt, node);
node              861 scripts/dtc/livetree.c 	for_each_child(node, c)
node              865 scripts/dtc/livetree.c static bool any_fixup_tree(struct dt_info *dti, struct node *node)
node              867 scripts/dtc/livetree.c 	struct node *c;
node              871 scripts/dtc/livetree.c 	for_each_property(node, prop) {
node              879 scripts/dtc/livetree.c 	for_each_child(node, c) {
node              887 scripts/dtc/livetree.c static void add_fixup_entry(struct dt_info *dti, struct node *fn,
node              888 scripts/dtc/livetree.c 			    struct node *node, struct property *prop,
node              897 scripts/dtc/livetree.c 	if (strchr(node->fullpath, ':') || strchr(prop->name, ':'))
node              901 scripts/dtc/livetree.c 			node->fullpath, prop->name, m->offset);
node              908 scripts/dtc/livetree.c 					  struct node *fn,
node              909 scripts/dtc/livetree.c 					  struct node *node)
node              911 scripts/dtc/livetree.c 	struct node *dt = dti->dt;
node              912 scripts/dtc/livetree.c 	struct node *c;
node              915 scripts/dtc/livetree.c 	struct node *refnode;
node              917 scripts/dtc/livetree.c 	for_each_property(node, prop) {
node              922 scripts/dtc/livetree.c 				add_fixup_entry(dti, fn, node, prop, m);
node              926 scripts/dtc/livetree.c 	for_each_child(node, c)
node              930 scripts/dtc/livetree.c static bool any_local_fixup_tree(struct dt_info *dti, struct node *node)
node              932 scripts/dtc/livetree.c 	struct node *c;
node              936 scripts/dtc/livetree.c 	for_each_property(node, prop) {
node              944 scripts/dtc/livetree.c 	for_each_child(node, c) {
node              953 scripts/dtc/livetree.c 		struct node *lfn, struct node *node,
node              955 scripts/dtc/livetree.c 		struct node *refnode)
node              957 scripts/dtc/livetree.c 	struct node *wn, *nwn;	/* local fixup node, walk node, new */
node              964 scripts/dtc/livetree.c 	for (wn = node; wn; wn = wn->parent)
node              971 scripts/dtc/livetree.c 	for (wn = node, i = depth - 1; wn; wn = wn->parent, i--)
node              989 scripts/dtc/livetree.c 						struct node *lfn,
node              990 scripts/dtc/livetree.c 						struct node *node)
node              992 scripts/dtc/livetree.c 	struct node *dt = dti->dt;
node              993 scripts/dtc/livetree.c 	struct node *c;
node              996 scripts/dtc/livetree.c 	struct node *refnode;
node              998 scripts/dtc/livetree.c 	for_each_property(node, prop) {
node             1003 scripts/dtc/livetree.c 				add_local_fixup_entry(dti, lfn, node, prop, m, refnode);
node             1007 scripts/dtc/livetree.c 	for_each_child(node, c)
node              130 scripts/dtc/srcpos.c 	struct search_path *node;
node              140 scripts/dtc/srcpos.c 	for (node = search_path_head; !*fp && node; node = node->next)
node              141 scripts/dtc/srcpos.c 		fullname = try_open(node->dirname, fname, fp);
node              217 scripts/dtc/srcpos.c 	struct search_path *node;
node              220 scripts/dtc/srcpos.c 	node = xmalloc(sizeof(*node));
node              221 scripts/dtc/srcpos.c 	node->next = NULL;
node              222 scripts/dtc/srcpos.c 	node->dirname = xstrdup(dirname);
node              226 scripts/dtc/srcpos.c 		*search_path_tail = node;
node              228 scripts/dtc/srcpos.c 		search_path_head = node;
node              229 scripts/dtc/srcpos.c 	search_path_tail = &node->next;
node              281 scripts/dtc/treesource.c static void write_tree_source_node(FILE *f, struct node *tree, int level)
node              284 scripts/dtc/treesource.c 	struct node *child;
node              176 scripts/dtc/yamltree.c static void yaml_tree(struct node *tree, yaml_emitter_t *emitter)
node              179 scripts/dtc/yamltree.c 	struct node *child;
node              162 scripts/gcc-plugins/gcc-common.h #define DECL_NAME_POINTER(node) IDENTIFIER_POINTER(DECL_NAME(node))
node              163 scripts/gcc-plugins/gcc-common.h #define DECL_NAME_LENGTH(node) IDENTIFIER_LENGTH(DECL_NAME(node))
node              164 scripts/gcc-plugins/gcc-common.h #define TYPE_NAME_POINTER(node) IDENTIFIER_POINTER(TYPE_NAME(node))
node              165 scripts/gcc-plugins/gcc-common.h #define TYPE_NAME_LENGTH(node) IDENTIFIER_LENGTH(TYPE_NAME(node))
node              312 scripts/gcc-plugins/gcc-common.h 	struct cgraph_node *node = cgraph_get_node(decl);
node              314 scripts/gcc-plugins/gcc-common.h 	return node ? node : cgraph_node(decl);
node              317 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_function_with_gimple_body_p(struct cgraph_node *node)
node              319 scripts/gcc-plugins/gcc-common.h 	return node->analyzed && !node->thunk.thunk_p && !node->alias;
node              324 scripts/gcc-plugins/gcc-common.h 	struct cgraph_node *node;
node              326 scripts/gcc-plugins/gcc-common.h 	for (node = cgraph_nodes; node; node = node->next)
node              327 scripts/gcc-plugins/gcc-common.h 		if (cgraph_function_with_gimple_body_p(node))
node              328 scripts/gcc-plugins/gcc-common.h 			return node;
node              332 scripts/gcc-plugins/gcc-common.h static inline struct cgraph_node *cgraph_next_function_with_gimple_body(struct cgraph_node *node)
node              334 scripts/gcc-plugins/gcc-common.h 	for (node = node->next; node; node = node->next)
node              335 scripts/gcc-plugins/gcc-common.h 		if (cgraph_function_with_gimple_body_p(node))
node              336 scripts/gcc-plugins/gcc-common.h 			return node;
node              340 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_for_node_and_aliases(cgraph_node_ptr node, bool (*callback)(cgraph_node_ptr, void *), void *data, bool include_overwritable)
node              344 scripts/gcc-plugins/gcc-common.h 	if (callback(node, data))
node              347 scripts/gcc-plugins/gcc-common.h 	for (alias = node->same_body; alias; alias = alias->next) {
node              356 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_FUNCTION_WITH_GIMPLE_BODY(node) \
node              357 scripts/gcc-plugins/gcc-common.h 	for ((node) = cgraph_first_function_with_gimple_body(); (node); \
node              358 scripts/gcc-plugins/gcc-common.h 		(node) = cgraph_next_function_with_gimple_body(node))
node              367 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_FUNCTION(node)	\
node              368 scripts/gcc-plugins/gcc-common.h 	for (node = cgraph_nodes; node; node = node->next)
node              369 scripts/gcc-plugins/gcc-common.h #define FOR_EACH_VARIABLE(node)	\
node              370 scripts/gcc-plugins/gcc-common.h 	for (node = varpool_nodes; node; node = node->next)
node              372 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (node)
node              373 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->decl
node              411 scripts/gcc-plugins/gcc-common.h #define NODE_IMPLICIT_ALIAS(node)	(node)->same_body_alias
node              547 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (&(node)->symbol)
node              548 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->symbol.decl
node              554 scripts/gcc-plugins/gcc-common.h #define varpool_mark_needed_node(node)
node              579 scripts/gcc-plugins/gcc-common.h static inline void set_decl_section_name(tree node, const char *value)
node              582 scripts/gcc-plugins/gcc-common.h 		DECL_SECTION_NAME(node) = build_string(strlen(value) + 1, value);
node              584 scripts/gcc-plugins/gcc-common.h 		DECL_SECTION_NAME(node) = NULL;
node              681 scripts/gcc-plugins/gcc-common.h #define NODE_SYMBOL(node) (node)
node              682 scripts/gcc-plugins/gcc-common.h #define NODE_DECL(node) (node)->decl
node              683 scripts/gcc-plugins/gcc-common.h #define cgraph_node_name(node) (node)->name()
node              684 scripts/gcc-plugins/gcc-common.h #define NODE_IMPLICIT_ALIAS(node) (node)->cpp_implicit_alias
node              716 scripts/gcc-plugins/gcc-common.h #define debug_cgraph_node(node) (node)->debug()
node              723 scripts/gcc-plugins/gcc-common.h #define dump_varpool_node(file, node) (node)->dump(file)
node              767 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_function_node(cgraph_node_ptr node, enum availability *availability)
node              769 scripts/gcc-plugins/gcc-common.h 	return node->function_symbol(availability);
node              772 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_function_or_thunk_node(cgraph_node_ptr node, enum availability *availability = NULL)
node              774 scripts/gcc-plugins/gcc-common.h 	return node->ultimate_alias_target(availability);
node              777 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_only_called_directly_p(cgraph_node_ptr node)
node              779 scripts/gcc-plugins/gcc-common.h 	return node->only_called_directly_p();
node              782 scripts/gcc-plugins/gcc-common.h static inline enum availability cgraph_function_body_availability(cgraph_node_ptr node)
node              784 scripts/gcc-plugins/gcc-common.h 	return node->get_availability();
node              787 scripts/gcc-plugins/gcc-common.h static inline cgraph_node_ptr cgraph_alias_target(cgraph_node_ptr node)
node              789 scripts/gcc-plugins/gcc-common.h 	return node->get_alias_target();
node              792 scripts/gcc-plugins/gcc-common.h static inline bool cgraph_for_node_and_aliases(cgraph_node_ptr node, bool (*callback)(cgraph_node_ptr, void *), void *data, bool include_overwritable)
node              794 scripts/gcc-plugins/gcc-common.h 	return node->call_for_symbol_thunks_and_aliases(callback, data, include_overwritable);
node              827 scripts/gcc-plugins/gcc-common.h static inline void cgraph_call_node_duplication_hooks(cgraph_node_ptr node, cgraph_node_ptr node2)
node              829 scripts/gcc-plugins/gcc-common.h 	symtab->call_cgraph_duplication_hooks(node, node2);
node              122 scripts/gcc-plugins/latent_entropy_plugin.c static tree handle_latent_entropy_attribute(tree *node, tree name,
node              134 scripts/gcc-plugins/latent_entropy_plugin.c 	switch (TREE_CODE(*node)) {
node              142 scripts/gcc-plugins/latent_entropy_plugin.c 		if (DECL_INITIAL(*node)) {
node              145 scripts/gcc-plugins/latent_entropy_plugin.c 				*node, name);
node              149 scripts/gcc-plugins/latent_entropy_plugin.c 		if (!TREE_STATIC(*node)) {
node              152 scripts/gcc-plugins/latent_entropy_plugin.c 				*node, name);
node              156 scripts/gcc-plugins/latent_entropy_plugin.c 		type = TREE_TYPE(*node);
node              161 scripts/gcc-plugins/latent_entropy_plugin.c 				*node, name);
node              177 scripts/gcc-plugins/latent_entropy_plugin.c 					*node, name, fld);
node              198 scripts/gcc-plugins/latent_entropy_plugin.c 			DECL_INITIAL(*node) = build_constructor(type, vals);
node              204 scripts/gcc-plugins/latent_entropy_plugin.c 			DECL_INITIAL(*node) = tree_get_random_const(type);
node              220 scripts/gcc-plugins/latent_entropy_plugin.c 					*node, name);
node              245 scripts/gcc-plugins/latent_entropy_plugin.c 			DECL_INITIAL(*node) = build_constructor(type, vals);
node              485 scripts/gcc-plugins/latent_entropy_plugin.c 	varpool_node_ptr node;
node              490 scripts/gcc-plugins/latent_entropy_plugin.c 	FOR_EACH_VARIABLE(node) {
node              491 scripts/gcc-plugins/latent_entropy_plugin.c 		tree name, var = NODE_DECL(node);
node               26 scripts/gcc-plugins/randomize_layout_plugin.c #define ORIG_TYPE_NAME(node) \
node               27 scripts/gcc-plugins/randomize_layout_plugin.c 	(TYPE_NAME(TYPE_MAIN_VARIANT(node)) != NULL_TREE ? ((const unsigned char *)IDENTIFIER_POINTER(TYPE_NAME(TYPE_MAIN_VARIANT(node)))) : (const unsigned char *)"anonymous")
node               76 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_layout_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
node               81 scripts/gcc-plugins/randomize_layout_plugin.c 	if (TREE_CODE(*node) == FUNCTION_DECL) {
node               82 scripts/gcc-plugins/randomize_layout_plugin.c 		error("%qE attribute does not apply to functions (%qF)", name, *node);
node               86 scripts/gcc-plugins/randomize_layout_plugin.c 	if (TREE_CODE(*node) == PARM_DECL) {
node               87 scripts/gcc-plugins/randomize_layout_plugin.c 		error("%qE attribute does not apply to function parameters (%qD)", name, *node);
node               91 scripts/gcc-plugins/randomize_layout_plugin.c 	if (TREE_CODE(*node) == VAR_DECL) {
node               92 scripts/gcc-plugins/randomize_layout_plugin.c 		error("%qE attribute does not apply to variables (%qD)", name, *node);
node               96 scripts/gcc-plugins/randomize_layout_plugin.c 	if (TYPE_P(*node)) {
node               97 scripts/gcc-plugins/randomize_layout_plugin.c 		type = *node;
node               99 scripts/gcc-plugins/randomize_layout_plugin.c 		gcc_assert(TREE_CODE(*node) == TYPE_DECL);
node              100 scripts/gcc-plugins/randomize_layout_plugin.c 		type = TREE_TYPE(*node);
node              119 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_considered_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
node              129 scripts/gcc-plugins/randomize_layout_plugin.c static tree handle_randomize_performed_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
node              433 scripts/gcc-plugins/randomize_layout_plugin.c static int is_pure_ops_struct(const_tree node)
node              437 scripts/gcc-plugins/randomize_layout_plugin.c 	gcc_assert(TREE_CODE(node) == RECORD_TYPE || TREE_CODE(node) == UNION_TYPE);
node              439 scripts/gcc-plugins/randomize_layout_plugin.c 	for (field = TYPE_FIELDS(node); field; field = TREE_CHAIN(field)) {
node              443 scripts/gcc-plugins/randomize_layout_plugin.c 		if (node == fieldtype)
node              662 scripts/gcc-plugins/randomize_layout_plugin.c 	struct varpool_node *node;
node              665 scripts/gcc-plugins/randomize_layout_plugin.c 	FOR_EACH_VARIABLE(node) {
node              666 scripts/gcc-plugins/randomize_layout_plugin.c 		tree var = NODE_DECL(node);
node               53 scripts/gcc-plugins/stackleak_plugin.c 	cgraph_node_ptr node;
node               69 scripts/gcc-plugins/stackleak_plugin.c 	node = cgraph_get_create_node(track_function_decl);
node               70 scripts/gcc-plugins/stackleak_plugin.c 	gcc_assert(node);
node               72 scripts/gcc-plugins/stackleak_plugin.c 	cgraph_create_edge(cgraph_get_node(current_function_decl), node,
node               53 scripts/gcc-plugins/structleak_plugin.c static tree handle_user_attribute(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
node               58 scripts/gcc-plugins/structleak_plugin.c 	if (TREE_CODE(*node) != FIELD_DECL)
node              323 scripts/genksyms/genksyms.c void free_node(struct string_list *node)
node              325 scripts/genksyms/genksyms.c 	free(node->string);
node              326 scripts/genksyms/genksyms.c 	free(node);
node              367 scripts/genksyms/genksyms.c struct string_list *copy_node(struct string_list *node)
node              372 scripts/genksyms/genksyms.c 	newnode->string = xstrdup(node->string);
node              373 scripts/genksyms/genksyms.c 	newnode->tag = node->tag;
node              411 scripts/genksyms/genksyms.c 	struct string_list node = {
node              418 scripts/genksyms/genksyms.c 			if (node.string == buffer)
node              424 scripts/genksyms/genksyms.c 			if (node.string == buffer)
node              429 scripts/genksyms/genksyms.c 		if (node.string >= buffer + sizeof(buffer) - 1) {
node              433 scripts/genksyms/genksyms.c 		*node.string++ = c;
node              435 scripts/genksyms/genksyms.c 	if (node.string == buffer)
node              437 scripts/genksyms/genksyms.c 	*node.string = 0;
node              438 scripts/genksyms/genksyms.c 	node.string = buffer;
node              440 scripts/genksyms/genksyms.c 	if (node.string[1] == '#') {
node              444 scripts/genksyms/genksyms.c 			if (node.string[0] == symbol_types[n].n) {
node              445 scripts/genksyms/genksyms.c 				node.tag = n;
node              446 scripts/genksyms/genksyms.c 				node.string += 2;
node              447 scripts/genksyms/genksyms.c 				return copy_node(&node);
node              450 scripts/genksyms/genksyms.c 		fprintf(stderr, "Unknown type %c\n", node.string[0]);
node              453 scripts/genksyms/genksyms.c 	return copy_node(&node);
node               29 scripts/genksyms/parse.y   struct string_list *node = *p;
node               30 scripts/genksyms/parse.y   *p = node->next;
node               31 scripts/genksyms/parse.y   free_node(node);
node               73 scripts/kconfig/gconf.c static void set_node(GtkTreeIter * node, struct menu *menu, gchar ** row);
node             1161 scripts/kconfig/gconf.c static void set_node(GtkTreeIter * node, struct menu *menu, gchar ** row)
node             1174 scripts/kconfig/gconf.c 	gtk_tree_store_set(tree, node,
node             1200 scripts/kconfig/gconf.c 	GtkTreeIter *node = parents[indent];
node             1202 scripts/kconfig/gconf.c 	gtk_tree_store_append(tree, node, parent);
node             1203 scripts/kconfig/gconf.c 	set_node(node, menu, row);
node              162 scripts/kconfig/lxdialog/dialog.h 	struct dialog_item node;
node              606 scripts/kconfig/lxdialog/util.c 	vsnprintf(item_cur->node.str, sizeof(item_cur->node.str), fmt, ap);
node              615 scripts/kconfig/lxdialog/util.c 	avail = sizeof(item_cur->node.str) - strlen(item_cur->node.str);
node              618 scripts/kconfig/lxdialog/util.c 	vsnprintf(item_cur->node.str + strlen(item_cur->node.str),
node              620 scripts/kconfig/lxdialog/util.c 	item_cur->node.str[sizeof(item_cur->node.str) - 1] = '\0';
node              626 scripts/kconfig/lxdialog/util.c 	item_cur->node.tag = tag;
node              630 scripts/kconfig/lxdialog/util.c 	item_cur->node.data = ptr;
node              635 scripts/kconfig/lxdialog/util.c 	item_cur->node.selected = val;
node              648 scripts/kconfig/lxdialog/util.c 	return item_cur->node.data;
node              653 scripts/kconfig/lxdialog/util.c 	return item_cur->node.tag;
node              689 scripts/kconfig/lxdialog/util.c 	return item_cur->node.str;
node              694 scripts/kconfig/lxdialog/util.c 	return (item_cur->node.selected != 0);
node              699 scripts/kconfig/lxdialog/util.c 	return (item_cur->node.tag == tag);
node               41 scripts/kconfig/preprocess.c 	struct list_head node;
node               52 scripts/kconfig/preprocess.c 	list_add_tail(&e->node, &env_list);
node               57 scripts/kconfig/preprocess.c 	list_del(&e->node);
node               72 scripts/kconfig/preprocess.c 	list_for_each_entry(e, &env_list, node) {
node               94 scripts/kconfig/preprocess.c 	list_for_each_entry_safe(e, tmp, &env_list, node) {
node              236 scripts/kconfig/preprocess.c 	struct list_head node;
node              243 scripts/kconfig/preprocess.c 	list_for_each_entry(v, &variable_list, node) {
node              303 scripts/kconfig/preprocess.c 		list_add_tail(&v->node, &variable_list);
node              326 scripts/kconfig/preprocess.c 	list_del(&v->node);
node              336 scripts/kconfig/preprocess.c 	list_for_each_entry_safe(v, tmp, &variable_list, node)
node              127 security/apparmor/include/label.h 	struct rb_node node;
node              412 security/apparmor/label.c 	RB_CLEAR_NODE(&label->node);
node              578 security/apparmor/label.c 		rb_erase(&label->node, &ls->root);
node              613 security/apparmor/label.c 		rb_replace_node(&old->node, &new->node, &ls->root);
node              649 security/apparmor/label.c 		struct aa_label *this = rb_entry(*new, struct aa_label, node);
node              673 security/apparmor/label.c 	rb_link_node(&label->node, parent, new);
node              674 security/apparmor/label.c 	rb_insert_color(&label->node, &ls->root);
node              694 security/apparmor/label.c 	struct rb_node *node;
node              700 security/apparmor/label.c 	node = vec_labelset(vec, n)->root.rb_node;
node              701 security/apparmor/label.c 	while (node) {
node              702 security/apparmor/label.c 		struct aa_label *this = rb_entry(node, struct aa_label, node);
node              706 security/apparmor/label.c 			node = node->rb_left;
node              708 security/apparmor/label.c 			node = node->rb_right;
node             1116 security/apparmor/label.c 	struct rb_node *node;
node             1125 security/apparmor/label.c 	node  = ls->root.rb_node;
node             1126 security/apparmor/label.c 	while (node) {
node             1127 security/apparmor/label.c 		struct aa_label *this = container_of(node, struct aa_label,
node             1128 security/apparmor/label.c 						     node);
node             1132 security/apparmor/label.c 			node = node->rb_left;
node             1134 security/apparmor/label.c 			node = node->rb_right;
node             1954 security/apparmor/label.c 	struct rb_node *node;
node             1960 security/apparmor/label.c 	for (node = rb_first(&ls->root); node; node = rb_first(&ls->root)) {
node             1961 security/apparmor/label.c 		struct aa_label *this = rb_entry(node, struct aa_label, node);
node             1986 security/apparmor/label.c 	struct rb_node *node;
node             1993 security/apparmor/label.c 	__labelset_for_each(ls, node) {
node             1994 security/apparmor/label.c 		label = rb_entry(node, struct aa_label, node);
node               98 security/integrity/iint.c 	struct rb_node *node, *parent = NULL;
node              123 security/integrity/iint.c 	node = &iint->rb_node;
node              125 security/integrity/iint.c 	rb_link_node(node, parent, p);
node              126 security/integrity/iint.c 	rb_insert_color(node, &integrity_iint_tree);
node               54 security/keys/internal.h 	struct rb_node		node;
node               63 security/keys/key.c 		user = rb_entry(parent, struct key_user, node);
node              101 security/keys/key.c 	rb_link_node(&candidate->node, parent, p);
node              102 security/keys/key.c 	rb_insert_color(&candidate->node, &key_user_tree);
node              122 security/keys/key.c 		rb_erase(&user->node, &key_user_tree);
node             1190 security/keys/key.c 	rb_link_node(&root_key_user.node,
node             1194 security/keys/key.c 	rb_insert_color(&root_key_user.node,
node              668 security/keys/keyring.c 		struct assoc_array_node *node;
node              673 security/keys/keyring.c 	struct assoc_array_node *node;
node              746 security/keys/keyring.c 		node = assoc_array_ptr_to_node(ptr);
node              750 security/keys/keyring.c 	node = assoc_array_ptr_to_node(ptr);
node              751 security/keys/keyring.c 	ptr = node->slots[0];
node              765 security/keys/keyring.c 	node = assoc_array_ptr_to_node(ptr);
node              773 security/keys/keyring.c 		ptr = READ_ONCE(node->slots[slot]);
node              775 security/keys/keyring.c 		if (assoc_array_ptr_is_meta(ptr) && node->back_pointer)
node              799 security/keys/keyring.c 		stack[sp].node = node;
node              811 security/keys/keyring.c 	ptr = READ_ONCE(node->back_pointer);
node              812 security/keys/keyring.c 	slot = node->parent_slot;
node              821 security/keys/keyring.c 	node = assoc_array_ptr_to_node(ptr);
node              828 security/keys/keyring.c 	if (node->back_pointer) {
node              846 security/keys/keyring.c 	node = stack[sp].node;
node              255 security/keys/proc.c 		struct key_user *user = rb_entry(n, struct key_user, node);
node              306 security/keys/proc.c 	struct key_user *user = rb_entry(_p, struct key_user, node);
node              153 security/selinux/avc.c 	struct avc_node *node;
node              165 security/selinux/avc.c 			hlist_for_each_entry_rcu(node, head, list)
node              326 security/selinux/avc.c static int avc_add_xperms_decision(struct avc_node *node,
node              331 security/selinux/avc.c 	node->ae.xp_node->xp.len++;
node              336 security/selinux/avc.c 	list_add(&dest_xpd->xpd_list, &node->ae.xp_node->xpd_head);
node              351 security/selinux/avc.c static int avc_xperms_populate(struct avc_node *node,
node              375 security/selinux/avc.c 	node->ae.xp_node = dest;
node              432 security/selinux/avc.c 	struct avc_node *node = container_of(rhead, struct avc_node, rhead);
node              433 security/selinux/avc.c 	avc_xperms_free(node->ae.xp_node);
node              434 security/selinux/avc.c 	kmem_cache_free(avc_node_cachep, node);
node              438 security/selinux/avc.c static void avc_node_delete(struct selinux_avc *avc, struct avc_node *node)
node              440 security/selinux/avc.c 	hlist_del_rcu(&node->list);
node              441 security/selinux/avc.c 	call_rcu(&node->rhead, avc_node_free);
node              445 security/selinux/avc.c static void avc_node_kill(struct selinux_avc *avc, struct avc_node *node)
node              447 security/selinux/avc.c 	avc_xperms_free(node->ae.xp_node);
node              448 security/selinux/avc.c 	kmem_cache_free(avc_node_cachep, node);
node              463 security/selinux/avc.c 	struct avc_node *node;
node              479 security/selinux/avc.c 		hlist_for_each_entry(node, head, list) {
node              480 security/selinux/avc.c 			avc_node_delete(avc, node);
node              498 security/selinux/avc.c 	struct avc_node *node;
node              500 security/selinux/avc.c 	node = kmem_cache_zalloc(avc_node_cachep, GFP_NOWAIT);
node              501 security/selinux/avc.c 	if (!node)
node              504 security/selinux/avc.c 	INIT_HLIST_NODE(&node->list);
node              512 security/selinux/avc.c 	return node;
node              515 security/selinux/avc.c static void avc_node_populate(struct avc_node *node, u32 ssid, u32 tsid, u16 tclass, struct av_decision *avd)
node              517 security/selinux/avc.c 	node->ae.ssid = ssid;
node              518 security/selinux/avc.c 	node->ae.tsid = tsid;
node              519 security/selinux/avc.c 	node->ae.tclass = tclass;
node              520 security/selinux/avc.c 	memcpy(&node->ae.avd, avd, sizeof(node->ae.avd));
node              526 security/selinux/avc.c 	struct avc_node *node, *ret = NULL;
node              532 security/selinux/avc.c 	hlist_for_each_entry_rcu(node, head, list) {
node              533 security/selinux/avc.c 		if (ssid == node->ae.ssid &&
node              534 security/selinux/avc.c 		    tclass == node->ae.tclass &&
node              535 security/selinux/avc.c 		    tsid == node->ae.tsid) {
node              536 security/selinux/avc.c 			ret = node;
node              559 security/selinux/avc.c 	struct avc_node *node;
node              562 security/selinux/avc.c 	node = avc_search_node(avc, ssid, tsid, tclass);
node              564 security/selinux/avc.c 	if (node)
node              565 security/selinux/avc.c 		return node;
node              617 security/selinux/avc.c 	struct avc_node *pos, *node = NULL;
node              626 security/selinux/avc.c 	node = avc_alloc_node(avc);
node              627 security/selinux/avc.c 	if (!node)
node              630 security/selinux/avc.c 	avc_node_populate(node, ssid, tsid, tclass, avd);
node              631 security/selinux/avc.c 	if (avc_xperms_populate(node, xp_node)) {
node              632 security/selinux/avc.c 		avc_node_kill(avc, node);
node              644 security/selinux/avc.c 			avc_node_replace(avc, node, pos);
node              648 security/selinux/avc.c 	hlist_add_head_rcu(&node->list, head);
node              651 security/selinux/avc.c 	return node;
node              836 security/selinux/avc.c 	struct avc_node *pos, *node, *orig = NULL;
node              855 security/selinux/avc.c 	node = avc_alloc_node(avc);
node              856 security/selinux/avc.c 	if (!node) {
node              881 security/selinux/avc.c 		avc_node_kill(avc, node);
node              889 security/selinux/avc.c 	avc_node_populate(node, ssid, tsid, tclass, &orig->ae.avd);
node              892 security/selinux/avc.c 		rc = avc_xperms_populate(node, orig->ae.xp_node);
node              894 security/selinux/avc.c 			avc_node_kill(avc, node);
node              901 security/selinux/avc.c 		node->ae.avd.allowed |= perms;
node              902 security/selinux/avc.c 		if (node->ae.xp_node && (flags & AVC_EXTENDED_PERMS))
node              903 security/selinux/avc.c 			avc_xperms_allow_perm(node->ae.xp_node, driver, xperm);
node              907 security/selinux/avc.c 		node->ae.avd.allowed &= ~perms;
node              910 security/selinux/avc.c 		node->ae.avd.auditallow |= perms;
node              913 security/selinux/avc.c 		node->ae.avd.auditallow &= ~perms;
node              916 security/selinux/avc.c 		node->ae.avd.auditdeny |= perms;
node              919 security/selinux/avc.c 		node->ae.avd.auditdeny &= ~perms;
node              922 security/selinux/avc.c 		avc_add_xperms_decision(node, xpd);
node              925 security/selinux/avc.c 	avc_node_replace(avc, node, orig);
node              938 security/selinux/avc.c 	struct avc_node *node;
node              953 security/selinux/avc.c 		hlist_for_each_entry(node, head, list)
node              954 security/selinux/avc.c 			avc_node_delete(avc, node);
node             1036 security/selinux/avc.c 	struct avc_node *node;
node             1054 security/selinux/avc.c 	node = avc_lookup(state->avc, ssid, tsid, tclass);
node             1055 security/selinux/avc.c 	if (unlikely(!node)) {
node             1056 security/selinux/avc.c 		node = avc_compute_av(state, ssid, tsid, tclass, &avd, xp_node);
node             1058 security/selinux/avc.c 		memcpy(&avd, &node->ae.avd, sizeof(avd));
node             1059 security/selinux/avc.c 		xp_node = node->ae.xp_node;
node             1135 security/selinux/avc.c 	struct avc_node *node;
node             1145 security/selinux/avc.c 	node = avc_lookup(state->avc, ssid, tsid, tclass);
node             1146 security/selinux/avc.c 	if (unlikely(!node))
node             1147 security/selinux/avc.c 		node = avc_compute_av(state, ssid, tsid, tclass, avd, &xp_node);
node             1149 security/selinux/avc.c 		memcpy(avd, &node->ae.avd, sizeof(*avd));
node              107 security/selinux/netnode.c 	struct sel_netnode *node;
node              121 security/selinux/netnode.c 	list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list)
node              122 security/selinux/netnode.c 		if (node->nsec.family == family)
node              125 security/selinux/netnode.c 				if (node->nsec.addr.ipv4 == *(__be32 *)addr)
node              126 security/selinux/netnode.c 					return node;
node              129 security/selinux/netnode.c 				if (ipv6_addr_equal(&node->nsec.addr.ipv6,
node              131 security/selinux/netnode.c 					return node;
node              146 security/selinux/netnode.c static void sel_netnode_insert(struct sel_netnode *node)
node              150 security/selinux/netnode.c 	switch (node->nsec.family) {
node              152 security/selinux/netnode.c 		idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4);
node              155 security/selinux/netnode.c 		idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6);
node              164 security/selinux/netnode.c 	list_add_rcu(&node->list, &sel_netnode_hash[idx].list);
node              193 security/selinux/netnode.c 	struct sel_netnode *node;
node              197 security/selinux/netnode.c 	node = sel_netnode_find(addr, family);
node              198 security/selinux/netnode.c 	if (node != NULL) {
node              199 security/selinux/netnode.c 		*sid = node->nsec.sid;
node              252 security/selinux/netnode.c 	struct sel_netnode *node;
node              255 security/selinux/netnode.c 	node = sel_netnode_find(addr, family);
node              256 security/selinux/netnode.c 	if (node != NULL) {
node              257 security/selinux/netnode.c 		*sid = node->nsec.sid;
node              276 security/selinux/netnode.c 	struct sel_netnode *node, *node_tmp;
node              280 security/selinux/netnode.c 		list_for_each_entry_safe(node, node_tmp,
node              282 security/selinux/netnode.c 				list_del_rcu(&node->list);
node              283 security/selinux/netnode.c 				kfree_rcu(node, rcu);
node              249 security/selinux/ss/avtab.c avtab_search_node_next(struct avtab_node *node, int specified)
node              253 security/selinux/ss/avtab.c 	if (!node)
node              257 security/selinux/ss/avtab.c 	for (cur = node->next; cur; cur = cur->next) {
node              258 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              259 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
node              260 security/selinux/ss/avtab.c 		    node->key.target_class == cur->key.target_class &&
node              264 security/selinux/ss/avtab.c 		if (node->key.source_type < cur->key.source_type)
node              266 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              267 security/selinux/ss/avtab.c 		    node->key.target_type < cur->key.target_type)
node              269 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              270 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
node              271 security/selinux/ss/avtab.c 		    node->key.target_class < cur->key.target_class)
node              111 security/selinux/ss/avtab.h struct avtab_node *avtab_search_node_next(struct avtab_node *node, int specified);
node               88 security/selinux/ss/conditional.c int evaluate_cond_node(struct policydb *p, struct cond_node *node)
node               93 security/selinux/ss/conditional.c 	new_state = cond_evaluate_expr(p, node->expr);
node               94 security/selinux/ss/conditional.c 	if (new_state != node->cur_state) {
node               95 security/selinux/ss/conditional.c 		node->cur_state = new_state;
node               99 security/selinux/ss/conditional.c 		for (cur = node->true_list; cur; cur = cur->next) {
node              101 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
node              103 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
node              106 security/selinux/ss/conditional.c 		for (cur = node->false_list; cur; cur = cur->next) {
node              109 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
node              111 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
node              141 security/selinux/ss/conditional.c static void cond_node_destroy(struct cond_node *node)
node              145 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr; cur_expr = next_expr) {
node              149 security/selinux/ss/conditional.c 	cond_av_list_destroy(node->true_list);
node              150 security/selinux/ss/conditional.c 	cond_av_list_destroy(node->false_list);
node              151 security/selinux/ss/conditional.c 	kfree(node);
node              304 security/selinux/ss/conditional.c 					if (cur->node == node_ptr) {
node              335 security/selinux/ss/conditional.c 	list->node = node_ptr;
node              395 security/selinux/ss/conditional.c static int cond_read_node(struct policydb *p, struct cond_node *node, void *fp)
node              406 security/selinux/ss/conditional.c 	node->cur_state = le32_to_cpu(buf[0]);
node              431 security/selinux/ss/conditional.c 			node->expr = expr;
node              437 security/selinux/ss/conditional.c 	rc = cond_read_av_list(p, fp, &node->true_list, NULL);
node              440 security/selinux/ss/conditional.c 	rc = cond_read_av_list(p, fp, &node->false_list, node->true_list);
node              445 security/selinux/ss/conditional.c 	cond_node_destroy(node);
node              451 security/selinux/ss/conditional.c 	struct cond_node *node, *last = NULL;
node              468 security/selinux/ss/conditional.c 		node = kzalloc(sizeof(*node), GFP_KERNEL);
node              469 security/selinux/ss/conditional.c 		if (!node)
node              472 security/selinux/ss/conditional.c 		rc = cond_read_node(p, node, fp);
node              477 security/selinux/ss/conditional.c 			p->cond_list = node;
node              479 security/selinux/ss/conditional.c 			last->next = node;
node              480 security/selinux/ss/conditional.c 		last = node;
node              542 security/selinux/ss/conditional.c 		rc = avtab_write_item(p, cur_list->node, fp);
node              550 security/selinux/ss/conditional.c static int cond_write_node(struct policydb *p, struct cond_node *node,
node              558 security/selinux/ss/conditional.c 	buf[0] = cpu_to_le32(node->cur_state);
node              563 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next)
node              571 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) {
node              579 security/selinux/ss/conditional.c 	rc = cond_write_av_list(p, node->true_list, fp);
node              582 security/selinux/ss/conditional.c 	rc = cond_write_av_list(p, node->false_list, fp);
node              616 security/selinux/ss/conditional.c 	struct avtab_node *node;
node              621 security/selinux/ss/conditional.c 	for (node = avtab_search_node(ctab, key); node;
node              622 security/selinux/ss/conditional.c 			node = avtab_search_node_next(node, key->specified)) {
node              623 security/selinux/ss/conditional.c 		if (node->key.specified & AVTAB_ENABLED)
node              624 security/selinux/ss/conditional.c 			services_compute_xperms_decision(xpermd, node);
node              635 security/selinux/ss/conditional.c 	struct avtab_node *node;
node              640 security/selinux/ss/conditional.c 	for (node = avtab_search_node(ctab, key); node;
node              641 security/selinux/ss/conditional.c 				node = avtab_search_node_next(node, key->specified)) {
node              643 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_ALLOWED|AVTAB_ENABLED)))
node              644 security/selinux/ss/conditional.c 			avd->allowed |= node->datum.u.data;
node              646 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_AUDITDENY|AVTAB_ENABLED)))
node              652 security/selinux/ss/conditional.c 			avd->auditdeny &= node->datum.u.data;
node              654 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_AUDITALLOW|AVTAB_ENABLED)))
node              655 security/selinux/ss/conditional.c 			avd->auditallow |= node->datum.u.data;
node              656 security/selinux/ss/conditional.c 		if (xperms && (node->key.specified & AVTAB_ENABLED) &&
node              657 security/selinux/ss/conditional.c 				(node->key.specified & AVTAB_XPERMS))
node              658 security/selinux/ss/conditional.c 			services_compute_xperms_drivers(xperms, node);
node               42 security/selinux/ss/conditional.h 	struct avtab_node *node;
node               78 security/selinux/ss/conditional.h int evaluate_cond_node(struct policydb *p, struct cond_node *node);
node               37 security/selinux/ss/ebitmap.c 	n1 = e1->node;
node               38 security/selinux/ss/ebitmap.c 	n2 = e2->node;
node               57 security/selinux/ss/ebitmap.c 	n = src->node;
node               71 security/selinux/ss/ebitmap.c 			dst->node = new;
node               94 security/selinux/ss/ebitmap.c 	struct ebitmap_node *e_iter = ebmap->node;
node              173 security/selinux/ss/ebitmap.c 				ebmap->node = e_iter;
node              209 security/selinux/ss/ebitmap.c 	n1 = e1->node;
node              210 security/selinux/ss/ebitmap.c 	n2 = e2->node;
node              249 security/selinux/ss/ebitmap.c 	n = e->node;
node              264 security/selinux/ss/ebitmap.c 	n = e->node;
node              293 security/selinux/ss/ebitmap.c 					e->node = n->next;
node              320 security/selinux/ss/ebitmap.c 		new->next = e->node;
node              321 security/selinux/ss/ebitmap.c 		e->node = new;
node              334 security/selinux/ss/ebitmap.c 	n = e->node;
node              342 security/selinux/ss/ebitmap.c 	e->node = NULL;
node              378 security/selinux/ss/ebitmap.c 		e->node = NULL;
node              419 security/selinux/ss/ebitmap.c 				e->node = tmp;
node               41 security/selinux/ss/ebitmap.h 	struct ebitmap_node *node;	/* first node in the bitmap */
node               52 security/selinux/ss/ebitmap.h 	for (*n = e->node; *n; *n = (*n)->next) {
node               83 security/selinux/ss/ebitmap.h #define EBITMAP_NODE_INDEX(node, bit)	\
node               84 security/selinux/ss/ebitmap.h 	(((bit) - (node)->startbit) / EBITMAP_UNIT_SIZE)
node               85 security/selinux/ss/ebitmap.h #define EBITMAP_NODE_OFFSET(node, bit)	\
node               86 security/selinux/ss/ebitmap.h 	(((bit) - (node)->startbit) % EBITMAP_UNIT_SIZE)
node               41 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node               55 security/selinux/ss/mls.c 		ebitmap_for_each_positive_bit(e, node, i) {
node               96 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node              115 security/selinux/ss/mls.c 		ebitmap_for_each_positive_bit(e, node, i) {
node              454 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node              470 security/selinux/ss/mls.c 					      node, i) {
node             1648 security/selinux/ss/policydb.c 		struct ebitmap_node *node;
node             1659 security/selinux/ss/policydb.c 		ebitmap_for_each_positive_bit(&user->roles, node, bit) {
node             1684 security/selinux/ss/policydb.c 		struct ebitmap_node *node;
node             1695 security/selinux/ss/policydb.c 		ebitmap_for_each_positive_bit(&role->types, node, bit) {
node             2140 security/selinux/ss/policydb.c 				c->u.node.addr = nodebuf[0]; /* network order */
node             2141 security/selinux/ss/policydb.c 				c->u.node.mask = nodebuf[1]; /* network order */
node             2806 security/selinux/ss/policydb.c static int write_cons_helper(struct policydb *p, struct constraint_node *node,
node             2815 security/selinux/ss/policydb.c 	for (c = node; c; c = c->next) {
node             3132 security/selinux/ss/policydb.c 				nodebuf[0] = c->u.node.addr; /* network order */
node             3133 security/selinux/ss/policydb.c 				nodebuf[1] = c->u.node.mask; /* network order */
node              181 security/selinux/ss/policydb.h 		} node;		/* node information */
node              590 security/selinux/ss/services.c 		struct avtab_node *node)
node              594 security/selinux/ss/services.c 	if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) {
node              597 security/selinux/ss/services.c 			xperms->drivers.p[i] |= node->datum.u.xperms->perms.p[i];
node              598 security/selinux/ss/services.c 	} else if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) {
node              601 security/selinux/ss/services.c 					node->datum.u.xperms->driver);
node              605 security/selinux/ss/services.c 	if (node->key.specified & AVTAB_XPERMS_ALLOWED)
node              623 security/selinux/ss/services.c 	struct avtab_node *node;
node              657 security/selinux/ss/services.c 			for (node = avtab_search_node(&policydb->te_avtab,
node              659 security/selinux/ss/services.c 			     node;
node              660 security/selinux/ss/services.c 			     node = avtab_search_node_next(node, avkey.specified)) {
node              661 security/selinux/ss/services.c 				if (node->key.specified == AVTAB_ALLOWED)
node              662 security/selinux/ss/services.c 					avd->allowed |= node->datum.u.data;
node              663 security/selinux/ss/services.c 				else if (node->key.specified == AVTAB_AUDITALLOW)
node              664 security/selinux/ss/services.c 					avd->auditallow |= node->datum.u.data;
node              665 security/selinux/ss/services.c 				else if (node->key.specified == AVTAB_AUDITDENY)
node              666 security/selinux/ss/services.c 					avd->auditdeny &= node->datum.u.data;
node              667 security/selinux/ss/services.c 				else if (xperms && (node->key.specified & AVTAB_XPERMS))
node              668 security/selinux/ss/services.c 					services_compute_xperms_drivers(xperms, node);
node              945 security/selinux/ss/services.c 					struct avtab_node *node)
node              949 security/selinux/ss/services.c 	if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) {
node              950 security/selinux/ss/services.c 		if (xpermd->driver != node->datum.u.xperms->driver)
node              952 security/selinux/ss/services.c 	} else if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) {
node              953 security/selinux/ss/services.c 		if (!security_xperm_test(node->datum.u.xperms->perms.p,
node              960 security/selinux/ss/services.c 	if (node->key.specified == AVTAB_XPERMS_ALLOWED) {
node              962 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) {
node              966 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) {
node              969 security/selinux/ss/services.c 					node->datum.u.xperms->perms.p[i];
node              971 security/selinux/ss/services.c 	} else if (node->key.specified == AVTAB_XPERMS_AUDITALLOW) {
node              973 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) {
node              977 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) {
node              980 security/selinux/ss/services.c 					node->datum.u.xperms->perms.p[i];
node              982 security/selinux/ss/services.c 	} else if (node->key.specified == AVTAB_XPERMS_DONTAUDIT) {
node              984 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLDRIVER) {
node              988 security/selinux/ss/services.c 		if (node->datum.u.xperms->specified == AVTAB_XPERMS_IOCTLFUNCTION) {
node              991 security/selinux/ss/services.c 					node->datum.u.xperms->perms.p[i];
node             1010 security/selinux/ss/services.c 	struct avtab_node *node;
node             1063 security/selinux/ss/services.c 			for (node = avtab_search_node(&policydb->te_avtab,
node             1065 security/selinux/ss/services.c 			     node;
node             1066 security/selinux/ss/services.c 			     node = avtab_search_node_next(node, avkey.specified))
node             1067 security/selinux/ss/services.c 				services_compute_xperms_decision(xpermd, node);
node             1652 security/selinux/ss/services.c 	struct avtab_node *node;
node             1757 security/selinux/ss/services.c 		node = avtab_search_node(&policydb->te_cond_avtab, &avkey);
node             1758 security/selinux/ss/services.c 		for (; node; node = avtab_search_node_next(node, specified)) {
node             1759 security/selinux/ss/services.c 			if (node->key.specified & AVTAB_ENABLED) {
node             1760 security/selinux/ss/services.c 				avdatum = &node->datum;
node             2054 security/selinux/ss/services.c 	struct ebitmap_node *node;
node             2064 security/selinux/ss/services.c 	ebitmap_for_each_positive_bit(&p->policycaps, node, i) {
node             2487 security/selinux/ss/services.c 			if (c->u.node.addr == (addr & c->u.node.mask))
node               37 security/selinux/ss/services.h 				struct avtab_node *node);
node               40 security/selinux/ss/services.h 					struct avtab_node *node);
node              181 security/selinux/ss/sidtab.c 		struct sidtab_node_inner *node = entry.ptr_inner;
node              185 security/selinux/ss/sidtab.c 			rc = sidtab_find_context(node->entries[i],
node              193 security/selinux/ss/sidtab.c 		struct sidtab_node_leaf *node = entry.ptr_leaf;
node              197 security/selinux/ss/sidtab.c 			if (context_cmp(&node->entries[i].context, context)) {
node              461 security/selinux/ss/sidtab.c 		struct sidtab_node_inner *node = entry.ptr_inner;
node              463 security/selinux/ss/sidtab.c 		if (!node)
node              467 security/selinux/ss/sidtab.c 			sidtab_destroy_tree(node->entries[i], level - 1);
node              468 security/selinux/ss/sidtab.c 		kfree(node);
node              470 security/selinux/ss/sidtab.c 		struct sidtab_node_leaf *node = entry.ptr_leaf;
node              472 security/selinux/ss/sidtab.c 		if (!node)
node              476 security/selinux/ss/sidtab.c 			context_destroy(&node->entries[i].context);
node              477 security/selinux/ss/sidtab.c 		kfree(node);
node               34 security/yama/yama_lsm.c 	struct list_head node;
node              121 security/yama/yama_lsm.c 	list_for_each_entry_rcu(relation, &ptracer_relations, node) {
node              123 security/yama/yama_lsm.c 			list_del_rcu(&relation->node);
node              156 security/yama/yama_lsm.c 	list_for_each_entry_rcu(relation, &ptracer_relations, node) {
node              160 security/yama/yama_lsm.c 			list_replace_rcu(&relation->node, &added->node);
node              166 security/yama/yama_lsm.c 	list_add_rcu(&added->node, &ptracer_relations);
node              186 security/yama/yama_lsm.c 	list_for_each_entry_rcu(relation, &ptracer_relations, node) {
node              328 security/yama/yama_lsm.c 	list_for_each_entry_rcu(relation, &ptracer_relations, node) {
node               73 sound/ac97/bus.c 	struct device_node *node;
node               80 sound/ac97/bus.c 	for_each_child_of_node(ac97_ctrl->parent->of_node, node) {
node               81 sound/ac97/bus.c 		if ((idx != of_property_read_u32(node, "reg", &reg)) ||
node               82 sound/ac97/bus.c 		    !of_device_is_compatible(node, compat))
node               84 sound/ac97/bus.c 		return node;
node               72 sound/aoa/aoa-gpio.h 	struct device_node *node;
node               35 sound/aoa/aoa.h 	struct device_node *node;
node              996 sound/aoa/codecs/onyx.c 	struct device_node *node = client->dev.of_node;
node             1020 sound/aoa/codecs/onyx.c 	onyx->codec.node = of_node_get(node);
node             1037 sound/aoa/codecs/onyx.c 	of_node_put(onyx->codec.node);
node              882 sound/aoa/codecs/tas.c 	struct device_node *node = client->dev.of_node;
node              901 sound/aoa/codecs/tas.c 	tas->codec.node = of_node_get(node);
node              908 sound/aoa/codecs/tas.c 	       (unsigned int)client->addr, node);
node              922 sound/aoa/codecs/tas.c 	of_node_put(tas->codec.node);
node               20 sound/aoa/core/gpio-pmf.c 	rc = pmf_call_function(rt->node, #name "-mute", &args);	\
node               43 sound/aoa/core/gpio-pmf.c 	rc = pmf_call_function(rt->node, "hw-reset", &args);
node              188 sound/aoa/core/gpio-pmf.c 		err = pmf_register_irq_client(rt->node,
node              229 sound/aoa/core/gpio-pmf.c 	err = pmf_call_function(rt->node, name, &args);
node              777 sound/aoa/fabrics/layout.c 	if (of_node_name_eq(codec->node, "codec")) {
node              786 sound/aoa/fabrics/layout.c 		if (*ref != codec->node->phandle) {
node             1038 sound/aoa/fabrics/layout.c 	ldev->gpio.node = sound->parent;
node             2137 sound/pci/asihpi/hpifunc.c 	u32 node, index;
node             2139 sound/pci/asihpi/hpifunc.c 		HPI_MULTIPLEXER_SOURCE, &node,
node             2142 sound/pci/asihpi/hpifunc.c 		*source_node_type = (u16)node;
node              820 sound/ppc/pmac.c 		ppc_md.feature_call(PMAC_FTR_SOUND_CHIP_ENABLE, chip->node, 0, enable);
node              836 sound/ppc/pmac.c 	if (chip->node)
node              862 sound/ppc/pmac.c 	if (chip->node) {
node              872 sound/ppc/pmac.c 	of_node_put(chip->node);
node              897 sound/ppc/pmac.c 	for (mio = chip->node->parent; mio; mio = mio->parent) {
node              949 sound/ppc/pmac.c 	chip->node = of_find_node_by_name(NULL, "awacs");
node              950 sound/ppc/pmac.c 	sound = of_node_get(chip->node);
node              956 sound/ppc/pmac.c 	if (!chip->node)
node              957 sound/ppc/pmac.c 		chip->node = of_find_node_by_name(NULL, "davbus");
node              962 sound/ppc/pmac.c 	if (! chip->node) {
node              963 sound/ppc/pmac.c 		chip->node = of_find_node_by_name(NULL, "i2s-a");
node              964 sound/ppc/pmac.c 		if (chip->node && chip->node->parent &&
node              965 sound/ppc/pmac.c 		    chip->node->parent->parent) {
node              966 sound/ppc/pmac.c 			if (of_device_is_compatible(chip->node->parent->parent,
node              971 sound/ppc/pmac.c 	if (! chip->node)
node              976 sound/ppc/pmac.c 			if (sound->parent == chip->node)
node              980 sound/ppc/pmac.c 		of_node_put(chip->node);
node              981 sound/ppc/pmac.c 		chip->node = NULL;
node              995 sound/ppc/pmac.c 		of_node_put(chip->node);
node              996 sound/ppc/pmac.c 		chip->node = NULL;
node             1046 sound/ppc/pmac.c 	macio = macio_find(chip->node, macio_unknown);
node             1198 sound/ppc/pmac.c 	np = chip->node;
node             1302 sound/ppc/pmac.c 		for (mio = chip->node->parent; mio; mio = mio->parent) {
node               82 sound/ppc/pmac.h 	struct device_node *node;
node             1086 sound/ppc/tumbler.c 	struct device_node *node;
node             1092 sound/ppc/tumbler.c 		node = find_compatible_audio_device(device);
node             1094 sound/ppc/tumbler.c 		node = find_audio_device(device);
node             1095 sound/ppc/tumbler.c 	if (! node) {
node             1101 sound/ppc/tumbler.c 	base = of_get_property(node, "AAPL,address", NULL);
node             1103 sound/ppc/tumbler.c 		base = of_get_property(node, "reg", NULL);
node             1107 sound/ppc/tumbler.c 			of_node_put(node);
node             1118 sound/ppc/tumbler.c 	base = of_get_property(node, "audio-gpio-active-state", NULL);
node             1134 sound/ppc/tumbler.c 			prop = of_get_property(node, platform, NULL);
node             1150 sound/ppc/tumbler.c 	ret = irq_of_parse_and_map(node, 0);
node             1151 sound/ppc/tumbler.c 	of_node_put(node);
node             1354 sound/ppc/tumbler.c 	for_each_child_of_node(chip->node, np) {
node              311 sound/soc/codecs/max98504.c 	struct device_node *node = dev->of_node;
node              319 sound/soc/codecs/max98504.c 	if (node) {
node              320 sound/soc/codecs/max98504.c 		if (!of_property_read_u32(node, "maxim,brownout-threshold",
node              324 sound/soc/codecs/max98504.c 		of_property_read_u32(node, "maxim,brownout-attenuation",
node              326 sound/soc/codecs/max98504.c 		of_property_read_u32(node, "maxim,brownout-attack-hold-ms",
node              328 sound/soc/codecs/max98504.c 		of_property_read_u32(node, "maxim,brownout-timed-hold-ms",
node              330 sound/soc/codecs/max98504.c 		of_property_read_u32(node, "maxim,brownout-release-rate-ms",
node              197 sound/soc/codecs/twl4030.c 				   struct device_node *node)
node              201 sound/soc/codecs/twl4030.c 	of_property_read_u32(node, "ti,digimic_delay",
node              203 sound/soc/codecs/twl4030.c 	of_property_read_u32(node, "ti,ramp_delay_value",
node              205 sound/soc/codecs/twl4030.c 	of_property_read_u32(node, "ti,offset_cncl_path",
node              207 sound/soc/codecs/twl4030.c 	if (!of_property_read_u32(node, "ti,hs_extmute", &value))
node              210 sound/soc/codecs/twl4030.c 	pdata->hs_extmute_gpio = of_get_named_gpio(node,
node               62 sound/soc/generic/audio-graph-card.c 	struct device_node *node;
node               86 sound/soc/generic/audio-graph-card.c 		node = of_get_parent(ep);
node               87 sound/soc/generic/audio-graph-card.c 		reg = of_get_property(node, "reg", NULL);
node               88 sound/soc/generic/audio-graph-card.c 		of_node_put(node);
node               92 sound/soc/generic/audio-graph-card.c 	node = of_graph_get_port_parent(ep);
node              100 sound/soc/generic/audio-graph-card.c 	for_each_endpoint_of_node(node, endpoint) {
node              106 sound/soc/generic/audio-graph-card.c 	of_node_put(node);
node              118 sound/soc/generic/audio-graph-card.c 	struct device_node *node;
node              125 sound/soc/generic/audio-graph-card.c 	node = of_graph_get_port_parent(ep);
node              128 sound/soc/generic/audio-graph-card.c 	args.np		= node;
node              130 sound/soc/generic/audio-graph-card.c 	args.args_count	= (of_graph_get_endpoint_count(node) > 1);
node              155 sound/soc/generic/audio-graph-card.c 	dlc->of_node = node;
node              158 sound/soc/generic/audio-graph-card.c 		*is_single_link = of_graph_get_endpoint_count(node) == 1;
node              170 sound/soc/generic/audio-graph-card.c 	struct device_node *node = of_graph_get_port_parent(ep);
node              173 sound/soc/generic/audio-graph-card.c 	asoc_simple_parse_convert(dev, node,  PREFIX, adata);
node              180 sound/soc/generic/audio-graph-card.c 	of_node_put(node);
node              189 sound/soc/generic/audio-graph-card.c 	struct device_node *node	= of_graph_get_port_parent(ep);
node              198 sound/soc/generic/audio-graph-card.c 	of_node_put(node);
node              214 sound/soc/generic/audio-graph-card.c 	struct device_node *node;
node              226 sound/soc/generic/audio-graph-card.c 	node	= of_graph_get_port_parent(ep);
node              298 sound/soc/generic/audio-graph-card.c 		snd_soc_of_parse_node_prefix(node, cconf, codecs->of_node,
node              328 sound/soc/generic/audio-graph-card.c 	of_node_put(node);
node              420 sound/soc/generic/audio-graph-card.c 	struct device_node *node = dev->of_node;
node              431 sound/soc/generic/audio-graph-card.c 	of_for_each_phandle(&it, rc, node, "dais", NULL, 0) {
node              432 sound/soc/generic/audio-graph-card.c 		cpu_port = it.node;
node               56 sound/soc/generic/simple-card-utils.c 			     struct device_node *node,
node               65 sound/soc/generic/simple-card-utils.c 	daifmt = snd_soc_of_parse_daifmt(node, prefix,
node              161 sound/soc/generic/simple-card-utils.c 			  struct device_node *node,
node              174 sound/soc/generic/simple-card-utils.c 	clk = devm_get_clk_from_child(dev, node, NULL);
node              179 sound/soc/generic/simple-card-utils.c 	} else if (!of_property_read_u32(node, "system-clock-frequency", &val)) {
node              187 sound/soc/generic/simple-card-utils.c 	if (of_property_read_bool(node, "system-clock-direction-out"))
node              402 sound/soc/generic/simple-card-utils.c 	struct device_node *node = card->dev->of_node;
node              410 sound/soc/generic/simple-card-utils.c 	if (!of_property_read_bool(node, prop))
node              420 sound/soc/generic/simple-card-utils.c 	struct device_node *node = card->dev->of_node;
node              428 sound/soc/generic/simple-card-utils.c 	if (of_property_read_bool(node, prop))
node               31 sound/soc/generic/simple-card.c static int asoc_simple_parse_dai(struct device_node *node,
node               38 sound/soc/generic/simple-card.c 	if (!node)
node               45 sound/soc/generic/simple-card.c 	ret = of_parse_phandle_with_args(node, DAI, CELL, 0, &args);
node               68 sound/soc/generic/simple-card.c 	ret = snd_soc_of_get_dai_name(node, &dlc->dai_name);
node               85 sound/soc/generic/simple-card.c 	struct device_node *node = of_get_parent(np);
node               88 sound/soc/generic/simple-card.c 	asoc_simple_parse_convert(dev, node, PREFIX, adata);
node               89 sound/soc/generic/simple-card.c 	asoc_simple_parse_convert(dev, node, NULL,   adata);
node               92 sound/soc/generic/simple-card.c 	of_node_put(node);
node              101 sound/soc/generic/simple-card.c 	struct device_node *node = of_get_parent(cpu);
node              108 sound/soc/generic/simple-card.c 	of_property_read_u32(node,	prop, &props->mclk_fs);
node              112 sound/soc/generic/simple-card.c 	of_node_put(node);
node              128 sound/soc/generic/simple-card.c 	struct device_node *node = of_get_parent(np);
node              214 sound/soc/generic/simple-card.c 		snd_soc_of_parse_node_prefix(node, cconf, codecs->of_node,
node              229 sound/soc/generic/simple-card.c 	ret = asoc_simple_parse_daifmt(dev, node, codec,
node              240 sound/soc/generic/simple-card.c 	of_node_put(node);
node              257 sound/soc/generic/simple-card.c 	struct device_node *node = NULL;
node              273 sound/soc/generic/simple-card.c 	node = of_get_parent(np);
node              276 sound/soc/generic/simple-card.c 	dev_dbg(dev, "link_of (%pOF)\n", node);
node              283 sound/soc/generic/simple-card.c 	plat = of_get_child_by_name(node, prop);
node              290 sound/soc/generic/simple-card.c 	ret = asoc_simple_parse_daifmt(dev, node, codec,
node              340 sound/soc/generic/simple-card.c 	of_node_put(node);
node              358 sound/soc/generic/simple-card.c 	struct device_node *node;
node              364 sound/soc/generic/simple-card.c 	node = of_get_child_by_name(top, PREFIX "dai-link");
node              365 sound/soc/generic/simple-card.c 	if (!node) {
node              366 sound/soc/generic/simple-card.c 		node = of_node_get(top);
node              375 sound/soc/generic/simple-card.c 		int num = of_get_child_count(node);
node              378 sound/soc/generic/simple-card.c 		codec = of_get_child_by_name(node, is_top ?
node              387 sound/soc/generic/simple-card.c 		for_each_child_of_node(node, np)
node              391 sound/soc/generic/simple-card.c 		for_each_child_of_node(node, np) {
node              413 sound/soc/generic/simple-card.c 		node = of_get_next_child(top, node);
node              414 sound/soc/generic/simple-card.c 	} while (!is_top && node);
node              417 sound/soc/generic/simple-card.c 	of_node_put(node);
node              421 sound/soc/generic/simple-card.c static int simple_parse_aux_devs(struct device_node *node,
node              429 sound/soc/generic/simple-card.c 	if (!of_find_property(node, PREFIX "aux-devs", &len))
node              442 sound/soc/generic/simple-card.c 		aux_node = of_parse_phandle(node, PREFIX "aux-devs", i);
node              544 sound/soc/hisilicon/hi6210-i2s.c 	struct device_node *node = pdev->dev.of_node;
node              567 sound/soc/hisilicon/hi6210-i2s.c 	i2s->sysctrl = syscon_regmap_lookup_by_phandle(node,
node              295 sound/soc/intel/atom/sst-atom-controls.c 	list_for_each_entry(algo, &ids->algo_list, node) {
node              589 sound/soc/intel/atom/sst-atom-controls.c 	list_for_each_entry(gain, &ids->gain_list, node) {
node             1386 sound/soc/intel/atom/sst-atom-controls.c 		list_add_tail(&module->node, &ids->gain_list);
node             1392 sound/soc/intel/atom/sst-atom-controls.c 		list_add_tail(&module->node, &ids->algo_list);
node              548 sound/soc/intel/atom/sst-atom-controls.h 	struct list_head node;
node              484 sound/soc/intel/atom/sst-mfld-dsp.h 	struct list_head node;
node              106 sound/soc/intel/atom/sst/sst.c 		list_add_tail(&msg->node, &drv->rx_list);
node              126 sound/soc/intel/atom/sst/sst.c 	list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) {
node              127 sound/soc/intel/atom/sst/sst.c 		list_del(&msg->node);
node              166 sound/soc/intel/atom/sst/sst.h 	struct list_head node;
node               43 sound/soc/intel/atom/sst/sst_ipc.c 	list_add_tail(&msg->node, &ctx->block_list);
node               72 sound/soc/intel/atom/sst/sst_ipc.c 	list_for_each_entry(block, &ctx->block_list, node) {
node               99 sound/soc/intel/atom/sst/sst_ipc.c 	list_for_each_entry_safe(block, __block, &ctx->block_list, node) {
node              103 sound/soc/intel/atom/sst/sst_ipc.c 			list_del(&freed->node);
node              157 sound/soc/intel/atom/sst/sst_ipc.c 				struct ipc_post, node);
node              158 sound/soc/intel/atom/sst/sst_ipc.c 		list_del(&msg->node);
node              404 sound/soc/intel/atom/sst/sst_pvt.c 	list_add_tail(&msg->node, &sst->ipc_dispatch_list);
node              142 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	struct list_head node;
node              203 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	list_for_each_entry(stream, &byt->stream_list, node) {
node              351 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	list_add(&stream->node, &byt->stream_list);
node              455 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	list_del(&stream->node);
node              243 sound/soc/intel/haswell/sst-haswell-ipc.c 	struct list_head node;
node              416 sound/soc/intel/haswell/sst-haswell-ipc.c 	list_for_each_entry(stream, &hsw->stream_list, node) {
node              976 sound/soc/intel/haswell/sst-haswell-ipc.c 	list_add(&stream->node, &hsw->stream_list);
node             1024 sound/soc/intel/haswell/sst-haswell-ipc.c 	list_del(&stream->node);
node              507 sound/soc/intel/skylake/skl-messages.c 		node_id.node.dma_type =
node              511 sound/soc/intel/skylake/skl-messages.c 		node_id.node.vindex = params->host_dma_id +
node              516 sound/soc/intel/skylake/skl-messages.c 		node_id.node.dma_type =
node              522 sound/soc/intel/skylake/skl-messages.c 		node_id.node.vindex = ssp_node.val;
node              526 sound/soc/intel/skylake/skl-messages.c 		node_id.node.dma_type = SKL_DMA_DMIC_LINK_INPUT_CLASS;
node              527 sound/soc/intel/skylake/skl-messages.c 		node_id.node.vindex = mconfig->vbus_id +
node              532 sound/soc/intel/skylake/skl-messages.c 		node_id.node.dma_type =
node              536 sound/soc/intel/skylake/skl-messages.c 		node_id.node.vindex = params->link_dma_id;
node              540 sound/soc/intel/skylake/skl-messages.c 		node_id.node.dma_type =
node              544 sound/soc/intel/skylake/skl-messages.c 		node_id.node.vindex = params->host_dma_id;
node             1388 sound/soc/intel/skylake/skl-pcm.c 	list_for_each_entry(p, &skl->ppl_list, node) {
node             1389 sound/soc/intel/skylake/skl-pcm.c 		list_for_each_entry(m, &p->pipe->w_list, node) {
node             1525 sound/soc/intel/skylake/skl-pcm.c 		list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) {
node             1526 sound/soc/intel/skylake/skl-pcm.c 			list_del(&modules->node);
node              478 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &pipe->w_list, node) {
node              555 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &pipe->w_list, node) {
node              674 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
node              693 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
node              697 sound/soc/intel/skylake/skl-topology.c 		list_for_each_entry(modules, &skl->bind_list, node) {
node              897 sound/soc/intel/skylake/skl-topology.c 				list_for_each_entry(modules, &skl->bind_list, node) {
node              910 sound/soc/intel/skylake/skl-topology.c 			list_add(&m_list->node, &skl->bind_list);
node             1178 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
node             1184 sound/soc/intel/skylake/skl-topology.c 		list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) {
node             1199 sound/soc/intel/skylake/skl-topology.c 				list_del(&modules->node);
node             1207 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
node             1221 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
node             1956 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(ppl, &skl->ppl_list, node) {
node             1980 sound/soc/intel/skylake/skl-topology.c 	list_add(&ppl->node, &skl->ppl_list);
node             2610 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(ppl, &skl->ppl_list, node) {
node             2638 sound/soc/intel/skylake/skl-topology.c 	list_add(&ppl->node, &skl->ppl_list);
node             3526 sound/soc/intel/skylake/skl-topology.c 			list_add_tail(&p_module->node, &pipe->w_list);
node             3540 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(w_module, &pipe->w_list, node) {
node             3595 sound/soc/intel/skylake/skl-topology.c 	list_for_each_entry(ppl, &skl->ppl_list, node)
node             3609 sound/soc/intel/skylake/skl-topology.c 		list_for_each_entry_safe(ppl, tmp, &skl->ppl_list, node)
node             3610 sound/soc/intel/skylake/skl-topology.c 			list_del(&ppl->node);
node              189 sound/soc/intel/skylake/skl-topology.h 	} node;
node              261 sound/soc/intel/skylake/skl-topology.h 	struct list_head node;
node              419 sound/soc/intel/skylake/skl-topology.h 	struct list_head node;
node              425 sound/soc/intel/skylake/skl-topology.h 	struct list_head node;
node               71 sound/soc/meson/axg-card.c 			      struct device_node *node,
node               78 sound/soc/meson/axg-card.c 	if (!dai_name || !dai_of_node || !node)
node               81 sound/soc/meson/axg-card.c 	ret = of_parse_phandle_with_args(node, "sound-dai",
node               95 sound/soc/meson/axg-card.c 				  struct device_node *node,
node               99 sound/soc/meson/axg-card.c 				    prefix, node->full_name);
node              137 sound/soc/meson/axg-card.c 	struct device_node *node = card->dev->of_node;
node              141 sound/soc/meson/axg-card.c 	num = of_count_phandle_with_args(node, "audio-aux-devs", NULL);
node              163 sound/soc/meson/axg-card.c 			of_parse_phandle(node, "audio-aux-devs", i);
node              307 sound/soc/meson/axg-card.c static unsigned int axg_card_parse_daifmt(struct device_node *node,
node              314 sound/soc/meson/axg-card.c 	daifmt = snd_soc_of_parse_daifmt(node, PREFIX,
node              335 sound/soc/meson/axg-card.c 					struct device_node *node,
node              351 sound/soc/meson/axg-card.c 		snd_soc_of_get_slot_mask(node, propname, &be->tx_mask[i]);
node              361 sound/soc/meson/axg-card.c 		snd_soc_of_get_slot_mask(node, propname, &be->rx_mask[i]);
node              375 sound/soc/meson/axg-card.c 	of_property_read_u32(node, "dai-tdm-slot-num", &be->slots);
node              391 sound/soc/meson/axg-card.c 	of_property_read_u32(node, "dai-tdm-slot-width", &be->slot_width);
node              398 sound/soc/meson/axg-card.c 				       struct device_node *node,
node              411 sound/soc/meson/axg-card.c 	for_each_child_of_node(node, np) {
node              424 sound/soc/meson/axg-card.c 			      struct device_node *node,
node              441 sound/soc/meson/axg-card.c 	link->dai_fmt = axg_card_parse_daifmt(node, link->cpus->of_node);
node              443 sound/soc/meson/axg-card.c 	of_property_read_u32(node, "mclk-fs", &be->mclk_fs);
node              445 sound/soc/meson/axg-card.c 	ret = axg_card_parse_cpu_tdm_slots(card, link, node, be);
node              451 sound/soc/meson/axg-card.c 	ret = axg_card_parse_codecs_masks(card, link, node, be);
node              467 sound/soc/meson/axg-card.c 				struct device_node *node)
node              477 sound/soc/meson/axg-card.c 	num_codecs = of_get_child_count(node);
node              480 sound/soc/meson/axg-card.c 			node->full_name);
node              491 sound/soc/meson/axg-card.c 	for_each_child_of_node(node, np) {
node              502 sound/soc/meson/axg-card.c 	ret = axg_card_set_link_name(card, link, node, "be");
node              511 sound/soc/meson/axg-card.c 				struct device_node *node,
node              535 sound/soc/meson/axg-card.c 	return axg_card_set_link_name(card, link, node, "fe");
node              598 sound/soc/meson/axg-card.c 	struct device_node *node = card->dev->of_node;
node              602 sound/soc/meson/axg-card.c 	num = of_get_child_count(node);
node              613 sound/soc/meson/axg-card.c 	for_each_child_of_node(node, np) {
node              118 sound/soc/qcom/apq8016_sbc.c 	struct device_node *np, *codec, *cpu, *node  = dev->of_node;
node              130 sound/soc/qcom/apq8016_sbc.c 	if (of_property_read_bool(node, "qcom,audio-routing")) {
node              139 sound/soc/qcom/apq8016_sbc.c 	num_links = of_get_child_count(node);
node              153 sound/soc/qcom/apq8016_sbc.c 	for_each_child_of_node(node, np) {
node               52 sound/soc/qcom/qdsp6/q6adm.c 	struct list_head node;
node               98 sound/soc/qcom/qdsp6/q6adm.c 	list_for_each_entry(c, &adm->copps_list, node) {
node              120 sound/soc/qcom/qdsp6/q6adm.c 	list_del(&c->node);
node              308 sound/soc/qcom/qdsp6/q6adm.c 	list_for_each_entry(c, &adm->copps_list, node) {
node              411 sound/soc/qcom/qdsp6/q6adm.c 	list_add_tail(&copp->node, &adm->copps_list);
node              465 sound/soc/qcom/qdsp6/q6adm.c 	struct q6adm_session_map_node_v5 *node;
node              473 sound/soc/qcom/qdsp6/q6adm.c 	pkt_size = (APR_HDR_SIZE + sizeof(*route) +  sizeof(*node) +
node              482 sound/soc/qcom/qdsp6/q6adm.c 	node = matrix_map + APR_HDR_SIZE + sizeof(*route);
node              483 sound/soc/qcom/qdsp6/q6adm.c 	copps_list = matrix_map + APR_HDR_SIZE + sizeof(*route) + sizeof(*node);
node              505 sound/soc/qcom/qdsp6/q6adm.c 	node->session_id = payload_map.session_id;
node              506 sound/soc/qcom/qdsp6/q6adm.c 	node->num_copps = payload_map.num_copps;
node             1368 sound/soc/qcom/qdsp6/q6afe-dai.c 	struct device_node *node;
node             1371 sound/soc/qcom/qdsp6/q6afe-dai.c 	for_each_child_of_node(dev->of_node, node) {
node             1376 sound/soc/qcom/qdsp6/q6afe-dai.c 		ret = of_property_read_u32(node, "reg", &id);
node             1386 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_variable_u32_array(node,
node             1403 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-sync-mode",
node             1409 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-sync-src",
node             1415 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-data-out",
node             1421 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-invert-sync",
node             1427 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-data-delay",
node             1433 sound/soc/qcom/qdsp6/q6afe-dai.c 			ret = of_property_read_u32(node, "qcom,tdm-data-align",
node              486 sound/soc/qcom/qdsp6/q6afe.c 	struct list_head node;
node              721 sound/soc/qcom/qdsp6/q6afe.c 	list_del(&port->node);
node              734 sound/soc/qcom/qdsp6/q6afe.c 	list_for_each_entry(p, &afe->port_list, node)
node             1442 sound/soc/qcom/qdsp6/q6afe.c 	list_add_tail(&port->node, &afe->port_list);
node              877 sound/soc/qcom/qdsp6/q6asm-dai.c 	struct device_node *node;
node              882 sound/soc/qcom/qdsp6/q6asm-dai.c 	for_each_child_of_node(dev->of_node, node) {
node              883 sound/soc/qcom/qdsp6/q6asm-dai.c 		ret = of_property_read_u32(node, "reg", &id);
node              891 sound/soc/qcom/qdsp6/q6asm-dai.c 		ret = of_property_read_u32(node, "direction", &dir);
node              900 sound/soc/qcom/qdsp6/q6asm-dai.c 		if (of_property_read_bool(node, "is-compress-dai"))
node              910 sound/soc/qcom/qdsp6/q6asm-dai.c 	struct device_node *node = dev->of_node;
node              919 sound/soc/qcom/qdsp6/q6asm-dai.c 	rc = of_parse_phandle_with_fixed_args(node, "iommus", 1, 0, &args);
node              579 sound/soc/rockchip/rockchip_i2s.c 	struct device_node *node = pdev->dev.of_node;
node              594 sound/soc/rockchip/rockchip_i2s.c 	i2s->grf = syscon_regmap_lookup_by_phandle(node, "rockchip,grf");
node              658 sound/soc/rockchip/rockchip_i2s.c 	if (!of_property_read_u32(node, "rockchip,playback-channels", &val)) {
node              663 sound/soc/rockchip/rockchip_i2s.c 	if (!of_property_read_u32(node, "rockchip,capture-channels", &val)) {
node             1079 sound/soc/sh/rcar/core.c 		struct device_node *node = is_play ?
node             1083 sound/soc/sh/rcar/core.c 		if (!node)
node             1088 sound/soc/sh/rcar/core.c 			if (np == node) {
node             1135 sound/soc/sh/rcar/core.c 		struct device_node *node,
node             1144 sound/soc/sh/rcar/core.c 	if (!node)
node             1148 sound/soc/sh/rcar/core.c 	for_each_child_of_node(node, np) {
node             1157 sound/soc/sh/rcar/core.c 	of_node_put(node);
node              298 sound/soc/sh/rcar/ctu.c 	struct device_node *node;
node              310 sound/soc/sh/rcar/ctu.c 	node = rsnd_ctu_of_node(priv);
node              311 sound/soc/sh/rcar/ctu.c 	if (!node)
node              314 sound/soc/sh/rcar/ctu.c 	nr = of_get_child_count(node);
node              331 sound/soc/sh/rcar/ctu.c 	for_each_child_of_node(node, np) {
node              360 sound/soc/sh/rcar/ctu.c 	of_node_put(node);
node              308 sound/soc/sh/rcar/dvc.c 	struct device_node *node;
node              320 sound/soc/sh/rcar/dvc.c 	node = rsnd_dvc_of_node(priv);
node              321 sound/soc/sh/rcar/dvc.c 	if (!node)
node              324 sound/soc/sh/rcar/dvc.c 	nr = of_get_child_count(node);
node              341 sound/soc/sh/rcar/dvc.c 	for_each_child_of_node(node, np) {
node              365 sound/soc/sh/rcar/dvc.c 	of_node_put(node);
node              272 sound/soc/sh/rcar/mix.c 	struct device_node *node;
node              284 sound/soc/sh/rcar/mix.c 	node = rsnd_mix_of_node(priv);
node              285 sound/soc/sh/rcar/mix.c 	if (!node)
node              288 sound/soc/sh/rcar/mix.c 	nr = of_get_child_count(node);
node              305 sound/soc/sh/rcar/mix.c 	for_each_child_of_node(node, np) {
node              329 sound/soc/sh/rcar/mix.c 	of_node_put(node);
node              451 sound/soc/sh/rcar/rsnd.h 		struct device_node *node,
node              475 sound/soc/sh/rcar/rsnd.h #define rsnd_parse_of_node(priv, node)					\
node              476 sound/soc/sh/rcar/rsnd.h 	of_get_child_by_name(rsnd_priv_to_dev(priv)->of_node, node)
node              614 sound/soc/sh/rcar/src.c 	struct device_node *node;
node              626 sound/soc/sh/rcar/src.c 	node = rsnd_src_of_node(priv);
node              627 sound/soc/sh/rcar/src.c 	if (!node)
node              630 sound/soc/sh/rcar/src.c 	nr = of_get_child_count(node);
node              646 sound/soc/sh/rcar/src.c 	for_each_child_of_node(node, np) {
node              683 sound/soc/sh/rcar/src.c 	of_node_put(node);
node             1066 sound/soc/sh/rcar/ssi.c 	struct device_node *node;
node             1071 sound/soc/sh/rcar/ssi.c 	node = rsnd_ssi_of_node(priv);
node             1072 sound/soc/sh/rcar/ssi.c 	if (!node)
node             1076 sound/soc/sh/rcar/ssi.c 	for_each_child_of_node(node, np) {
node             1085 sound/soc/sh/rcar/ssi.c 	of_node_put(node);
node             1106 sound/soc/sh/rcar/ssi.c 	struct device_node *node;
node             1115 sound/soc/sh/rcar/ssi.c 	node = rsnd_ssi_of_node(priv);
node             1116 sound/soc/sh/rcar/ssi.c 	if (!node)
node             1119 sound/soc/sh/rcar/ssi.c 	nr = of_get_child_count(node);
node             1135 sound/soc/sh/rcar/ssi.c 	for_each_child_of_node(node, np) {
node             1182 sound/soc/sh/rcar/ssi.c 	of_node_put(node);
node              361 sound/soc/sh/rcar/ssiu.c 	struct device_node *node = rsnd_ssiu_of_node(priv);
node              369 sound/soc/sh/rcar/ssiu.c 	if (node) {
node              371 sound/soc/sh/rcar/ssiu.c 		for_each_child_of_node(node, np) {
node              380 sound/soc/sh/rcar/ssiu.c 		of_node_put(node);
node              393 sound/soc/sh/rcar/ssiu.c 	struct device_node *node;
node              406 sound/soc/sh/rcar/ssiu.c 	node = rsnd_ssiu_of_node(priv);
node              407 sound/soc/sh/rcar/ssiu.c 	if (node)
node              408 sound/soc/sh/rcar/ssiu.c 		nr = of_get_child_count(node);
node              426 sound/soc/sh/rcar/ssiu.c 	if ((node) &&
node              444 sound/soc/sh/rcar/ssiu.c 		if (node) {
node              208 sound/soc/soc-dapm.c 	struct snd_soc_dapm_widget *node;
node              224 sound/soc/soc-dapm.c 			node = p->node[rdir];
node              225 sound/soc/soc-dapm.c 			if (node->endpoints[dir] != -1) {
node              226 sound/soc/soc-dapm.c 				node->endpoints[dir] = -1;
node              227 sound/soc/soc-dapm.c 				list_add_tail(&node->work_list, &list);
node             1173 sound/soc/soc-dapm.c 			invalidate_paths_ep(path->node[dir], dir);
node             1230 sound/soc/soc-dapm.c 			con += fn(path->node[dir], list, custom_stop_condition);
node             2151 sound/soc/soc-dapm.c 					p->node[rdir]->name);
node             2832 sound/soc/soc-dapm.c 	path->node[SND_SOC_DAPM_DIR_IN] = wsource;
node             2833 sound/soc/soc-dapm.c 	path->node[SND_SOC_DAPM_DIR_OUT] = wsink;
node              381 sound/soc/sti/sti_uniperif.c static int sti_uniperiph_cpu_dai_of(struct device_node *node,
node              395 sound/soc/sti/sti_uniperif.c 	of_id = of_match_node(snd_soc_sti_match, node);
node              436 sound/soc/sti/sti_uniperif.c 		if (!of_property_read_string(node, "st,tdm-mode", &mode))
node              473 sound/soc/sti/sti_uniperif.c 	struct device_node *node = pdev->dev.of_node;
node              486 sound/soc/sti/sti_uniperif.c 	ret = sti_uniperiph_cpu_dai_of(node, priv);
node             1013 sound/soc/sti/uniperif_player.c 	struct device_node *node = pdev->dev.of_node;
node             1024 sound/soc/sti/uniperif_player.c 	regmap = syscon_regmap_lookup_by_phandle(node, "st,syscfg");
node             1686 sound/soc/ti/davinci-mcasp.c 	struct device_node *node = pdev->dev.of_node;
node             1691 sound/soc/ti/davinci-mcasp.c 	if (!node)
node             1694 sound/soc/ti/davinci-mcasp.c 	parent_name = of_get_property(node, "fck_parent", NULL);
node              213 sound/soc/ti/omap-abe-twl6040.c 	struct device_node *node = pdev->dev.of_node;
node              220 sound/soc/ti/omap-abe-twl6040.c 	if (!node) {
node              248 sound/soc/ti/omap-abe-twl6040.c 	dai_node = of_parse_phandle(node, "ti,mcpdm", 0);
node              267 sound/soc/ti/omap-abe-twl6040.c 	dai_node = of_parse_phandle(node, "ti,dmic", 0);
node              286 sound/soc/ti/omap-abe-twl6040.c 	priv->jack_detection = of_property_read_bool(node, "ti,jack-detection");
node              287 sound/soc/ti/omap-abe-twl6040.c 	of_property_read_u32(node, "ti,mclk-freq", &priv->mclk_freq);
node             1384 sound/soc/ti/omap-mcbsp.c 		struct device_node *node = pdev->dev.of_node;
node             1395 sound/soc/ti/omap-mcbsp.c 		if (!of_property_read_u32(node, "ti,buffer-size", &buffer_size))
node              240 sound/soc/ti/omap-twl4030.c 	struct device_node *node = pdev->dev.of_node;
node              251 sound/soc/ti/omap-twl4030.c 	if (node) {
node              260 sound/soc/ti/omap-twl4030.c 		dai_node = of_parse_phandle(node, "ti,mcbsp", 0);
node              271 sound/soc/ti/omap-twl4030.c 		dai_node = of_parse_phandle(node, "ti,mcbsp-voice", 0);
node              282 sound/soc/ti/omap-twl4030.c 		priv->jack_detect = of_get_named_gpio(node,
node              286 sound/soc/ti/omap-twl4030.c 		prop = of_find_property(node, "ti,audio-routing", NULL);
node              103 sound/soc/xilinx/xlnx_i2s.c 	struct device_node *node = dev->of_node;
node              113 sound/soc/xilinx/xlnx_i2s.c 	ret = of_property_read_u32(node, "xlnx,num-channels", &ch);
node              120 sound/soc/xilinx/xlnx_i2s.c 	ret = of_property_read_u32(node, "xlnx,dwidth", &data_width);
node              136 sound/soc/xilinx/xlnx_i2s.c 	if (of_device_is_compatible(node, "xlnx,i2s-transmitter-1.0")) {
node              144 sound/soc/xilinx/xlnx_i2s.c 	} else if (of_device_is_compatible(node, "xlnx,i2s-receiver-1.0")) {
node              245 sound/soc/xilinx/xlnx_spdif.c 	struct device_node *node = dev->of_node;
node              268 sound/soc/xilinx/xlnx_spdif.c 	ret = of_property_read_u32(node, "xlnx,spdif-mode", &ctx->mode);
node              295 sound/soc/xilinx/xlnx_spdif.c 	ret = of_property_read_u32(node, "xlnx,aud_clk_i", &ctx->aclk);
node              393 sound/xen/xen_snd_front_cfg.c 	char node[3];
node              417 sound/xen/xen_snd_front_cfg.c 		snprintf(node, sizeof(node), "%d", num_streams);
node              418 sound/xen/xen_snd_front_cfg.c 		if (!xenbus_exists(XBT_NIL, device_path, node))
node              481 sound/xen/xen_snd_front_cfg.c 	char node[3];
node              486 sound/xen/xen_snd_front_cfg.c 		snprintf(node, sizeof(node), "%d", num_devices);
node              487 sound/xen/xen_snd_front_cfg.c 		if (!xenbus_exists(XBT_NIL, xb_dev->nodename, node))
node               57 tools/include/linux/hashtable.h #define hash_add(hashtable, node, key)						\
node               58 tools/include/linux/hashtable.h 	hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))])
node               64 tools/include/linux/hashtable.h static inline bool hash_hashed(struct hlist_node *node)
node               66 tools/include/linux/hashtable.h 	return !hlist_unhashed(node);
node               93 tools/include/linux/hashtable.h static inline void hash_del(struct hlist_node *node)
node               95 tools/include/linux/hashtable.h 	hlist_del_init(node);
node               42 tools/include/linux/rbtree.h #define RB_EMPTY_NODE(node)  \
node               43 tools/include/linux/rbtree.h 	((node)->__rb_parent_color == (unsigned long)(node))
node               44 tools/include/linux/rbtree.h #define RB_CLEAR_NODE(node)  \
node               45 tools/include/linux/rbtree.h 	((node)->__rb_parent_color = (unsigned long)(node))
node               66 tools/include/linux/rbtree.h static inline void rb_link_node(struct rb_node *node, struct rb_node *parent,
node               69 tools/include/linux/rbtree.h 	node->__rb_parent_color = (unsigned long)parent;
node               70 tools/include/linux/rbtree.h 	node->rb_left = node->rb_right = NULL;
node               72 tools/include/linux/rbtree.h 	*rb_link = node;
node              129 tools/include/linux/rbtree.h static inline void rb_insert_color_cached(struct rb_node *node,
node              134 tools/include/linux/rbtree.h 		root->rb_leftmost = node;
node              135 tools/include/linux/rbtree.h 	rb_insert_color(node, &root->rb_root);
node              138 tools/include/linux/rbtree.h static inline void rb_erase_cached(struct rb_node *node,
node              141 tools/include/linux/rbtree.h 	if (root->rb_leftmost == node)
node              142 tools/include/linux/rbtree.h 		root->rb_leftmost = rb_next(node);
node              143 tools/include/linux/rbtree.h 	rb_erase(node, &root->rb_root);
node               30 tools/include/linux/rbtree_augmented.h 	void (*propagate)(struct rb_node *node, struct rb_node *stop);
node               35 tools/include/linux/rbtree_augmented.h extern void __rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node               49 tools/include/linux/rbtree_augmented.h rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node               52 tools/include/linux/rbtree_augmented.h 	__rb_insert_augmented(node, root, augment->rotate);
node               56 tools/include/linux/rbtree_augmented.h rb_insert_augmented_cached(struct rb_node *node,
node               61 tools/include/linux/rbtree_augmented.h 		root->rb_leftmost = node;
node               62 tools/include/linux/rbtree_augmented.h 	rb_insert_augmented(node, &root->rb_root, augment);
node               82 tools/include/linux/rbtree_augmented.h 		RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD);	\
node               83 tools/include/linux/rbtree_augmented.h 		if (RBCOMPUTE(node, true))				\
node               85 tools/include/linux/rbtree_augmented.h 		rb = rb_parent(&node->RBFIELD);				\
node              124 tools/include/linux/rbtree_augmented.h static inline bool RBNAME ## _compute_max(RBSTRUCT *node, bool exit)	      \
node              127 tools/include/linux/rbtree_augmented.h 	RBTYPE max = RBCOMPUTE(node);					      \
node              128 tools/include/linux/rbtree_augmented.h 	if (node->RBFIELD.rb_left) {					      \
node              129 tools/include/linux/rbtree_augmented.h 		child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD);   \
node              133 tools/include/linux/rbtree_augmented.h 	if (node->RBFIELD.rb_right) {					      \
node              134 tools/include/linux/rbtree_augmented.h 		child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD);  \
node              138 tools/include/linux/rbtree_augmented.h 	if (exit && node->RBAUGMENTED == max)				      \
node              140 tools/include/linux/rbtree_augmented.h 	node->RBAUGMENTED = max;					      \
node              187 tools/include/linux/rbtree_augmented.h __rb_erase_augmented(struct rb_node *node, struct rb_root *root,
node              190 tools/include/linux/rbtree_augmented.h 	struct rb_node *child = node->rb_right;
node              191 tools/include/linux/rbtree_augmented.h 	struct rb_node *tmp = node->rb_left;
node              203 tools/include/linux/rbtree_augmented.h 		pc = node->__rb_parent_color;
node              205 tools/include/linux/rbtree_augmented.h 		__rb_change_child(node, child, parent, root);
node              214 tools/include/linux/rbtree_augmented.h 		tmp->__rb_parent_color = pc = node->__rb_parent_color;
node              216 tools/include/linux/rbtree_augmented.h 		__rb_change_child(node, tmp, parent, root);
node              236 tools/include/linux/rbtree_augmented.h 			augment->copy(node, successor);
node              262 tools/include/linux/rbtree_augmented.h 			augment->copy(node, successor);
node              266 tools/include/linux/rbtree_augmented.h 		tmp = node->rb_left;
node              270 tools/include/linux/rbtree_augmented.h 		pc = node->__rb_parent_color;
node              272 tools/include/linux/rbtree_augmented.h 		__rb_change_child(node, successor, tmp, root);
node              291 tools/include/linux/rbtree_augmented.h rb_erase_augmented(struct rb_node *node, struct rb_root *root,
node              294 tools/include/linux/rbtree_augmented.h 	struct rb_node *rebalance = __rb_erase_augmented(node, root, augment);
node              300 tools/include/linux/rbtree_augmented.h rb_erase_augmented_cached(struct rb_node *node, struct rb_root_cached *root,
node              303 tools/include/linux/rbtree_augmented.h 	if (root->rb_leftmost == node)
node              304 tools/include/linux/rbtree_augmented.h 		root->rb_leftmost = rb_next(node);
node              305 tools/include/linux/rbtree_augmented.h 	rb_erase_augmented(node, &root->rb_root, augment);
node              105 tools/lib/bpf/bpf.c 			__u32 map_flags, int node)
node              115 tools/lib/bpf/bpf.c 	if (node >= 0) {
node              116 tools/lib/bpf/bpf.c 		map_attr.numa_node = node;
node              155 tools/lib/bpf/bpf.c 			       __u32 map_flags, int node)
node              171 tools/lib/bpf/bpf.c 	if (node >= 0) {
node              173 tools/lib/bpf/bpf.c 		attr.numa_node = node;
node               58 tools/lib/bpf/bpf.h 				   int max_entries, __u32 map_flags, int node);
node               67 tools/lib/bpf/bpf.h 					  __u32 map_flags, int node);
node             1318 tools/lib/bpf/btf.c #define for_each_dedup_cand(d, node, hash) \
node             1319 tools/lib/bpf/btf.c 	hashmap__for_each_key_entry(d->dedup_table, node, (void *)hash)
node               33 tools/lib/lockdep/preload.c 	struct rb_node node;
node              103 tools/lib/lockdep/preload.c 	struct rb_node **node = &locks.rb_node;
node              108 tools/lib/lockdep/preload.c 	while (*node) {
node              109 tools/lib/lockdep/preload.c 		l = rb_entry(*node, struct lock_lookup, node);
node              111 tools/lib/lockdep/preload.c 		*parent = *node;
node              113 tools/lib/lockdep/preload.c 			node = &l->node.rb_left;
node              115 tools/lib/lockdep/preload.c 			node = &l->node.rb_right;
node              117 tools/lib/lockdep/preload.c 			return node;
node              120 tools/lib/lockdep/preload.c 	return node;
node              177 tools/lib/lockdep/preload.c 	struct rb_node **node, *parent;
node              181 tools/lib/lockdep/preload.c 	node = __get_lock_node(lock, &parent);
node              183 tools/lib/lockdep/preload.c 	if (*node) {
node              184 tools/lib/lockdep/preload.c 		return rb_entry(*node, struct lock_lookup, node);
node              204 tools/lib/lockdep/preload.c 	node = __get_lock_node(lock, &parent);
node              205 tools/lib/lockdep/preload.c 	rb_link_node(&l->node, parent, node);
node              206 tools/lib/lockdep/preload.c 	rb_insert_color(&l->node, &locks);
node              215 tools/lib/lockdep/preload.c 	rb_erase(&lock->node, &locks);
node               85 tools/lib/rbtree.c __rb_insert(struct rb_node *node, struct rb_root *root,
node               88 tools/lib/rbtree.c 	struct rb_node *parent = rb_red_parent(node), *gparent, *tmp;
node              100 tools/lib/rbtree.c 			rb_set_parent_color(node, NULL, RB_BLACK);
node              133 tools/lib/rbtree.c 				node = gparent;
node              134 tools/lib/rbtree.c 				parent = rb_parent(node);
node              135 tools/lib/rbtree.c 				rb_set_parent_color(node, parent, RB_RED);
node              140 tools/lib/rbtree.c 			if (node == tmp) {
node              154 tools/lib/rbtree.c 				tmp = node->rb_left;
node              156 tools/lib/rbtree.c 				WRITE_ONCE(node->rb_left, parent);
node              160 tools/lib/rbtree.c 				rb_set_parent_color(parent, node, RB_RED);
node              161 tools/lib/rbtree.c 				augment_rotate(parent, node);
node              162 tools/lib/rbtree.c 				parent = node;
node              163 tools/lib/rbtree.c 				tmp = node->rb_right;
node              189 tools/lib/rbtree.c 				node = gparent;
node              190 tools/lib/rbtree.c 				parent = rb_parent(node);
node              191 tools/lib/rbtree.c 				rb_set_parent_color(node, parent, RB_RED);
node              196 tools/lib/rbtree.c 			if (node == tmp) {
node              198 tools/lib/rbtree.c 				tmp = node->rb_right;
node              200 tools/lib/rbtree.c 				WRITE_ONCE(node->rb_right, parent);
node              204 tools/lib/rbtree.c 				rb_set_parent_color(parent, node, RB_RED);
node              205 tools/lib/rbtree.c 				augment_rotate(parent, node);
node              206 tools/lib/rbtree.c 				parent = node;
node              207 tools/lib/rbtree.c 				tmp = node->rb_left;
node              230 tools/lib/rbtree.c 	struct rb_node *node = NULL, *sibling, *tmp1, *tmp2;
node              241 tools/lib/rbtree.c 		if (node != sibling) {	/* node == parent->rb_left */
node              285 tools/lib/rbtree.c 						node = parent;
node              286 tools/lib/rbtree.c 						parent = rb_parent(node);
node              375 tools/lib/rbtree.c 						node = parent;
node              376 tools/lib/rbtree.c 						parent = rb_parent(node);
node              423 tools/lib/rbtree.c static inline void dummy_propagate(struct rb_node *node, struct rb_node *stop) {}
node              433 tools/lib/rbtree.c void rb_insert_color(struct rb_node *node, struct rb_root *root)
node              435 tools/lib/rbtree.c 	__rb_insert(node, root, dummy_rotate);
node              438 tools/lib/rbtree.c void rb_erase(struct rb_node *node, struct rb_root *root)
node              441 tools/lib/rbtree.c 	rebalance = __rb_erase_augmented(node, root, &dummy_callbacks);
node              453 tools/lib/rbtree.c void __rb_insert_augmented(struct rb_node *node, struct rb_root *root,
node              456 tools/lib/rbtree.c 	__rb_insert(node, root, augment_rotate);
node              486 tools/lib/rbtree.c struct rb_node *rb_next(const struct rb_node *node)
node              490 tools/lib/rbtree.c 	if (RB_EMPTY_NODE(node))
node              497 tools/lib/rbtree.c 	if (node->rb_right) {
node              498 tools/lib/rbtree.c 		node = node->rb_right;
node              499 tools/lib/rbtree.c 		while (node->rb_left)
node              500 tools/lib/rbtree.c 			node=node->rb_left;
node              501 tools/lib/rbtree.c 		return (struct rb_node *)node;
node              511 tools/lib/rbtree.c 	while ((parent = rb_parent(node)) && node == parent->rb_right)
node              512 tools/lib/rbtree.c 		node = parent;
node              517 tools/lib/rbtree.c struct rb_node *rb_prev(const struct rb_node *node)
node              521 tools/lib/rbtree.c 	if (RB_EMPTY_NODE(node))
node              528 tools/lib/rbtree.c 	if (node->rb_left) {
node              529 tools/lib/rbtree.c 		node = node->rb_left;
node              530 tools/lib/rbtree.c 		while (node->rb_right)
node              531 tools/lib/rbtree.c 			node=node->rb_right;
node              532 tools/lib/rbtree.c 		return (struct rb_node *)node;
node              539 tools/lib/rbtree.c 	while ((parent = rb_parent(node)) && node == parent->rb_left)
node              540 tools/lib/rbtree.c 		node = parent;
node              561 tools/lib/rbtree.c static struct rb_node *rb_left_deepest_node(const struct rb_node *node)
node              564 tools/lib/rbtree.c 		if (node->rb_left)
node              565 tools/lib/rbtree.c 			node = node->rb_left;
node              566 tools/lib/rbtree.c 		else if (node->rb_right)
node              567 tools/lib/rbtree.c 			node = node->rb_right;
node              569 tools/lib/rbtree.c 			return (struct rb_node *)node;
node              573 tools/lib/rbtree.c struct rb_node *rb_next_postorder(const struct rb_node *node)
node              576 tools/lib/rbtree.c 	if (!node)
node              578 tools/lib/rbtree.c 	parent = rb_parent(node);
node              581 tools/lib/rbtree.c 	if (parent && node == parent->rb_left && parent->rb_right) {
node              240 tools/perf/bench/numa.c static int is_node_present(int node)
node              242 tools/perf/bench/numa.c 	return numa_bitmask_isbitset(numa_nodes_ptr, node);
node              248 tools/perf/bench/numa.c static bool node_has_cpus(int node)
node              253 tools/perf/bench/numa.c 	if (cpu && !numa_node_to_cpus(node, cpu)) {
node              340 tools/perf/bench/numa.c static void bind_to_memnode(int node)
node              345 tools/perf/bench/numa.c 	if (node == NUMA_NO_NODE)
node              349 tools/perf/bench/numa.c 	nodemask = 1L << node;
node              352 tools/perf/bench/numa.c 	dprintf("binding to node %d, mask: %016lx => %d\n", node, nodemask, ret);
node              380 tools/perf/bench/numa.c 		int node = numa_node_of_cpu(0);
node              382 tools/perf/bench/numa.c 		orig_mask = bind_to_node(node);
node              383 tools/perf/bench/numa.c 		bind_to_memnode(node);
node              884 tools/perf/bench/numa.c 		int node;
node              889 tools/perf/bench/numa.c 		node = numa_node_of_cpu(td->curr_cpu);
node              890 tools/perf/bench/numa.c 		if (node < 0) /* curr_cpu was likely still -1 */
node              893 tools/perf/bench/numa.c 		node_present[node] = 1;
node              911 tools/perf/bench/numa.c static int count_node_processes(int node)
node              926 tools/perf/bench/numa.c 			if (n == node) {
node              976 tools/perf/bench/numa.c 	int node;
node              983 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++)
node              984 tools/perf/bench/numa.c 		nodes[node] = 0;
node              999 tools/perf/bench/numa.c 		node = numa_node_of_cpu(cpu);
node             1001 tools/perf/bench/numa.c 		nodes[node]++;
node             1012 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++) {
node             1013 tools/perf/bench/numa.c 		if (!is_node_present(node))
node             1015 tools/perf/bench/numa.c 		nr = nodes[node];
node             1034 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++) {
node             1037 tools/perf/bench/numa.c 		if (!is_node_present(node))
node             1039 tools/perf/bench/numa.c 		processes = count_node_processes(node);
node             1040 tools/perf/bench/numa.c 		nr = nodes[node];
node              223 tools/perf/builtin-c2c.c 	int node;
node              230 tools/perf/builtin-c2c.c 	node = mem2node__node(&c2c.mem2node, sample->phys_addr);
node              231 tools/perf/builtin-c2c.c 	if (WARN_ONCE(node < 0, "WARNING: failed to find node\n"))
node              234 tools/perf/builtin-c2c.c 	set_bit(node, c2c_he->nodeset);
node              320 tools/perf/builtin-c2c.c 		int node = c2c.cpu2node[cpu];
node              337 tools/perf/builtin-c2c.c 		c2c_add_stats(&c2c_he->node_stats[node], &stats);
node             1088 tools/perf/builtin-c2c.c 	int node;
node             1093 tools/perf/builtin-c2c.c 	for (node = 0; node < c2c.nodes_cnt; node++) {
node             1097 tools/perf/builtin-c2c.c 		bitmap_and(set, c2c_he->cpuset, c2c.nodes[node], c2c.cpus_cnt);
node             1114 tools/perf/builtin-c2c.c 			ret = scnprintf(hpp->buf, hpp->size, "%2d", node);
node             1120 tools/perf/builtin-c2c.c 			struct c2c_stats *stats = &c2c_he->node_stats[node];
node             1122 tools/perf/builtin-c2c.c 			ret = scnprintf(hpp->buf, hpp->size, "%2d{%2d ", node, num);
node             1161 tools/perf/builtin-c2c.c 			ret = scnprintf(hpp->buf, hpp->size, "%2d{", node);
node             2033 tools/perf/builtin-c2c.c 	int node, cpu;
node             2061 tools/perf/builtin-c2c.c 	for (node = 0; node < c2c.nodes_cnt; node++) {
node             2062 tools/perf/builtin-c2c.c 		struct perf_cpu_map *map = n[node].map;
node             2069 tools/perf/builtin-c2c.c 		nodes[node] = set;
node             2081 tools/perf/builtin-c2c.c 			cpu2node[map->map[cpu]] = node;
node              495 tools/perf/builtin-diff.c 		list_for_each_entry(pair, &he->pairs.head, pairs.node)
node               51 tools/perf/builtin-inject.c 	struct list_head node;
node              476 tools/perf/builtin-inject.c 	list_for_each_entry(ent, &inject->samples, node) {
node              478 tools/perf/builtin-inject.c 			list_del_init(&ent->node);
node              507 tools/perf/builtin-inject.c 	list_add(&ent->node, &inject->samples);
node              523 tools/perf/builtin-inject.c 	list_for_each_entry(ent, &inject->samples, node) {
node               70 tools/perf/builtin-kmem.c 	struct rb_node node;
node               88 tools/perf/builtin-kmem.c 	struct rb_node **node = &root_alloc_stat.rb_node;
node               92 tools/perf/builtin-kmem.c 	while (*node) {
node               93 tools/perf/builtin-kmem.c 		parent = *node;
node               94 tools/perf/builtin-kmem.c 		data = rb_entry(*node, struct alloc_stat, node);
node               97 tools/perf/builtin-kmem.c 			node = &(*node)->rb_right;
node               99 tools/perf/builtin-kmem.c 			node = &(*node)->rb_left;
node              120 tools/perf/builtin-kmem.c 		rb_link_node(&data->node, parent, node);
node              121 tools/perf/builtin-kmem.c 		rb_insert_color(&data->node, &root_alloc_stat);
node              133 tools/perf/builtin-kmem.c 	struct rb_node **node = &root_caller_stat.rb_node;
node              137 tools/perf/builtin-kmem.c 	while (*node) {
node              138 tools/perf/builtin-kmem.c 		parent = *node;
node              139 tools/perf/builtin-kmem.c 		data = rb_entry(*node, struct alloc_stat, node);
node              142 tools/perf/builtin-kmem.c 			node = &(*node)->rb_right;
node              144 tools/perf/builtin-kmem.c 			node = &(*node)->rb_left;
node              165 tools/perf/builtin-kmem.c 		rb_link_node(&data->node, parent, node);
node              166 tools/perf/builtin-kmem.c 		rb_insert_color(&data->node, &root_caller_stat);
node              215 tools/perf/builtin-kmem.c 	struct rb_node *node = root->rb_node;
node              218 tools/perf/builtin-kmem.c 	while (node) {
node              222 tools/perf/builtin-kmem.c 		data = rb_entry(node, struct alloc_stat, node);
node              226 tools/perf/builtin-kmem.c 			node = node->rb_left;
node              228 tools/perf/builtin-kmem.c 			node = node->rb_right;
node              281 tools/perf/builtin-kmem.c 	struct rb_node 	node;
node              338 tools/perf/builtin-kmem.c 	struct rb_node *node;
node              359 tools/perf/builtin-kmem.c 	map__for_each_symbol(kernel_map, sym, node) {
node              391 tools/perf/builtin-kmem.c 	struct callchain_cursor_node *node;
node              406 tools/perf/builtin-kmem.c 		node = callchain_cursor_current(&callchain_cursor);
node              407 tools/perf/builtin-kmem.c 		if (node == NULL)
node              410 tools/perf/builtin-kmem.c 		key.start = key.end = node->ip;
node              415 tools/perf/builtin-kmem.c 			if (node->map)
node              416 tools/perf/builtin-kmem.c 				addr = map__unmap_ip(node->map, node->ip);
node              418 tools/perf/builtin-kmem.c 				addr = node->ip;
node              444 tools/perf/builtin-kmem.c 	struct rb_node **node = &page_live_tree.rb_node;
node              448 tools/perf/builtin-kmem.c 	while (*node) {
node              451 tools/perf/builtin-kmem.c 		parent = *node;
node              452 tools/perf/builtin-kmem.c 		data = rb_entry(*node, struct page_stat, node);
node              456 tools/perf/builtin-kmem.c 			node = &parent->rb_left;
node              458 tools/perf/builtin-kmem.c 			node = &parent->rb_right;
node              473 tools/perf/builtin-kmem.c 		rb_link_node(&data->node, parent, node);
node              474 tools/perf/builtin-kmem.c 		rb_insert_color(&data->node, &page_live_tree);
node              493 tools/perf/builtin-kmem.c 	struct rb_node **node = &page_alloc_tree.rb_node;
node              498 tools/perf/builtin-kmem.c 	while (*node) {
node              501 tools/perf/builtin-kmem.c 		parent = *node;
node              502 tools/perf/builtin-kmem.c 		data = rb_entry(*node, struct page_stat, node);
node              511 tools/perf/builtin-kmem.c 			node = &parent->rb_left;
node              513 tools/perf/builtin-kmem.c 			node = &parent->rb_right;
node              528 tools/perf/builtin-kmem.c 		rb_link_node(&data->node, parent, node);
node              529 tools/perf/builtin-kmem.c 		rb_insert_color(&data->node, &page_alloc_tree);
node              548 tools/perf/builtin-kmem.c 	struct rb_node **node = &page_caller_tree.rb_node;
node              553 tools/perf/builtin-kmem.c 	while (*node) {
node              556 tools/perf/builtin-kmem.c 		parent = *node;
node              557 tools/perf/builtin-kmem.c 		data = rb_entry(*node, struct page_stat, node);
node              566 tools/perf/builtin-kmem.c 			node = &parent->rb_left;
node              568 tools/perf/builtin-kmem.c 			node = &parent->rb_right;
node              583 tools/perf/builtin-kmem.c 		rb_link_node(&data->node, parent, node);
node              584 tools/perf/builtin-kmem.c 		rb_insert_color(&data->node, &page_caller_tree);
node              895 tools/perf/builtin-kmem.c 	rb_erase(&pstat->node, &page_live_tree);
node              921 tools/perf/builtin-kmem.c 			rb_erase(&pstat->node, &page_caller_tree);
node             1005 tools/perf/builtin-kmem.c 						   node);
node             1077 tools/perf/builtin-kmem.c 		data = rb_entry(next, struct page_stat, node);
node             1119 tools/perf/builtin-kmem.c 		data = rb_entry(next, struct page_stat, node);
node             1264 tools/perf/builtin-kmem.c 		this = rb_entry(*new, struct alloc_stat, node);
node             1279 tools/perf/builtin-kmem.c 	rb_link_node(&data->node, parent, new);
node             1280 tools/perf/builtin-kmem.c 	rb_insert_color(&data->node, root);
node             1286 tools/perf/builtin-kmem.c 	struct rb_node *node;
node             1290 tools/perf/builtin-kmem.c 		node = rb_first(root);
node             1291 tools/perf/builtin-kmem.c 		if (!node)
node             1294 tools/perf/builtin-kmem.c 		rb_erase(node, root);
node             1295 tools/perf/builtin-kmem.c 		data = rb_entry(node, struct alloc_stat, node);
node             1311 tools/perf/builtin-kmem.c 		this = rb_entry(*new, struct page_stat, node);
node             1326 tools/perf/builtin-kmem.c 	rb_link_node(&data->node, parent, new);
node             1327 tools/perf/builtin-kmem.c 	rb_insert_color(&data->node, root);
node             1333 tools/perf/builtin-kmem.c 	struct rb_node *node;
node             1337 tools/perf/builtin-kmem.c 		node = rb_first(root);
node             1338 tools/perf/builtin-kmem.c 		if (!node)
node             1341 tools/perf/builtin-kmem.c 		rb_erase(node, root);
node             1342 tools/perf/builtin-kmem.c 		data = rb_entry(node, struct page_stat, node);
node              562 tools/perf/builtin-kvm.c 	struct rb_node *node = rb_first(result);
node              564 tools/perf/builtin-kvm.c 	if (!node)
node              567 tools/perf/builtin-kvm.c 	rb_erase(node, result);
node              568 tools/perf/builtin-kvm.c 	return container_of(node, struct kvm_event, rb);
node              125 tools/perf/builtin-lock.c 	struct rb_node *node;
node              128 tools/perf/builtin-lock.c 	node = thread_stats.rb_node;
node              129 tools/perf/builtin-lock.c 	while (node) {
node              130 tools/perf/builtin-lock.c 		st = container_of(node, struct thread_stat, rb);
node              134 tools/perf/builtin-lock.c 			node = node->rb_left;
node              136 tools/perf/builtin-lock.c 			node = node->rb_right;
node              306 tools/perf/builtin-lock.c 	struct rb_node *node = result.rb_node;
node              308 tools/perf/builtin-lock.c 	if (!node)
node              311 tools/perf/builtin-lock.c 	while (node->rb_left)
node              312 tools/perf/builtin-lock.c 		node = node->rb_left;
node              314 tools/perf/builtin-lock.c 	rb_erase(node, &result);
node              315 tools/perf/builtin-lock.c 	return container_of(node, struct lock_stat, rb);
node              766 tools/perf/builtin-lock.c 	struct rb_node *node;
node              771 tools/perf/builtin-lock.c 	node = rb_first(&thread_stats);
node              772 tools/perf/builtin-lock.c 	while (node) {
node              773 tools/perf/builtin-lock.c 		st = container_of(node, struct thread_stat, rb);
node              776 tools/perf/builtin-lock.c 		node = rb_next(node);
node              128 tools/perf/builtin-sched.c 	struct rb_node		node;
node              963 tools/perf/builtin-sched.c 	struct rb_node *node = root->rb_root.rb_node;
node              966 tools/perf/builtin-sched.c 	while (node) {
node              970 tools/perf/builtin-sched.c 		atoms = container_of(node, struct work_atoms, node);
node              974 tools/perf/builtin-sched.c 			node = node->rb_left;
node              976 tools/perf/builtin-sched.c 			node = node->rb_right;
node              996 tools/perf/builtin-sched.c 		this = container_of(*new, struct work_atoms, node);
node             1009 tools/perf/builtin-sched.c 	rb_link_node(&data->node, parent, new);
node             1010 tools/perf/builtin-sched.c 	rb_insert_color_cached(&data->node, root, leftmost);
node             1459 tools/perf/builtin-sched.c 	struct rb_node *node;
node             1464 tools/perf/builtin-sched.c 		node = rb_first_cached(root);
node             1465 tools/perf/builtin-sched.c 		if (!node)
node             1468 tools/perf/builtin-sched.c 		rb_erase_cached(node, root);
node             1469 tools/perf/builtin-sched.c 		data = rb_entry(node, struct work_atoms, node);
node             2168 tools/perf/builtin-sched.c 		struct callchain_cursor_node *node;
node             2171 tools/perf/builtin-sched.c 		node = callchain_cursor_current(cursor);
node             2172 tools/perf/builtin-sched.c 		if (node == NULL)
node             2175 tools/perf/builtin-sched.c 		sym = node->sym;
node             2745 tools/perf/builtin-sched.c static size_t callchain__fprintf_folded(FILE *fp, struct callchain_node *node)
node             2753 tools/perf/builtin-sched.c 	if (node == NULL)
node             2756 tools/perf/builtin-sched.c 	ret = callchain__fprintf_folded(fp, node->parent);
node             2759 tools/perf/builtin-sched.c 	list_for_each_entry(chain, &node->val, list) {
node             2938 tools/perf/builtin-sched.c 	list_for_each_entry(evsel, &evlist->core.entries, core.node) {
node             3093 tools/perf/builtin-sched.c 		this = container_of(*new, struct work_atoms, node);
node             3119 tools/perf/builtin-sched.c 	rb_link_node(&data->node, parent, new);
node             3120 tools/perf/builtin-sched.c 	rb_insert_color_cached(&data->node, root, leftmost);
node             3126 tools/perf/builtin-sched.c 	struct rb_node *node;
node             3131 tools/perf/builtin-sched.c 	while ((node = rb_first_cached(&sched->atom_root))) {
node             3132 tools/perf/builtin-sched.c 		rb_erase_cached(node, &sched->atom_root);
node             3133 tools/perf/builtin-sched.c 		data = rb_entry(node, struct work_atoms, node);
node             3159 tools/perf/builtin-sched.c 		work_list = rb_entry(next, struct work_atoms, node);
node             2527 tools/perf/builtin-script.c 	struct list_head	node;
node             2549 tools/perf/builtin-script.c 	list_add_tail(&s->node, &script_specs);
node             2556 tools/perf/builtin-script.c 	list_for_each_entry(s, &script_specs, node)
node             2596 tools/perf/builtin-script.c 	list_for_each_entry(s, &script_specs, node)
node             2827 tools/perf/builtin-script.c 	struct list_head	node;
node             2855 tools/perf/builtin-script.c 	list_add_tail(&s->node, &script_descs);
node             2862 tools/perf/builtin-script.c 	list_for_each_entry(s, &script_descs, node)
node             3001 tools/perf/builtin-script.c 	list_for_each_entry(desc, &script_descs, node) {
node             2651 tools/perf/builtin-trace.c 		list_del_init(&evsel->core.node);
node               62 tools/perf/lib/evlist.c 	list_add_tail(&evsel->node, &evlist->entries);
node               70 tools/perf/lib/evlist.c 	list_del_init(&evsel->node);
node               92 tools/perf/lib/evlist.c 					node);
node               94 tools/perf/lib/evlist.c 		next = list_next_entry(prev, node);
node               98 tools/perf/lib/evlist.c 	if (&next->node == &evlist->entries)
node              194 tools/perf/lib/evlist.c 	hlist_add_head(&sid->node, &evlist->heads[hash]);
node               21 tools/perf/lib/evsel.c 	INIT_LIST_HEAD(&evsel->node);
node               37 tools/perf/lib/include/internal/evlist.h 	list_for_each_entry(evsel, list, node)
node               53 tools/perf/lib/include/internal/evlist.h 	list_for_each_entry_reverse(evsel, list, node)
node               65 tools/perf/lib/include/internal/evlist.h 	return list_entry(evlist->entries.next, struct perf_evsel, node);
node               70 tools/perf/lib/include/internal/evlist.h 	return list_entry(evlist->entries.prev, struct perf_evsel, node);
node               19 tools/perf/lib/include/internal/evsel.h 	struct hlist_node	 node;
node               38 tools/perf/lib/include/internal/evsel.h 	struct list_head	 node;
node              167 tools/perf/tests/hists_common.c 	struct rb_node *node;
node              175 tools/perf/tests/hists_common.c 	node = rb_first_cached(root);
node              176 tools/perf/tests/hists_common.c 	while (node) {
node              179 tools/perf/tests/hists_common.c 		he = rb_entry(node, struct hist_entry, rb_node_in);
node              189 tools/perf/tests/hists_common.c 		node = rb_next(node);
node              197 tools/perf/tests/hists_common.c 	struct rb_node *node;
node              202 tools/perf/tests/hists_common.c 	node = rb_first_cached(root);
node              203 tools/perf/tests/hists_common.c 	while (node) {
node              206 tools/perf/tests/hists_common.c 		he = rb_entry(node, struct hist_entry, rb_node);
node              217 tools/perf/tests/hists_common.c 		node = rb_next(node);
node              131 tools/perf/tests/hists_cumulate.c 	struct rb_node *node;
node              141 tools/perf/tests/hists_cumulate.c 		node = rb_first_cached(root_out);
node              143 tools/perf/tests/hists_cumulate.c 		he = rb_entry(node, struct hist_entry, rb_node);
node              144 tools/perf/tests/hists_cumulate.c 		rb_erase_cached(node, root_out);
node              174 tools/perf/tests/hists_cumulate.c 	} node[10];
node              184 tools/perf/tests/hists_cumulate.c 	struct rb_node *node;
node              203 tools/perf/tests/hists_cumulate.c 	for (node = rb_first(root), i = 0;
node              204 tools/perf/tests/hists_cumulate.c 	     node && (he = rb_entry(node, struct hist_entry, rb_node));
node              205 tools/perf/tests/hists_cumulate.c 	     node = rb_next(node), i++) {
node              222 tools/perf/tests/hists_cumulate.c 		root = &he->callchain->node.rb_root;
node              234 tools/perf/tests/hists_cumulate.c 				!strcmp(CDSO(clist), expected_callchain[i].node[c].dso) &&
node              235 tools/perf/tests/hists_cumulate.c 				!strcmp(CSYM(clist), expected_callchain[i].node[c].sym));
node              145 tools/perf/tests/hists_link.c 	struct rb_node *node;
node              155 tools/perf/tests/hists_link.c 	node = rb_first_cached(root);
node              156 tools/perf/tests/hists_link.c 	while (node) {
node              159 tools/perf/tests/hists_link.c 		he = rb_entry(node, struct hist_entry, rb_node_in);
node              172 tools/perf/tests/hists_link.c 		node = rb_next(node);
node              195 tools/perf/tests/hists_link.c 	struct rb_node *node;
node              207 tools/perf/tests/hists_link.c 	node = rb_first_cached(root);
node              208 tools/perf/tests/hists_link.c 	while (node) {
node              211 tools/perf/tests/hists_link.c 		he = rb_entry(node, struct hist_entry, rb_node_in);
node              229 tools/perf/tests/hists_link.c 		node = rb_next(node);
node               97 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              107 tools/perf/tests/hists_output.c 		node = rb_first_cached(root_out);
node              109 tools/perf/tests/hists_output.c 		he = rb_entry(node, struct hist_entry, rb_node);
node              110 tools/perf/tests/hists_output.c 		rb_erase_cached(node, root_out);
node              131 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              166 tools/perf/tests/hists_output.c 	node = rb_first_cached(root);
node              167 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              172 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              173 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              178 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              179 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              184 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              185 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              190 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              191 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              196 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              197 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              202 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              203 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              208 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              209 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              214 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              215 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              233 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              266 tools/perf/tests/hists_output.c 	node = rb_first_cached(root);
node              267 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              271 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              272 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              289 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              320 tools/perf/tests/hists_output.c 	node = rb_first_cached(root);
node              321 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              326 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              327 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              332 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              333 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              338 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              339 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              344 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              345 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              363 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              398 tools/perf/tests/hists_output.c 	node = rb_first_cached(root);
node              399 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              404 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              405 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              410 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              411 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              416 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              417 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              422 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              423 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              428 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              429 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              434 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              435 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              440 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              441 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              446 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              447 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              465 tools/perf/tests/hists_output.c 	struct rb_node *node;
node              501 tools/perf/tests/hists_output.c 	node = rb_first_cached(root);
node              502 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              509 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              510 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              516 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              517 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              523 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              524 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              530 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              531 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              537 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              538 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              544 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              545 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              551 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              552 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              558 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              559 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node              565 tools/perf/tests/hists_output.c 	node = rb_next(node);
node              566 tools/perf/tests/hists_output.c 	he = rb_entry(node, struct hist_entry, rb_node);
node               14 tools/perf/tests/mem2node.c 	int		 node;
node               17 tools/perf/tests/mem2node.c 	{ .node = 0, .map = "0"     },
node               18 tools/perf/tests/mem2node.c 	{ .node = 1, .map = "1-2"   },
node               19 tools/perf/tests/mem2node.c 	{ .node = 3, .map = "5-7,9" },
node               58 tools/perf/tests/mem2node.c 		nodes[i].node = test_nodes[i].node;
node              215 tools/perf/tests/switch-tracking.c 	struct event_node *node;
node              217 tools/perf/tests/switch-tracking.c 	node = malloc(sizeof(struct event_node));
node              218 tools/perf/tests/switch-tracking.c 	if (!node) {
node              222 tools/perf/tests/switch-tracking.c 	node->event = event;
node              223 tools/perf/tests/switch-tracking.c 	list_add(&node->list, events);
node              235 tools/perf/tests/switch-tracking.c 	node->event_time = sample.time;
node              242 tools/perf/tests/switch-tracking.c 	struct event_node *node;
node              245 tools/perf/tests/switch-tracking.c 		node = list_entry(events->next, struct event_node, list);
node              246 tools/perf/tests/switch-tracking.c 		list_del_init(&node->list);
node              247 tools/perf/tests/switch-tracking.c 		free(node);
node              266 tools/perf/tests/switch-tracking.c 	struct event_node *events_array, *node;
node              293 tools/perf/tests/switch-tracking.c 	list_for_each_entry(node, &events, list)
node              294 tools/perf/tests/switch-tracking.c 		events_array[pos++] = *node;
node               49 tools/perf/ui/browsers/annotate.c 	struct annotation_line *al = list_entry(entry, struct annotation_line, node);
node              100 tools/perf/ui/browsers/annotate.c 	struct annotation_line *al = list_entry(entry, struct annotation_line, node);
node              130 tools/perf/ui/browsers/annotate.c 	struct disasm_line *pos = list_prev_entry(cursor, al.node);
node              276 tools/perf/ui/browsers/annotate.c 		pos = list_entry(pos->node.prev, struct annotation_line, node);
node              316 tools/perf/ui/browsers/annotate.c 	list_for_each_entry(pos, &notes->src->source, al.node) {
node              353 tools/perf/ui/browsers/annotate.c 	al = list_entry(browser->b.top, struct annotation_line, node);
node              448 tools/perf/ui/browsers/annotate.c 	list_for_each_entry(pos, &notes->src->source, al.node) {
node              494 tools/perf/ui/browsers/annotate.c 	list_for_each_entry_continue(al, &notes->src->source, node) {
node              531 tools/perf/ui/browsers/annotate.c 	list_for_each_entry_continue_reverse(al, &notes->src->source, node) {
node              173 tools/perf/ui/browsers/hists.c static int callchain_node__count_rows_rb_tree(struct callchain_node *node)
node              178 tools/perf/ui/browsers/hists.c 	for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
node              199 tools/perf/ui/browsers/hists.c static int callchain_node__count_flat_rows(struct callchain_node *node)
node              205 tools/perf/ui/browsers/hists.c 	list_for_each_entry(chain, &node->parent_val, list) {
node              215 tools/perf/ui/browsers/hists.c 	list_for_each_entry(chain, &node->val, list) {
node              228 tools/perf/ui/browsers/hists.c static int callchain_node__count_folded_rows(struct callchain_node *node __maybe_unused)
node              233 tools/perf/ui/browsers/hists.c static int callchain_node__count_rows(struct callchain_node *node)
node              240 tools/perf/ui/browsers/hists.c 		return callchain_node__count_flat_rows(node);
node              242 tools/perf/ui/browsers/hists.c 		return callchain_node__count_folded_rows(node);
node              244 tools/perf/ui/browsers/hists.c 	list_for_each_entry(chain, &node->val, list) {
node              251 tools/perf/ui/browsers/hists.c 		n += callchain_node__count_rows_rb_tree(node);
node              262 tools/perf/ui/browsers/hists.c 		struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
node              263 tools/perf/ui/browsers/hists.c 		n += callchain_node__count_rows(node);
node              273 tools/perf/ui/browsers/hists.c 	struct rb_node *node;
node              282 tools/perf/ui/browsers/hists.c 	node = rb_first_cached(&he->hroot_out);
node              283 tools/perf/ui/browsers/hists.c 	while (node) {
node              286 tools/perf/ui/browsers/hists.c 		child = rb_entry(node, struct hist_entry, rb_node);
node              296 tools/perf/ui/browsers/hists.c 		node = rb_next(node);
node              325 tools/perf/ui/browsers/hists.c static void callchain_node__init_have_children_rb_tree(struct callchain_node *node)
node              327 tools/perf/ui/browsers/hists.c 	struct rb_node *nd = rb_first(&node->rb_root);
node              329 tools/perf/ui/browsers/hists.c 	for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
node              348 tools/perf/ui/browsers/hists.c static void callchain_node__init_have_children(struct callchain_node *node,
node              353 tools/perf/ui/browsers/hists.c 	chain = list_entry(node->val.next, struct callchain_list, list);
node              356 tools/perf/ui/browsers/hists.c 	if (!list_empty(&node->val)) {
node              357 tools/perf/ui/browsers/hists.c 		chain = list_entry(node->val.prev, struct callchain_list, list);
node              358 tools/perf/ui/browsers/hists.c 		chain->has_children = !RB_EMPTY_ROOT(&node->rb_root);
node              361 tools/perf/ui/browsers/hists.c 	callchain_node__init_have_children_rb_tree(node);
node              370 tools/perf/ui/browsers/hists.c 		struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
node              371 tools/perf/ui/browsers/hists.c 		callchain_node__init_have_children(node, has_sibling);
node              374 tools/perf/ui/browsers/hists.c 			callchain_node__make_parent_list(node);
node              461 tools/perf/ui/browsers/hists.c static int callchain_node__set_folding_rb_tree(struct callchain_node *node, bool unfold)
node              466 tools/perf/ui/browsers/hists.c 	for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
node              484 tools/perf/ui/browsers/hists.c static int callchain_node__set_folding(struct callchain_node *node, bool unfold)
node              490 tools/perf/ui/browsers/hists.c 	list_for_each_entry(chain, &node->val, list) {
node              497 tools/perf/ui/browsers/hists.c 		n += callchain_node__set_folding_rb_tree(node, unfold);
node              508 tools/perf/ui/browsers/hists.c 		struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
node              509 tools/perf/ui/browsers/hists.c 		n += callchain_node__set_folding(node, unfold);
node              793 tools/perf/ui/browsers/hists.c 					     struct callchain_node *node,
node              827 tools/perf/ui/browsers/hists.c 		callchain_node__scnprintf_value(node, buf, sizeof(buf),
node              843 tools/perf/ui/browsers/hists.c static bool check_percent_display(struct rb_node *node, u64 parent_total)
node              847 tools/perf/ui/browsers/hists.c 	if (node == NULL)
node              850 tools/perf/ui/browsers/hists.c 	if (rb_next(node))
node              853 tools/perf/ui/browsers/hists.c 	child = rb_entry(node, struct callchain_node, rb_node);
node              865 tools/perf/ui/browsers/hists.c 	struct rb_node *node;
node              869 tools/perf/ui/browsers/hists.c 	node = rb_first(root);
node              870 tools/perf/ui/browsers/hists.c 	need_percent = check_percent_display(node, parent_total);
node              872 tools/perf/ui/browsers/hists.c 	while (node) {
node              873 tools/perf/ui/browsers/hists.c 		struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
node              874 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
node              929 tools/perf/ui/browsers/hists.c 		node = next;
node              969 tools/perf/ui/browsers/hists.c 	struct rb_node *node;
node              973 tools/perf/ui/browsers/hists.c 	node = rb_first(root);
node              974 tools/perf/ui/browsers/hists.c 	need_percent = check_percent_display(node, parent_total);
node              976 tools/perf/ui/browsers/hists.c 	while (node) {
node              977 tools/perf/ui/browsers/hists.c 		struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
node              978 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
node             1040 tools/perf/ui/browsers/hists.c 		node = next;
node             1054 tools/perf/ui/browsers/hists.c 	struct rb_node *node;
node             1062 tools/perf/ui/browsers/hists.c 	node = rb_first(root);
node             1063 tools/perf/ui/browsers/hists.c 	need_percent = check_percent_display(node, parent_total);
node             1065 tools/perf/ui/browsers/hists.c 	while (node) {
node             1066 tools/perf/ui/browsers/hists.c 		struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
node             1067 tools/perf/ui/browsers/hists.c 		struct rb_node *next = rb_next(node);
node             1106 tools/perf/ui/browsers/hists.c 		node = next;
node             3225 tools/perf/ui/browsers/hists.c 	struct evsel *evsel = list_entry(entry, struct evsel, core.node);
node             3322 tools/perf/ui/browsers/hists.c 				if (pos->core.node.next == &evlist->core.entries)
node             3328 tools/perf/ui/browsers/hists.c 				if (pos->core.node.prev == &evlist->core.entries)
node             3364 tools/perf/ui/browsers/hists.c 	struct evsel *evsel = list_entry(entry, struct evsel, core.node);
node              126 tools/perf/ui/gtk/annotate.c 	list_for_each_entry(pos, &notes->src->source, al.node) {
node              155 tools/perf/ui/gtk/annotate.c 	list_for_each_entry_safe(pos, n, &notes->src->source, al.node) {
node              156 tools/perf/ui/gtk/annotate.c 		list_del_init(&pos->al.node);
node              104 tools/perf/ui/gtk/hists.c 		struct callchain_node *node;
node              109 tools/perf/ui/gtk/hists.c 		node = rb_entry(nd, struct callchain_node, rb_node);
node              114 tools/perf/ui/gtk/hists.c 		callchain_node__make_parent_list(node);
node              116 tools/perf/ui/gtk/hists.c 		list_for_each_entry(chain, &node->parent_val, list) {
node              121 tools/perf/ui/gtk/hists.c 			callchain_node__scnprintf_value(node, buf, sizeof(buf), total);
node              137 tools/perf/ui/gtk/hists.c 		list_for_each_entry(chain, &node->val, list) {
node              142 tools/perf/ui/gtk/hists.c 			callchain_node__scnprintf_value(node, buf, sizeof(buf), total);
node              166 tools/perf/ui/gtk/hists.c 		struct callchain_node *node;
node              173 tools/perf/ui/gtk/hists.c 		node = rb_entry(nd, struct callchain_node, rb_node);
node              175 tools/perf/ui/gtk/hists.c 		callchain_node__make_parent_list(node);
node              177 tools/perf/ui/gtk/hists.c 		list_for_each_entry(chain, &node->parent_val, list) {
node              193 tools/perf/ui/gtk/hists.c 		list_for_each_entry(chain, &node->val, list) {
node              211 tools/perf/ui/gtk/hists.c 		callchain_node__scnprintf_value(node, buf, sizeof(buf), total);
node              227 tools/perf/ui/gtk/hists.c 		struct callchain_node *node;
node              233 tools/perf/ui/gtk/hists.c 		node = rb_entry(nd, struct callchain_node, rb_node);
node              236 tools/perf/ui/gtk/hists.c 		need_new_parent = !has_single_node && (node->val_nr > 1);
node              238 tools/perf/ui/gtk/hists.c 		list_for_each_entry(chain, &node->val, list) {
node              243 tools/perf/ui/gtk/hists.c 			callchain_node__scnprintf_value(node, buf, sizeof(buf), total);
node              260 tools/perf/ui/gtk/hists.c 			child_total = node->children_hit;
node              265 tools/perf/ui/gtk/hists.c 		perf_gtk__add_callchain_graph(&node->rb_root, store, &iter, col,
node              413 tools/perf/ui/gtk/hists.c 	struct rb_node *node;
node              420 tools/perf/ui/gtk/hists.c 	for (node = rb_first_cached(root); node; node = rb_next(node)) {
node              425 tools/perf/ui/gtk/hists.c 		he = rb_entry(node, struct hist_entry, rb_node);
node               53 tools/perf/ui/hist.c 		list_for_each_entry(pair, &he->pairs.head, pairs.node) {
node              178 tools/perf/ui/hist.c 	list_for_each_entry(pair, &a->pairs.head, pairs.node) {
node              183 tools/perf/ui/hist.c 	list_for_each_entry(pair, &b->pairs.head, pairs.node) {
node              730 tools/perf/ui/hist.c 	struct perf_hpp_list_node *node;
node              736 tools/perf/ui/hist.c 	list_for_each_entry(node, &hists->hpp_formats, list) {
node              737 tools/perf/ui/hist.c 		perf_hpp_list__for_each_format(&node->hpp, fmt)
node              762 tools/perf/ui/hist.c 	struct perf_hpp_list_node *node = NULL;
node              767 tools/perf/ui/hist.c 	list_for_each_entry(node, &hists->hpp_formats, list) {
node              768 tools/perf/ui/hist.c 		if (node->level == fmt->level) {
node              775 tools/perf/ui/hist.c 		node = malloc(sizeof(*node));
node              776 tools/perf/ui/hist.c 		if (node == NULL)
node              779 tools/perf/ui/hist.c 		node->skip = skip;
node              780 tools/perf/ui/hist.c 		node->level = fmt->level;
node              781 tools/perf/ui/hist.c 		perf_hpp_list__init(&node->hpp);
node              784 tools/perf/ui/hist.c 		list_add_tail(&node->list, &hists->hpp_formats);
node              792 tools/perf/ui/hist.c 		node->skip = false;
node              794 tools/perf/ui/hist.c 	list_add_tail(&fmt_copy->list, &node->hpp.fields);
node              795 tools/perf/ui/hist.c 	list_add_tail(&fmt_copy->sort_list, &node->hpp.sorts);
node               49 tools/perf/ui/stdio/hist.c static size_t ipchain__fprintf_graph(FILE *fp, struct callchain_node *node,
node               68 tools/perf/ui/stdio/hist.c 			ret += callchain_node__fprintf_value(node, fp, total_samples);
node              112 tools/perf/ui/stdio/hist.c 	struct rb_node *node, *next;
node              124 tools/perf/ui/stdio/hist.c 	node = rb_first(root);
node              125 tools/perf/ui/stdio/hist.c 	while (node) {
node              129 tools/perf/ui/stdio/hist.c 		child = rb_entry(node, struct callchain_node, rb_node);
node              141 tools/perf/ui/stdio/hist.c 		next = rb_next(node);
node              168 tools/perf/ui/stdio/hist.c 		node = next;
node              205 tools/perf/ui/stdio/hist.c static bool need_percent_display(struct rb_node *node, u64 parent_samples)
node              209 tools/perf/ui/stdio/hist.c 	if (rb_next(node))
node              212 tools/perf/ui/stdio/hist.c 	cnode = rb_entry(node, struct callchain_node, rb_node);
node              224 tools/perf/ui/stdio/hist.c 	struct rb_node *node;
node              229 tools/perf/ui/stdio/hist.c 	node = rb_first(root);
node              230 tools/perf/ui/stdio/hist.c 	if (node && !need_percent_display(node, parent_samples)) {
node              231 tools/perf/ui/stdio/hist.c 		cnode = rb_entry(node, struct callchain_node, rb_node);
node              281 tools/perf/ui/stdio/hist.c static size_t __callchain__fprintf_flat(FILE *fp, struct callchain_node *node,
node              288 tools/perf/ui/stdio/hist.c 	if (!node)
node              291 tools/perf/ui/stdio/hist.c 	ret += __callchain__fprintf_flat(fp, node->parent, total_samples);
node              294 tools/perf/ui/stdio/hist.c 	list_for_each_entry(chain, &node->val, list) {
node              329 tools/perf/ui/stdio/hist.c static size_t __callchain__fprintf_folded(FILE *fp, struct callchain_node *node)
node              337 tools/perf/ui/stdio/hist.c 	if (!node)
node              340 tools/perf/ui/stdio/hist.c 	ret += __callchain__fprintf_folded(fp, node->parent);
node              343 tools/perf/ui/stdio/hist.c 	list_for_each_entry(chain, &node->val, list) {
node             1258 tools/perf/util/annotate.c 	list_add_tail(&al->node, head);
node             1264 tools/perf/util/annotate.c 	list_for_each_entry_continue(pos, head, node)
node             1402 tools/perf/util/annotate.c 			list_for_each_entry_from(queue, &notes->src->source, node) {
node             1583 tools/perf/util/annotate.c 		dl = list_entry(list->prev, struct disasm_line, al.node);
node             1595 tools/perf/util/annotate.c 		list_del_init(&dl->al.node);
node             1774 tools/perf/util/annotate.c 		struct btf_node *node;
node             1776 tools/perf/util/annotate.c 		node = perf_env__find_btf(dso->bpf_prog.env,
node             1778 tools/perf/util/annotate.c 		if (node)
node             1779 tools/perf/util/annotate.c 			btf = btf__new((__u8 *)(node->data),
node             1780 tools/perf/util/annotate.c 				       node->data_size);
node             2042 tools/perf/util/annotate.c 	list_for_each_entry(al, &notes->src->source, node) {
node             2187 tools/perf/util/annotate.c 	struct rb_node *node;
node             2189 tools/perf/util/annotate.c 	node = rb_first(src_root);
node             2190 tools/perf/util/annotate.c 	while (node) {
node             2193 tools/perf/util/annotate.c 		al = rb_entry(node, struct annotation_line, rb_node);
node             2194 tools/perf/util/annotate.c 		next = rb_next(node);
node             2195 tools/perf/util/annotate.c 		rb_erase(node, src_root);
node             2198 tools/perf/util/annotate.c 		node = next;
node             2205 tools/perf/util/annotate.c 	struct rb_node *node;
node             2215 tools/perf/util/annotate.c 	node = rb_first(root);
node             2216 tools/perf/util/annotate.c 	while (node) {
node             2222 tools/perf/util/annotate.c 		al = rb_entry(node, struct annotation_line, rb_node);
node             2236 tools/perf/util/annotate.c 		node = rb_next(node);
node             2258 tools/perf/util/annotate.c 	list_for_each_entry_reverse(line, lines, node) {
node             2318 tools/perf/util/annotate.c 	list_for_each_entry(pos, &notes->src->source, node) {
node             2352 tools/perf/util/annotate.c 				queue = list_entry(queue->node.next, typeof(*queue), node);
node             2420 tools/perf/util/annotate.c 	list_for_each_entry(al, &notes->src->source, node) {
node             2488 tools/perf/util/annotate.c 	list_for_each_entry_safe(al, n, &as->source, node) {
node             2489 tools/perf/util/annotate.c 		list_del_init(&al->node);
node             2516 tools/perf/util/annotate.c 	list_for_each_entry(pos, head, al.node)
node             2571 tools/perf/util/annotate.c 	list_for_each_entry(al, &notes->src->source, node) {
node             2607 tools/perf/util/annotate.c         list_for_each_entry(al, &notes->src->source, node) {
node             2648 tools/perf/util/annotate.c 	list_for_each_entry(al, &notes->src->source, node) {
node              131 tools/perf/util/annotate.h 	struct list_head	 node;
node               23 tools/perf/util/block-range.c 		struct block_range *entry = rb_entry(rb, struct block_range, node);
node               41 tools/perf/util/block-range.c 		entry = rb_entry(parent, struct block_range, node);
node               54 tools/perf/util/block-range.c static inline void rb_link_left_of_node(struct rb_node *left, struct rb_node *node)
node               56 tools/perf/util/block-range.c 	struct rb_node **p = &node->rb_left;
node               58 tools/perf/util/block-range.c 		node = *p;
node               59 tools/perf/util/block-range.c 		p = &node->rb_right;
node               61 tools/perf/util/block-range.c 	rb_link_node(left, node, p);
node               64 tools/perf/util/block-range.c static inline void rb_link_right_of_node(struct rb_node *right, struct rb_node *node)
node               66 tools/perf/util/block-range.c 	struct rb_node **p = &node->rb_right;
node               68 tools/perf/util/block-range.c 		node = *p;
node               69 tools/perf/util/block-range.c 		p = &node->rb_left;
node               71 tools/perf/util/block-range.c 	rb_link_node(right, node, p);
node               90 tools/perf/util/block-range.c 		entry = rb_entry(parent, struct block_range, node);
node              117 tools/perf/util/block-range.c 		next = rb_entry(n, struct block_range, node);
node              131 tools/perf/util/block-range.c 			rb_link_left_of_node(&head->node, &next->node);
node              132 tools/perf/util/block-range.c 			rb_insert_color(&head->node, &block_ranges.root);
node              154 tools/perf/util/block-range.c 		rb_link_node(&entry->node, parent, p);
node              155 tools/perf/util/block-range.c 		rb_insert_color(&entry->node, &block_ranges.root);
node              185 tools/perf/util/block-range.c 		rb_link_left_of_node(&head->node, &entry->node);
node              186 tools/perf/util/block-range.c 		rb_insert_color(&head->node, &block_ranges.root);
node              225 tools/perf/util/block-range.c 			rb_link_right_of_node(&tail->node, &entry->node);
node              226 tools/perf/util/block-range.c 			rb_insert_color(&tail->node, &block_ranges.root);
node              263 tools/perf/util/block-range.c 			rb_link_right_of_node(&tail->node, &entry->node);
node              264 tools/perf/util/block-range.c 			rb_insert_color(&tail->node, &block_ranges.root);
node              286 tools/perf/util/block-range.c 			rb_link_left_of_node(&hole->node, &next->node);
node              287 tools/perf/util/block-range.c 			rb_insert_color(&hole->node, &block_ranges.root);
node               23 tools/perf/util/block-range.h 	struct rb_node node;
node               40 tools/perf/util/block-range.h 	struct rb_node *n = rb_next(&br->node);
node               43 tools/perf/util/block-range.h 	return rb_entry(n, struct block_range, node);
node               97 tools/perf/util/bpf-event.c 	struct btf_node *node;
node              103 tools/perf/util/bpf-event.c 	node = malloc(data_size + sizeof(struct btf_node));
node              104 tools/perf/util/bpf-event.c 	if (!node)
node              107 tools/perf/util/bpf-event.c 	node->id = btf_id;
node              108 tools/perf/util/bpf-event.c 	node->data_size = data_size;
node              109 tools/perf/util/bpf-event.c 	memcpy(node->data, data, data_size);
node              111 tools/perf/util/bpf-event.c 	perf_env__insert_btf(env, node);
node              458 tools/perf/util/bpf-event.c 		struct btf_node *node;
node              460 tools/perf/util/bpf-event.c 		node = perf_env__find_btf(env, info->btf_id);
node              461 tools/perf/util/bpf-event.c 		if (node)
node              462 tools/perf/util/bpf-event.c 			btf = btf__new((__u8 *)(node->data),
node              463 tools/perf/util/bpf-event.c 				       node->data_size);
node              291 tools/perf/util/build-id.c 	list_for_each_entry(pos, head, node)	\
node              386 tools/perf/util/build-id.c 	list_for_each_entry(pos, head, node)
node               42 tools/perf/util/call-path.c 	list_for_each_entry_safe(pos, n, &cpr->blocks, node) {
node               43 tools/perf/util/call-path.c 		list_del_init(&pos->node);
node               60 tools/perf/util/call-path.c 				      node);
node               65 tools/perf/util/call-path.c 		list_add_tail(&cpb->node, &cpr->blocks);
node               44 tools/perf/util/call-path.h 	struct list_head node;
node              413 tools/perf/util/callchain.c __sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node,
node              419 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node              427 tools/perf/util/callchain.c 	if (node->hit && node->hit >= min_hit)
node              428 tools/perf/util/callchain.c 		rb_insert_callchain(rb_root, node, CHAIN_FLAT);
node              440 tools/perf/util/callchain.c 	__sort_chain_flat(rb_root, &root->node, min_hit);
node              443 tools/perf/util/callchain.c static void __sort_chain_graph_abs(struct callchain_node *node,
node              449 tools/perf/util/callchain.c 	node->rb_root = RB_ROOT;
node              450 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node              458 tools/perf/util/callchain.c 			rb_insert_callchain(&node->rb_root, child,
node              467 tools/perf/util/callchain.c 	__sort_chain_graph_abs(&chain_root->node, min_hit);
node              468 tools/perf/util/callchain.c 	rb_root->rb_node = chain_root->node.rb_root.rb_node;
node              471 tools/perf/util/callchain.c static void __sort_chain_graph_rel(struct callchain_node *node,
node              478 tools/perf/util/callchain.c 	node->rb_root = RB_ROOT;
node              479 tools/perf/util/callchain.c 	min_hit = ceil(node->children_hit * min_percent);
node              481 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node              488 tools/perf/util/callchain.c 			rb_insert_callchain(&node->rb_root, child,
node              497 tools/perf/util/callchain.c 	__sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0);
node              498 tools/perf/util/callchain.c 	rb_root->rb_node = chain_root->node.rb_root.rb_node;
node              566 tools/perf/util/callchain.c fill_node(struct callchain_node *node, struct callchain_cursor *cursor)
node              570 tools/perf/util/callchain.c 	node->val_nr = cursor->nr - cursor->pos;
node              571 tools/perf/util/callchain.c 	if (!node->val_nr)
node              621 tools/perf/util/callchain.c 		list_add_tail(&call->list, &node->val);
node              710 tools/perf/util/callchain.c static enum match_result match_chain(struct callchain_cursor_node *node,
node              717 tools/perf/util/callchain.c 		match = match_chain_strings(cnode->srcline, node->srcline);
node              723 tools/perf/util/callchain.c 		if (node->sym && cnode->ms.sym) {
node              730 tools/perf/util/callchain.c 			if (cnode->ms.sym->inlined || node->sym->inlined) {
node              732 tools/perf/util/callchain.c 							    node->sym->name);
node              737 tools/perf/util/callchain.c 								  node->map, node->sym->start);
node              745 tools/perf/util/callchain.c 		match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->map, node->ip);
node              749 tools/perf/util/callchain.c 	if (match == MATCH_EQ && node->branch) {
node              752 tools/perf/util/callchain.c 		if (node->branch_from) {
node              758 tools/perf/util/callchain.c 			if (node->branch_flags.predicted)
node              761 tools/perf/util/callchain.c 			if (node->branch_flags.abort)
node              765 tools/perf/util/callchain.c 					  &node->branch_flags,
node              766 tools/perf/util/callchain.c 					  node->branch_from,
node              767 tools/perf/util/callchain.c 					  node->ip);
node              773 tools/perf/util/callchain.c 			cnode->cycles_count += node->branch_flags.cycles;
node              774 tools/perf/util/callchain.c 			cnode->iter_count += node->nr_loop_iter;
node              775 tools/perf/util/callchain.c 			cnode->iter_cycles += node->iter_cycles;
node              825 tools/perf/util/callchain.c 		struct callchain_cursor_node *node;
node              833 tools/perf/util/callchain.c 		node = callchain_cursor_current(cursor);
node              847 tools/perf/util/callchain.c 		if (match_chain(node, cnode) == MATCH_LT)
node              872 tools/perf/util/callchain.c 	struct callchain_cursor_node *node;
node              876 tools/perf/util/callchain.c 	node = callchain_cursor_current(cursor);
node              877 tools/perf/util/callchain.c 	if (!node)
node              931 tools/perf/util/callchain.c 		struct callchain_cursor_node *node;
node              933 tools/perf/util/callchain.c 		node = callchain_cursor_current(cursor);
node              934 tools/perf/util/callchain.c 		if (!node)
node              937 tools/perf/util/callchain.c 		cmp = match_chain(node, cnode);
node              986 tools/perf/util/callchain.c 	if (append_chain_children(&root->node, cursor, period) < 0)
node             1043 tools/perf/util/callchain.c 	return merge_chain_branch(cursor, &dst->node, &src->node);
node             1052 tools/perf/util/callchain.c 	struct callchain_cursor_node *node = *cursor->last;
node             1054 tools/perf/util/callchain.c 	if (!node) {
node             1055 tools/perf/util/callchain.c 		node = calloc(1, sizeof(*node));
node             1056 tools/perf/util/callchain.c 		if (!node)
node             1059 tools/perf/util/callchain.c 		*cursor->last = node;
node             1062 tools/perf/util/callchain.c 	node->ip = ip;
node             1063 tools/perf/util/callchain.c 	map__zput(node->map);
node             1064 tools/perf/util/callchain.c 	node->map = map__get(map);
node             1065 tools/perf/util/callchain.c 	node->sym = sym;
node             1066 tools/perf/util/callchain.c 	node->branch = branch;
node             1067 tools/perf/util/callchain.c 	node->nr_loop_iter = nr_loop_iter;
node             1068 tools/perf/util/callchain.c 	node->iter_cycles = iter_cycles;
node             1069 tools/perf/util/callchain.c 	node->srcline = srcline;
node             1072 tools/perf/util/callchain.c 		memcpy(&node->branch_flags, flags,
node             1075 tools/perf/util/callchain.c 	node->branch_from = branch_from;
node             1078 tools/perf/util/callchain.c 	cursor->last = &node->next;
node             1107 tools/perf/util/callchain.c int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node,
node             1110 tools/perf/util/callchain.c 	al->map = node->map;
node             1111 tools/perf/util/callchain.c 	al->sym = node->sym;
node             1112 tools/perf/util/callchain.c 	al->srcline = node->srcline;
node             1113 tools/perf/util/callchain.c 	al->addr = node->ip;
node             1176 tools/perf/util/callchain.c char *callchain_node__scnprintf_value(struct callchain_node *node,
node             1180 tools/perf/util/callchain.c 	u64 period = callchain_cumul_hits(node);
node             1181 tools/perf/util/callchain.c 	unsigned count = callchain_cumul_counts(node);
node             1184 tools/perf/util/callchain.c 		period = node->hit;
node             1185 tools/perf/util/callchain.c 		count = node->count;
node             1205 tools/perf/util/callchain.c int callchain_node__fprintf_value(struct callchain_node *node,
node             1209 tools/perf/util/callchain.c 	u64 period = callchain_cumul_hits(node);
node             1210 tools/perf/util/callchain.c 	unsigned count = callchain_cumul_counts(node);
node             1213 tools/perf/util/callchain.c 		period = node->hit;
node             1214 tools/perf/util/callchain.c 		count = node->count;
node             1231 tools/perf/util/callchain.c static void callchain_counts_value(struct callchain_node *node,
node             1237 tools/perf/util/callchain.c 	list_for_each_entry(clist, &node->val, list) {
node             1252 tools/perf/util/callchain.c static int callchain_node_branch_counts_cumul(struct callchain_node *node,
node             1261 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node             1295 tools/perf/util/callchain.c 	return callchain_node_branch_counts_cumul(&root->node,
node             1454 tools/perf/util/callchain.c static void free_callchain_node(struct callchain_node *node)
node             1460 tools/perf/util/callchain.c 	list_for_each_entry_safe(list, tmp, &node->parent_val, list) {
node             1466 tools/perf/util/callchain.c 	list_for_each_entry_safe(list, tmp, &node->val, list) {
node             1472 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node             1476 tools/perf/util/callchain.c 		rb_erase(&child->rb_node_in, &node->rb_root_in);
node             1488 tools/perf/util/callchain.c 	free_callchain_node(&root->node);
node             1491 tools/perf/util/callchain.c static u64 decay_callchain_node(struct callchain_node *node)
node             1497 tools/perf/util/callchain.c 	n = rb_first(&node->rb_root_in);
node             1505 tools/perf/util/callchain.c 	node->hit = (node->hit * 7) / 8;
node             1506 tools/perf/util/callchain.c 	node->children_hit = child_hits;
node             1508 tools/perf/util/callchain.c 	return node->hit;
node             1516 tools/perf/util/callchain.c 	decay_callchain_node(&root->node);
node             1519 tools/perf/util/callchain.c int callchain_node__make_parent_list(struct callchain_node *node)
node             1521 tools/perf/util/callchain.c 	struct callchain_node *parent = node->parent;
node             1539 tools/perf/util/callchain.c 		list_move_tail(&chain->list, &node->parent_val);
node             1541 tools/perf/util/callchain.c 	if (!list_empty(&node->parent_val)) {
node             1542 tools/perf/util/callchain.c 		chain = list_first_entry(&node->parent_val, struct callchain_list, list);
node             1543 tools/perf/util/callchain.c 		chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node);
node             1545 tools/perf/util/callchain.c 		chain = list_first_entry(&node->val, struct callchain_list, list);
node             1568 tools/perf/util/callchain.c 		struct callchain_cursor_node *node;
node             1570 tools/perf/util/callchain.c 		node = callchain_cursor_current(src);
node             1571 tools/perf/util/callchain.c 		if (node == NULL)
node             1574 tools/perf/util/callchain.c 		rc = callchain_cursor_append(dst, node->ip, node->map, node->sym,
node             1575 tools/perf/util/callchain.c 					     node->branch, &node->branch_flags,
node             1576 tools/perf/util/callchain.c 					     node->nr_loop_iter,
node             1577 tools/perf/util/callchain.c 					     node->iter_cycles,
node             1578 tools/perf/util/callchain.c 					     node->branch_from, node->srcline);
node             1594 tools/perf/util/callchain.c 	struct callchain_cursor_node *node;
node             1599 tools/perf/util/callchain.c 	for (node = cursor->first; node != NULL; node = node->next)
node             1600 tools/perf/util/callchain.c 		map__zput(node->map);
node               76 tools/perf/util/callchain.h 	struct callchain_node	node;
node              167 tools/perf/util/callchain.h 	INIT_LIST_HEAD(&root->node.val);
node              168 tools/perf/util/callchain.h 	INIT_LIST_HEAD(&root->node.parent_val);
node              170 tools/perf/util/callchain.h 	root->node.parent = NULL;
node              171 tools/perf/util/callchain.h 	root->node.hit = 0;
node              172 tools/perf/util/callchain.h 	root->node.children_hit = 0;
node              173 tools/perf/util/callchain.h 	root->node.rb_root_in = RB_ROOT;
node              177 tools/perf/util/callchain.h static inline u64 callchain_cumul_hits(struct callchain_node *node)
node              179 tools/perf/util/callchain.h 	return node->hit + node->children_hit;
node              182 tools/perf/util/callchain.h static inline unsigned callchain_cumul_counts(struct callchain_node *node)
node              184 tools/perf/util/callchain.h 	return node->count + node->children_count;
node              246 tools/perf/util/callchain.h int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node,
node              277 tools/perf/util/callchain.h char *callchain_node__scnprintf_value(struct callchain_node *node,
node              279 tools/perf/util/callchain.h int callchain_node__fprintf_value(struct callchain_node *node,
node              287 tools/perf/util/callchain.h int callchain_node__make_parent_list(struct callchain_node *node);
node              515 tools/perf/util/config.c 	list_for_each_entry(section, sections, node)
node              527 tools/perf/util/config.c 	list_for_each_entry(item, &section->items, node)
node              550 tools/perf/util/config.c 	list_add_tail(&section->node, sections);
node              569 tools/perf/util/config.c 	list_add_tail(&item->node, &section->items);
node              780 tools/perf/util/config.c 	list_for_each_entry_safe(item, tmp, &section->items, node) {
node              781 tools/perf/util/config.c 		list_del_init(&item->node);
node              797 tools/perf/util/config.c 	list_for_each_entry_safe(section, tmp, &set->sections, node) {
node              798 tools/perf/util/config.c 		list_del_init(&section->node);
node               12 tools/perf/util/config.h 	struct list_head node;
node               19 tools/perf/util/config.h 	struct list_head node;
node               50 tools/perf/util/config.h         list_for_each_entry(section, list, node)
node               58 tools/perf/util/config.h         list_for_each_entry(item, list, node)
node              233 tools/perf/util/cputopo.c static int load_numa_node(struct numa_topology_node *node, int nr)
node              243 tools/perf/util/cputopo.c 	node->node = (u32) nr;
node              258 tools/perf/util/cputopo.c 			node->mem_total = mem;
node              260 tools/perf/util/cputopo.c 			node->mem_free = mem;
node              261 tools/perf/util/cputopo.c 		if (node->mem_total && node->mem_free)
node              282 tools/perf/util/cputopo.c 	node->cpus = buf;
node               18 tools/perf/util/cputopo.h 	u32		 node;
node             2566 tools/perf/util/cs-etm.c 	INIT_LIST_HEAD(&etm->unknown_thread->node);
node              238 tools/perf/util/db-export.c 		struct callchain_cursor_node *node;
node              244 tools/perf/util/db-export.c 		node = callchain_cursor_current(&callchain_cursor);
node              245 tools/perf/util/db-export.c 		if (!node)
node              252 tools/perf/util/db-export.c 		al.sym = node->sym;
node              253 tools/perf/util/db-export.c 		al.map = node->map;
node              255 tools/perf/util/db-export.c 		al.addr = node->ip;
node              264 tools/perf/util/db-export.c 					     al.sym, node->ip,
node              717 tools/perf/util/dso.c 	struct bpf_prog_info_node *node;
node              722 tools/perf/util/dso.c 	node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id);
node              723 tools/perf/util/dso.c 	if (!node || !node->info_linear) {
node              728 tools/perf/util/dso.c 	len = node->info_linear->info.jited_prog_len;
node              729 tools/perf/util/dso.c 	buf = (u8 *)(uintptr_t)node->info_linear->info.jited_prog_insns;
node              741 tools/perf/util/dso.c 	struct bpf_prog_info_node *node;
node              743 tools/perf/util/dso.c 	node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id);
node              744 tools/perf/util/dso.c 	if (!node || !node->info_linear) {
node              749 tools/perf/util/dso.c 	dso->data.file_size = node->info_linear->info.jited_prog_len;
node             1193 tools/perf/util/dso.c 		INIT_LIST_HEAD(&dso->node);
node              136 tools/perf/util/dso.h 	struct list_head node;
node               18 tools/perf/util/dsos.c 	list_for_each_entry(pos, head, node) {
node               92 tools/perf/util/dsos.c 	list_add_tail(&dso->node, &dsos->head);
node              129 tools/perf/util/dsos.c 		list_for_each_entry(pos, &dsos->head, node)
node              213 tools/perf/util/dsos.c 	list_for_each_entry(pos, head, node) {
node              227 tools/perf/util/dsos.c 	list_for_each_entry(pos, head, node) {
node               20 tools/perf/util/env.c 	struct bpf_prog_info_node *node;
node               29 tools/perf/util/env.c 		node = rb_entry(parent, struct bpf_prog_info_node, rb_node);
node               30 tools/perf/util/env.c 		if (prog_id < node->info_linear->info.id) {
node               32 tools/perf/util/env.c 		} else if (prog_id > node->info_linear->info.id) {
node               50 tools/perf/util/env.c 	struct bpf_prog_info_node *node = NULL;
node               57 tools/perf/util/env.c 		node = rb_entry(n, struct bpf_prog_info_node, rb_node);
node               58 tools/perf/util/env.c 		if (prog_id < node->info_linear->info.id)
node               60 tools/perf/util/env.c 		else if (prog_id > node->info_linear->info.id)
node               65 tools/perf/util/env.c 	node = NULL;
node               69 tools/perf/util/env.c 	return node;
node               76 tools/perf/util/env.c 	struct btf_node *node;
node               84 tools/perf/util/env.c 		node = rb_entry(parent, struct btf_node, rb_node);
node               85 tools/perf/util/env.c 		if (btf_id < node->id) {
node               87 tools/perf/util/env.c 		} else if (btf_id > node->id) {
node              104 tools/perf/util/env.c 	struct btf_node *node = NULL;
node              111 tools/perf/util/env.c 		node = rb_entry(n, struct btf_node, rb_node);
node              112 tools/perf/util/env.c 		if (btf_id < node->id)
node              114 tools/perf/util/env.c 		else if (btf_id > node->id)
node              119 tools/perf/util/env.c 	node = NULL;
node              123 tools/perf/util/env.c 	return node;
node              138 tools/perf/util/env.c 		struct bpf_prog_info_node *node;
node              140 tools/perf/util/env.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
node              141 tools/perf/util/env.c 		next = rb_next(&node->rb_node);
node              142 tools/perf/util/env.c 		rb_erase(&node->rb_node, root);
node              143 tools/perf/util/env.c 		free(node);
node              152 tools/perf/util/env.c 		struct btf_node *node;
node              154 tools/perf/util/env.c 		node = rb_entry(next, struct btf_node, rb_node);
node              155 tools/perf/util/env.c 		next = rb_next(&node->rb_node);
node              156 tools/perf/util/env.c 		rb_erase(&node->rb_node, root);
node              157 tools/perf/util/env.c 		free(node);
node               28 tools/perf/util/env.h 	u32		 node;
node               35 tools/perf/util/env.h 	u64		 node;
node              129 tools/perf/util/evlist.c 		list_del_init(&pos->core.node);
node              184 tools/perf/util/evlist.c 		list_del_init(&evsel->core.node);
node              193 tools/perf/util/evlist.c 	leader = list_entry(list->next, struct evsel, core.node);
node              194 tools/perf/util/evlist.c 	evsel = list_entry(list->prev, struct evsel, core.node);
node              249 tools/perf/util/evlist.c 		list_add_tail(&evsel->core.node, &head);
node              451 tools/perf/util/evlist.c 	hlist_for_each_entry(sid, head, node)
node              537 tools/perf/util/evlist.c 	hlist_for_each_entry(sid, head, node) {
node             1520 tools/perf/util/evlist.c 			list_move_tail(&evsel->core.node, &move);
node              264 tools/perf/util/evlist.h         list_for_each_entry(evsel, list, core.node)
node              280 tools/perf/util/evlist.h         list_for_each_entry_continue(evsel, list, core.node)
node              296 tools/perf/util/evlist.h         list_for_each_entry_reverse(evsel, list, core.node)
node              313 tools/perf/util/evlist.h         list_for_each_entry_safe(evsel, tmp, list, core.node)
node             1245 tools/perf/util/evsel.c 	assert(list_empty(&evsel->core.node));
node              309 tools/perf/util/evsel.h 	return list_entry(evsel->core.node.next, struct evsel, core.node);
node              314 tools/perf/util/evsel.h 	return list_entry(evsel->core.node.prev, struct evsel, core.node);
node              370 tools/perf/util/evsel.h for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \
node              372 tools/perf/util/evsel.h      (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
node              378 tools/perf/util/evsel.h      (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
node              109 tools/perf/util/evsel_fprintf.c 	struct callchain_cursor_node *node;
node              130 tools/perf/util/evsel_fprintf.c 			node = callchain_cursor_current(cursor);
node              131 tools/perf/util/evsel_fprintf.c 			if (!node)
node              134 tools/perf/util/evsel_fprintf.c 			if (node->sym && node->sym->ignore && print_skip_ignored)
node              143 tools/perf/util/evsel_fprintf.c 				printed += fprintf(fp, "%c%16" PRIx64, s, node->ip);
node              145 tools/perf/util/evsel_fprintf.c 			if (node->map)
node              146 tools/perf/util/evsel_fprintf.c 				addr = node->map->map_ip(node->map, node->ip);
node              151 tools/perf/util/evsel_fprintf.c 				node_al.map  = node->map;
node              154 tools/perf/util/evsel_fprintf.c 					printed += __symbol__fprintf_symname_offs(node->sym, &node_al,
node              158 tools/perf/util/evsel_fprintf.c 					printed += __symbol__fprintf_symname(node->sym, &node_al,
node              163 tools/perf/util/evsel_fprintf.c 			if (print_dso && (!node->sym || !node->sym->inlined)) {
node              165 tools/perf/util/evsel_fprintf.c 				printed += map__fprintf_dsoname(node->map, fp);
node              170 tools/perf/util/evsel_fprintf.c 				printed += map__fprintf_srcline(node->map, addr, "\n  ", fp);
node              172 tools/perf/util/evsel_fprintf.c 			if (node->sym && node->sym->inlined)
node              179 tools/perf/util/evsel_fprintf.c 			if (bt_stop_list && node->sym &&
node              180 tools/perf/util/evsel_fprintf.c 			    strlist__has_entry(bt_stop_list, node->sym->name)) {
node              689 tools/perf/util/header.c 		ret = do_write(ff, &n->node, sizeof(u32));
node              933 tools/perf/util/header.c 		struct bpf_prog_info_node *node;
node              936 tools/perf/util/header.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
node              937 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
node              939 tools/perf/util/header.c 			node->info_linear->data_len;
node              942 tools/perf/util/header.c 		bpf_program__bpil_addr_to_offs(node->info_linear);
node              943 tools/perf/util/header.c 		ret = do_write(ff, node->info_linear, len);
node              948 tools/perf/util/header.c 		bpf_program__bpil_offs_to_addr(node->info_linear);
node              983 tools/perf/util/header.c 		struct btf_node *node;
node              985 tools/perf/util/header.c 		node = rb_entry(next, struct btf_node, rb_node);
node              986 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
node              987 tools/perf/util/header.c 		ret = do_write(ff, &node->id,
node              988 tools/perf/util/header.c 			       sizeof(u32) * 2 + node->data_size);
node             1238 tools/perf/util/header.c 	n->node = idx;
node             1256 tools/perf/util/header.c 	return na->node - nb->node;
node             1360 tools/perf/util/header.c 		_W(node)
node             1544 tools/perf/util/header.c 		struct bpf_prog_info_node *node;
node             1546 tools/perf/util/header.c 		node = rb_entry(next, struct bpf_prog_info_node, rb_node);
node             1547 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
node             1549 tools/perf/util/header.c 		bpf_event__print_bpf_prog_info(&node->info_linear->info,
node             1568 tools/perf/util/header.c 		struct btf_node *node;
node             1570 tools/perf/util/header.c 		node = rb_entry(next, struct btf_node, rb_node);
node             1571 tools/perf/util/header.c 		next = rb_next(&node->rb_node);
node             1572 tools/perf/util/header.c 		fprintf(fp, "# btf info of id %u\n", node->id);
node             1730 tools/perf/util/header.c 			n->node, n->mem_total, n->mem_free);
node             1732 tools/perf/util/header.c 		fprintf(fp, "# node%u cpu list : ", n->node);
node             1867 tools/perf/util/header.c 	fprintf(fp, "#  %3" PRIu64 " [%s]: %s\n", n->node, buf_size, buf_map);
node             2347 tools/perf/util/header.c 		if (do_read_u32(ff, &n->node))
node             2623 tools/perf/util/header.c 		_R(node)
node             2748 tools/perf/util/header.c 	struct btf_node *node = NULL;
node             2770 tools/perf/util/header.c 		node = malloc(sizeof(struct btf_node) + data_size);
node             2771 tools/perf/util/header.c 		if (!node)
node             2774 tools/perf/util/header.c 		node->id = id;
node             2775 tools/perf/util/header.c 		node->data_size = data_size;
node             2777 tools/perf/util/header.c 		if (__do_read(ff, node->data, data_size))
node             2780 tools/perf/util/header.c 		perf_env__insert_btf(env, node);
node             2781 tools/perf/util/header.c 		node = NULL;
node             2787 tools/perf/util/header.c 	free(node);
node              321 tools/perf/util/hist.c 		struct rb_node *node = rb_first_cached(&he->hroot_out);
node              322 tools/perf/util/hist.c 		while (node) {
node              323 tools/perf/util/hist.c 			child = rb_entry(node, struct hist_entry, rb_node);
node              324 tools/perf/util/hist.c 			node = rb_next(node);
node              474 tools/perf/util/hist.c 	INIT_LIST_HEAD(&he->pairs.node);
node             1041 tools/perf/util/hist.c 	struct callchain_cursor_node *node;
node             1043 tools/perf/util/hist.c 	node = callchain_cursor_current(&callchain_cursor);
node             1044 tools/perf/util/hist.c 	if (node == NULL)
node             1047 tools/perf/util/hist.c 	return fill_callchain_info(al, node, iter->hide_unresolved);
node             1470 tools/perf/util/hist.c 	struct perf_hpp_list_node *node;
node             1476 tools/perf/util/hist.c 	list_for_each_entry(node, &hists->hpp_formats, list) {
node             1478 tools/perf/util/hist.c 		if (node->level == 0 || node->skip)
node             1482 tools/perf/util/hist.c 		new_he = hierarchy_insert_entry(hists, root, he, parent, &node->hpp);
node             1677 tools/perf/util/hist.c 	struct rb_node *node;
node             1680 tools/perf/util/hist.c 	node = rb_first_cached(&hists->entries);
node             1690 tools/perf/util/hist.c 	while (node) {
node             1691 tools/perf/util/hist.c 		he = rb_entry(node, struct hist_entry, rb_node);
node             1692 tools/perf/util/hist.c 		node = rb_next(node);
node             1738 tools/perf/util/hist.c 	struct rb_node *node;
node             1742 tools/perf/util/hist.c 	node = rb_first_cached(root_in);
node             1744 tools/perf/util/hist.c 	while (node) {
node             1745 tools/perf/util/hist.c 		he = rb_entry(node, struct hist_entry, rb_node_in);
node             1746 tools/perf/util/hist.c 		node = rb_next(node);
node             1928 tools/perf/util/hist.c struct rb_node *rb_hierarchy_last(struct rb_node *node)
node             1930 tools/perf/util/hist.c 	struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
node             1933 tools/perf/util/hist.c 		node = rb_last(&he->hroot_out.rb_root);
node             1934 tools/perf/util/hist.c 		he = rb_entry(node, struct hist_entry, rb_node);
node             1936 tools/perf/util/hist.c 	return node;
node             1939 tools/perf/util/hist.c struct rb_node *__rb_hierarchy_next(struct rb_node *node, enum hierarchy_move_dir hmd)
node             1941 tools/perf/util/hist.c 	struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
node             1944 tools/perf/util/hist.c 		node = rb_first_cached(&he->hroot_out);
node             1946 tools/perf/util/hist.c 		node = rb_next(node);
node             1948 tools/perf/util/hist.c 	while (node == NULL) {
node             1953 tools/perf/util/hist.c 		node = rb_next(&he->rb_node);
node             1955 tools/perf/util/hist.c 	return node;
node             1958 tools/perf/util/hist.c struct rb_node *rb_hierarchy_prev(struct rb_node *node)
node             1960 tools/perf/util/hist.c 	struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
node             1962 tools/perf/util/hist.c 	node = rb_prev(node);
node             1963 tools/perf/util/hist.c 	if (node)
node             1964 tools/perf/util/hist.c 		return rb_hierarchy_last(node);
node             1975 tools/perf/util/hist.c 	struct rb_node *node;
node             1982 tools/perf/util/hist.c 	node = rb_first_cached(&he->hroot_out);
node             1983 tools/perf/util/hist.c 	child = rb_entry(node, struct hist_entry, rb_node);
node             1985 tools/perf/util/hist.c 	while (node && child->filtered) {
node             1986 tools/perf/util/hist.c 		node = rb_next(node);
node             1987 tools/perf/util/hist.c 		child = rb_entry(node, struct hist_entry, rb_node);
node             1990 tools/perf/util/hist.c 	if (node)
node             1995 tools/perf/util/hist.c 	return node && percent >= limit;
node             2485 tools/perf/util/hist.c 			list_for_each_entry(leader, &pos->pairs.head, pairs.node) {
node             2565 tools/perf/util/hist.c 		list_del_init(&pos->pairs.node);
node             2743 tools/perf/util/hist.c 	struct rb_node *node;
node             2747 tools/perf/util/hist.c 		node = rb_first_cached(root);
node             2748 tools/perf/util/hist.c 		rb_erase_cached(node, root);
node             2750 tools/perf/util/hist.c 		he = rb_entry(node, struct hist_entry, rb_node_in);
node             2767 tools/perf/util/hist.c 	struct perf_hpp_list_node *node, *tmp;
node             2771 tools/perf/util/hist.c 	list_for_each_entry_safe(node, tmp, &hists->hpp_formats, list) {
node             2772 tools/perf/util/hist.c 		perf_hpp_list__for_each_format_safe(&node->hpp, fmt, pos) {
node             2776 tools/perf/util/hist.c 		list_del_init(&node->list);
node             2777 tools/perf/util/hist.c 		free(node);
node              544 tools/perf/util/hist.h struct rb_node *rb_hierarchy_last(struct rb_node *node);
node              545 tools/perf/util/hist.h struct rb_node *__rb_hierarchy_next(struct rb_node *node,
node              547 tools/perf/util/hist.h struct rb_node *rb_hierarchy_prev(struct rb_node *node);
node              549 tools/perf/util/hist.h static inline struct rb_node *rb_hierarchy_next(struct rb_node *node)
node              551 tools/perf/util/hist.h 	return __rb_hierarchy_next(node, HMD_NORMAL);
node             3196 tools/perf/util/intel-pt.c 	INIT_LIST_HEAD(&pt->unknown_thread->node);
node               18 tools/perf/util/intlist.c 	struct int_node *node = malloc(sizeof(*node));
node               20 tools/perf/util/intlist.c 	if (node != NULL) {
node               21 tools/perf/util/intlist.c 		node->i = i;
node               22 tools/perf/util/intlist.c 		node->priv = NULL;
node               23 tools/perf/util/intlist.c 		rc = &node->rb_node;
node               37 tools/perf/util/intlist.c 	struct int_node *node = container_of(rb_node, struct int_node, rb_node);
node               39 tools/perf/util/intlist.c 	int_node__delete(node);
node               45 tools/perf/util/intlist.c 	struct int_node *node = container_of(rb_node, struct int_node, rb_node);
node               47 tools/perf/util/intlist.c 	return node->i - i;
node               55 tools/perf/util/intlist.c void intlist__remove(struct intlist *ilist, struct int_node *node)
node               57 tools/perf/util/intlist.c 	rblist__remove_node(&ilist->rblist, &node->rb_node);
node               63 tools/perf/util/intlist.c 	struct int_node *node = NULL;
node               75 tools/perf/util/intlist.c 		node = container_of(rb_node, struct int_node, rb_node);
node               77 tools/perf/util/intlist.c 	return node;
node              137 tools/perf/util/intlist.c 	struct int_node *node = NULL;
node              142 tools/perf/util/intlist.c 		node = container_of(rb_node, struct int_node, rb_node);
node              144 tools/perf/util/intlist.c 	return node;
node              172 tools/perf/util/machine.c 	list_for_each_entry_safe(pos, n, &dsos->head, node) {
node              175 tools/perf/util/machine.c 		list_del_init(&pos->node);
node              233 tools/perf/util/machine.c 		list_for_each_entry_safe(thread, n, &threads->dead, node)
node              234 tools/perf/util/machine.c 			list_del_init(&thread->node);
node              379 tools/perf/util/machine.c 	struct rb_node *node;
node              384 tools/perf/util/machine.c 	for (node = rb_first_cached(&machines->guests); node;
node              385 tools/perf/util/machine.c 	     node = rb_next(node)) {
node              386 tools/perf/util/machine.c 		machine = rb_entry(node, struct machine, rb_node);
node             1503 tools/perf/util/machine.c 	list_for_each_entry(dso, &machine->dsos.head, node) {
node             1576 tools/perf/util/machine.c 		list_for_each_entry(dso, &machine->dsos.head, node) {
node             1774 tools/perf/util/machine.c 	list_add_tail(&th->node, &threads->dead);
node             2558 tools/perf/util/machine.c 		list_for_each_entry(thread, &threads->dead, node) {
node               24 tools/perf/util/map.h 		struct list_head node;
node               14 tools/perf/util/mem2node.c 	u64	node;
node               38 tools/perf/util/mem2node.c phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node)
node               42 tools/perf/util/mem2node.c 	entry->node  = node;
node               86 tools/perf/util/mem2node.c 				    (prev->node == n->node)) {
node               92 tools/perf/util/mem2node.c 			phys_entry__init(&entries[j++], start, bsize, n->node);
node              103 tools/perf/util/mem2node.c 			 entries[i].node, entries[i].start, entries[i].end);
node              136 tools/perf/util/mem2node.c 	return entry ? (int) entry->node : -1;
node              306 tools/perf/util/metricgroup.c 	struct rb_node *node, *next;
node              387 tools/perf/util/metricgroup.c 	for (node = rb_first_cached(&groups.entries); node; node = next) {
node              388 tools/perf/util/metricgroup.c 		struct mep *me = container_of(node, struct mep, nd);
node              394 tools/perf/util/metricgroup.c 		next = rb_next(node);
node              395 tools/perf/util/metricgroup.c 		rblist__remove_node(&groups, node);
node              330 tools/perf/util/mmap.c static void build_node_mask(int node, cpu_set_t *mask)
node              342 tools/perf/util/mmap.c 		if (cpu__get_node(cpu) == node)
node              348 tools/perf/util/parse-events.c 	list_add_tail(&evsel->core.node, list);
node              531 tools/perf/util/parse-events.c 	list_add_tail(&evsel->core.node, list);
node              665 tools/perf/util/parse-events.c 		list_for_each_entry_safe(evsel, tmp, &new_evsels, core.node) {
node              666 tools/perf/util/parse-events.c 			list_del_init(&evsel->core.node);
node              673 tools/perf/util/parse-events.c 	list_for_each_entry(pos, &new_evsels, core.node) {
node             1478 tools/perf/util/parse-events.c 	leader = list_first_entry(list, struct evsel, core.node);
node             1479 tools/perf/util/parse-events.c 	evsel = list_last_entry(list, struct evsel, core.node);
node             1575 tools/perf/util/parse-events.c 	leader = list_entry(list->next, struct evsel, core.node);
node             2075 tools/perf/util/parse-events.c 		if (last->core.node.prev == &evlist->core.entries)
node             2077 tools/perf/util/parse-events.c 		last = list_entry(last->core.node.prev, struct evsel, core.node);
node             2392 tools/perf/util/parse-events.c 		list_for_each_entry(ent, &pcache->entries, node) {
node             1047 tools/perf/util/probe-event.c 	struct str_node *node;
node             1089 tools/perf/util/probe-event.c 			strlist__for_each_entry(node, vl->vars) {
node             1090 tools/perf/util/probe-event.c 				var = strchr(node->s, '\t') + 1;
node             1092 tools/perf/util/probe-event.c 					fprintf(stdout, "\t\t%s\n", node->s);
node             2345 tools/perf/util/probe-event.c 	struct kprobe_blacklist_node *node;
node             2348 tools/perf/util/probe-event.c 		node = list_first_entry(blacklist,
node             2350 tools/perf/util/probe-event.c 		list_del_init(&node->list);
node             2351 tools/perf/util/probe-event.c 		zfree(&node->symbol);
node             2352 tools/perf/util/probe-event.c 		free(node);
node             2358 tools/perf/util/probe-event.c 	struct kprobe_blacklist_node *node;
node             2377 tools/perf/util/probe-event.c 		node = zalloc(sizeof(*node));
node             2378 tools/perf/util/probe-event.c 		if (!node) {
node             2382 tools/perf/util/probe-event.c 		INIT_LIST_HEAD(&node->list);
node             2383 tools/perf/util/probe-event.c 		list_add_tail(&node->list, blacklist);
node             2384 tools/perf/util/probe-event.c 		if (sscanf(buf, "0x%lx-0x%lx", &node->start, &node->end) != 2) {
node             2395 tools/perf/util/probe-event.c 		node->symbol = strdup(p);
node             2396 tools/perf/util/probe-event.c 		if (!node->symbol) {
node             2401 tools/perf/util/probe-event.c 			  node->start, node->end, node->symbol);
node             2415 tools/perf/util/probe-event.c 	struct kprobe_blacklist_node *node;
node             2417 tools/perf/util/probe-event.c 	list_for_each_entry(node, blacklist, list) {
node             2418 tools/perf/util/probe-event.c 		if (node->start <= address && address < node->end)
node             2419 tools/perf/util/probe-event.c 			return node;
node             3262 tools/perf/util/probe-event.c 	struct str_node *node;
node             3298 tools/perf/util/probe-event.c 	strlist__for_each_entry(node, entry->tevlist) {
node             3300 tools/perf/util/probe-event.c 		ret = parse_probe_trace_command(node->s, tev);
node              352 tools/perf/util/probe-file.c 		BUG_ON(!list_empty(&entry->node));
node              367 tools/perf/util/probe-file.c 		INIT_LIST_HEAD(&entry->node);
node              388 tools/perf/util/probe-file.c 	struct str_node *node;
node              400 tools/perf/util/probe-file.c 	strlist__for_each_entry(node, entry->tevlist) {
node              402 tools/perf/util/probe-file.c 		ret = parse_probe_trace_command(node->s, tev);
node              511 tools/perf/util/probe-file.c 			list_add_tail(&entry->node, &pcache->entries);
node              540 tools/perf/util/probe-file.c 	list_for_each_entry_safe(entry, n, &pcache->entries, node) {
node              541 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
node              662 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
node              681 tools/perf/util/probe-file.c 	list_add_tail(&entry->node, &pcache->entries);
node              851 tools/perf/util/probe-file.c 			list_add_tail(&entry->node, &pcache->entries);
node              864 tools/perf/util/probe-file.c 		list_del_init(&entry->node);
node              953 tools/perf/util/probe-file.c 	list_for_each_entry_safe(entry, tmp, &pcache->entries, node) {
node              956 tools/perf/util/probe-file.c 			list_del_init(&entry->node);
node               12 tools/perf/util/probe-file.h 	struct list_head	node;
node               36 tools/perf/util/probe-file.h 	list_for_each_entry(entry, &pcache->entries, node)
node              131 tools/perf/util/rblist.c 	struct rb_node *node;
node              133 tools/perf/util/rblist.c 	for (node = rb_first_cached(&rblist->entries); node;
node              134 tools/perf/util/rblist.c 	     node = rb_next(node)) {
node              136 tools/perf/util/rblist.c 			return node;
node              283 tools/perf/util/scripting-engines/trace-event-perl.c 		struct callchain_cursor_node *node;
node              284 tools/perf/util/scripting-engines/trace-event-perl.c 		node = callchain_cursor_current(&callchain_cursor);
node              285 tools/perf/util/scripting-engines/trace-event-perl.c 		if (!node)
node              292 tools/perf/util/scripting-engines/trace-event-perl.c 		if (!hv_stores(elem, "ip", newSVuv(node->ip))) {
node              297 tools/perf/util/scripting-engines/trace-event-perl.c 		if (node->sym) {
node              303 tools/perf/util/scripting-engines/trace-event-perl.c 			if (!hv_stores(sym, "start",   newSVuv(node->sym->start)) ||
node              304 tools/perf/util/scripting-engines/trace-event-perl.c 			    !hv_stores(sym, "end",     newSVuv(node->sym->end)) ||
node              305 tools/perf/util/scripting-engines/trace-event-perl.c 			    !hv_stores(sym, "binding", newSVuv(node->sym->binding)) ||
node              306 tools/perf/util/scripting-engines/trace-event-perl.c 			    !hv_stores(sym, "name",    newSVpvn(node->sym->name,
node              307 tools/perf/util/scripting-engines/trace-event-perl.c 								node->sym->namelen)) ||
node              315 tools/perf/util/scripting-engines/trace-event-perl.c 		if (node->map) {
node              316 tools/perf/util/scripting-engines/trace-event-perl.c 			struct map *map = node->map;
node              418 tools/perf/util/scripting-engines/trace-event-python.c 		struct callchain_cursor_node *node;
node              419 tools/perf/util/scripting-engines/trace-event-python.c 		node = callchain_cursor_current(&callchain_cursor);
node              420 tools/perf/util/scripting-engines/trace-event-python.c 		if (!node)
node              429 tools/perf/util/scripting-engines/trace-event-python.c 				PyLong_FromUnsignedLongLong(node->ip));
node              431 tools/perf/util/scripting-engines/trace-event-python.c 		if (node->sym) {
node              436 tools/perf/util/scripting-engines/trace-event-python.c 					PyLong_FromUnsignedLongLong(node->sym->start));
node              438 tools/perf/util/scripting-engines/trace-event-python.c 					PyLong_FromUnsignedLongLong(node->sym->end));
node              440 tools/perf/util/scripting-engines/trace-event-python.c 					_PyLong_FromLong(node->sym->binding));
node              442 tools/perf/util/scripting-engines/trace-event-python.c 					_PyUnicode_FromStringAndSize(node->sym->name,
node              443 tools/perf/util/scripting-engines/trace-event-python.c 							node->sym->namelen));
node              447 tools/perf/util/scripting-engines/trace-event-python.c 		if (node->map) {
node              448 tools/perf/util/scripting-engines/trace-event-python.c 			const char *dsoname = get_dsoname(node->map);
node               91 tools/perf/util/sort.h 		struct list_head node;
node              165 tools/perf/util/sort.h 	return !list_empty(&he->pairs.node);
node              171 tools/perf/util/sort.h 		return list_entry(he->pairs.node.next, struct hist_entry, pairs.node);
node              178 tools/perf/util/sort.h 	list_add_tail(&pair->pairs.node, &he->pairs.head);
node               40 tools/perf/util/srcline.c 			       struct inline_node *node)
node               52 tools/perf/util/srcline.c 		list_add_tail(&ilist->list, &node->val);
node               54 tools/perf/util/srcline.c 		list_add(&ilist->list, &node->val);
node              279 tools/perf/util/srcline.c 				       struct inline_node *node,
node              289 tools/perf/util/srcline.c 	return inline_list__append(inline_sym, srcline, node);
node              294 tools/perf/util/srcline.c 		     bool unwind_inlines, struct inline_node *node,
node              322 tools/perf/util/srcline.c 		if (node && inline_list__append_dso_a2l(dso, node, sym))
node              332 tools/perf/util/srcline.c 			if (node != NULL) {
node              333 tools/perf/util/srcline.c 				if (inline_list__append_dso_a2l(dso, node, sym))
node              367 tools/perf/util/srcline.c 	struct inline_node *node;
node              369 tools/perf/util/srcline.c 	node = zalloc(sizeof(*node));
node              370 tools/perf/util/srcline.c 	if (node == NULL) {
node              375 tools/perf/util/srcline.c 	INIT_LIST_HEAD(&node->val);
node              376 tools/perf/util/srcline.c 	node->addr = addr;
node              378 tools/perf/util/srcline.c 	addr2line(dso_name, addr, NULL, NULL, dso, true, node, sym);
node              379 tools/perf/util/srcline.c 	return node;
node              409 tools/perf/util/srcline.c 		     struct inline_node *node __maybe_unused,
node              455 tools/perf/util/srcline.c 	struct inline_node *node;
node              470 tools/perf/util/srcline.c 	node = zalloc(sizeof(*node));
node              471 tools/perf/util/srcline.c 	if (node == NULL) {
node              476 tools/perf/util/srcline.c 	INIT_LIST_HEAD(&node->val);
node              477 tools/perf/util/srcline.c 	node->addr = addr;
node              495 tools/perf/util/srcline.c 		if (inline_list__append(inline_sym, srcline, node) != 0) {
node              508 tools/perf/util/srcline.c 	return node;
node              618 tools/perf/util/srcline.c 	struct srcline_node *i, *node;
node              621 tools/perf/util/srcline.c 	node = zalloc(sizeof(struct srcline_node));
node              622 tools/perf/util/srcline.c 	if (!node) {
node              627 tools/perf/util/srcline.c 	node->addr = addr;
node              628 tools/perf/util/srcline.c 	node->srcline = srcline;
node              640 tools/perf/util/srcline.c 	rb_link_node(&node->rb_node, parent, p);
node              641 tools/perf/util/srcline.c 	rb_insert_color_cached(&node->rb_node, tree, leftmost);
node              689 tools/perf/util/srcline.c void inline_node__delete(struct inline_node *node)
node              693 tools/perf/util/srcline.c 	list_for_each_entry_safe(ilist, tmp, &node->val, list) {
node              702 tools/perf/util/srcline.c 	free(node);
node               46 tools/perf/util/srcline.h void inline_node__delete(struct inline_node *node);
node              552 tools/perf/util/stat-display.c 	alias = list_prepare_entry(counter, &(evlist->core.entries), core.node);
node              553 tools/perf/util/stat-display.c 	list_for_each_entry_continue (alias, &evlist->core.entries, core.node) {
node               19 tools/perf/util/strfilter.c static void strfilter_node__delete(struct strfilter_node *node)
node               21 tools/perf/util/strfilter.c 	if (node) {
node               22 tools/perf/util/strfilter.c 		if (node->p && !is_operator(*node->p))
node               23 tools/perf/util/strfilter.c 			zfree((char **)&node->p);
node               24 tools/perf/util/strfilter.c 		strfilter_node__delete(node->l);
node               25 tools/perf/util/strfilter.c 		strfilter_node__delete(node->r);
node               26 tools/perf/util/strfilter.c 		free(node);
node               70 tools/perf/util/strfilter.c 	struct strfilter_node *node = zalloc(sizeof(*node));
node               72 tools/perf/util/strfilter.c 	if (node) {
node               73 tools/perf/util/strfilter.c 		node->p = op;
node               74 tools/perf/util/strfilter.c 		node->l = l;
node               75 tools/perf/util/strfilter.c 		node->r = r;
node               78 tools/perf/util/strfilter.c 	return node;
node              218 tools/perf/util/strfilter.c static bool strfilter_node__compare(struct strfilter_node *node,
node              221 tools/perf/util/strfilter.c 	if (!node || !node->p)
node              224 tools/perf/util/strfilter.c 	switch (*node->p) {
node              226 tools/perf/util/strfilter.c 		return strfilter_node__compare(node->l, str) ||
node              227 tools/perf/util/strfilter.c 			strfilter_node__compare(node->r, str);
node              229 tools/perf/util/strfilter.c 		return strfilter_node__compare(node->l, str) &&
node              230 tools/perf/util/strfilter.c 			strfilter_node__compare(node->r, str);
node              232 tools/perf/util/strfilter.c 		return !strfilter_node__compare(node->r, str);
node              234 tools/perf/util/strfilter.c 		return strglobmatch(str, node->p);
node              246 tools/perf/util/strfilter.c static int strfilter_node__sprint(struct strfilter_node *node, char *buf);
node              249 tools/perf/util/strfilter.c static int strfilter_node__sprint_pt(struct strfilter_node *node, char *buf)
node              252 tools/perf/util/strfilter.c 	int pt = node->r ? 2 : 0;	/* don't need to check node->l */
node              256 tools/perf/util/strfilter.c 	len = strfilter_node__sprint(node, buf);
node              264 tools/perf/util/strfilter.c static int strfilter_node__sprint(struct strfilter_node *node, char *buf)
node              268 tools/perf/util/strfilter.c 	if (!node || !node->p)
node              271 tools/perf/util/strfilter.c 	switch (*node->p) {
node              274 tools/perf/util/strfilter.c 		len = strfilter_node__sprint_pt(node->l, buf);
node              280 tools/perf/util/strfilter.c 			*(buf + len++) = *node->p;
node              284 tools/perf/util/strfilter.c 		rlen = strfilter_node__sprint_pt(node->r, buf);
node              290 tools/perf/util/strfilter.c 		len = strlen(node->p);
node              292 tools/perf/util/strfilter.c 			strcpy(buf, node->p);
node             1437 tools/perf/util/symbol-elf.c 	struct list_head node;
node             1443 tools/perf/util/symbol-elf.c 	struct list_head node;
node             1459 tools/perf/util/symbol-elf.c 	list_for_each_entry((p), &(k)->phdrs, node)
node             1481 tools/perf/util/symbol-elf.c 		list_add_tail(&p->node, &kci->phdrs);
node             1490 tools/perf/util/symbol-elf.c 	list_for_each_entry_safe(p, tmp, &kci->phdrs, node) {
node             1491 tools/perf/util/symbol-elf.c 		list_del_init(&p->node);
node             1503 tools/perf/util/symbol-elf.c 		list_add_tail(&s->node, &kci->syms);
node             1513 tools/perf/util/symbol-elf.c 	list_for_each_entry_safe(s, tmp, &kci->syms, node) {
node             1514 tools/perf/util/symbol-elf.c 		list_del_init(&s->node);
node             1628 tools/perf/util/symbol-elf.c 	list_for_each_entry(sdat, &kci->syms, node) {
node             1178 tools/perf/util/symbol.c 	list_add(&map->node, &md->maps);
node             1222 tools/perf/util/symbol.c 				list_add_tail(&m->node, &merged);
node             1249 tools/perf/util/symbol.c 		old_map = list_entry(merged.next, struct map, node);
node             1250 tools/perf/util/symbol.c 		list_del_init(&old_map->node);
node             1331 tools/perf/util/symbol.c 		list_for_each_entry(new_map, &md.maps, node) {
node             1340 tools/perf/util/symbol.c 		replacement_map = list_entry(md.maps.next, struct map, node);
node             1344 tools/perf/util/symbol.c 		new_map = list_entry(md.maps.next, struct map, node);
node             1345 tools/perf/util/symbol.c 		list_del_init(&new_map->node);
node             1403 tools/perf/util/symbol.c 		map = list_entry(md.maps.next, struct map, node);
node             1404 tools/perf/util/symbol.c 		list_del_init(&map->node);
node             1791 tools/perf/util/symbol.c 	struct rb_node *node;
node             1795 tools/perf/util/symbol.c 	for (node = maps->names.rb_node; node; ) {
node             1798 tools/perf/util/symbol.c 		map = rb_entry(node, struct map, rb_node_name);
node             1802 tools/perf/util/symbol.c 			node = node->rb_left;
node             1804 tools/perf/util/symbol.c 			node = node->rb_right;
node              146 tools/perf/util/thread.c 		if (!list_empty(&thread->node))
node              147 tools/perf/util/thread.c 			list_del_init(&thread->node);
node               26 tools/perf/util/thread.h 		struct list_head node;
node              410 tools/perf/util/trace-event-info.c 	list_for_each_entry(pos, pattrs, core.node) {
node              445 tools/perf/util/trace-event-info.c 	list_for_each_entry(pos, pattrs, core.node)
node              349 tools/perf/util/util.c 	struct str_node *node;
node              364 tools/perf/util/util.c 	node = strlist__entry(tips, random() % strlist__nr_entries(tips));
node              365 tools/perf/util/util.c 	if (asprintf(&tip, "Tip: %s", node->s) < 0)
node             2573 tools/power/x86/turbostat/turbostat.c 	int pkg, node, lnode, cpu, cpux;
node             2590 tools/power/x86/turbostat/turbostat.c 			node = cpus[cpu].physical_node_id;
node             2598 tools/power/x86/turbostat/turbostat.c 				   (cpus[cpux].physical_node_id == node)) {
node              122 tools/testing/nvdimm/test/nfit.c #define NFIT_DIMM_HANDLE(node, socket, imc, chan, dimm) \
node              123 tools/testing/nvdimm/test/nfit.c 	(((node & 0xfff) << 16) | ((socket & 0xf) << 12) \
node               30 tools/testing/radix-tree/linux.c 	struct radix_tree_node *node;
node               38 tools/testing/radix-tree/linux.c 		node = cachep->objs;
node               39 tools/testing/radix-tree/linux.c 		cachep->objs = node->parent;
node               41 tools/testing/radix-tree/linux.c 		node->parent = NULL;
node               44 tools/testing/radix-tree/linux.c 		node = malloc(cachep->size);
node               46 tools/testing/radix-tree/linux.c 			cachep->ctor(node);
node               51 tools/testing/radix-tree/linux.c 		printf("Allocating %p from slab\n", node);
node               52 tools/testing/radix-tree/linux.c 	return node;
node               66 tools/testing/radix-tree/linux.c 		struct radix_tree_node *node = objp;
node               68 tools/testing/radix-tree/linux.c 		node->parent = cachep->objs;
node               69 tools/testing/radix-tree/linux.c 		cachep->objs = node;
node              249 tools/testing/radix-tree/test.c 	struct radix_tree_node *node = root->xa_head;
node              250 tools/testing/radix-tree/test.c 	if (!radix_tree_is_internal_node(node))
node              252 tools/testing/radix-tree/test.c 	verify_node(node, tag, !!root_tag_get(root, tag));
node              273 tools/testing/radix-tree/test.c 	struct radix_tree_node *node = root->xa_head;
node              274 tools/testing/radix-tree/test.c 	if (!radix_tree_is_internal_node(node)) {
node              279 tools/testing/radix-tree/test.c 	node = entry_to_node(node);
node              280 tools/testing/radix-tree/test.c 	assert(maxindex <= node_maxindex(node));
node              282 tools/testing/radix-tree/test.c 	shift = node->shift;
node               45 tools/testing/selftests/bpf/test_lpm_map.c 	struct tlpm_node *node;
node               51 tools/testing/selftests/bpf/test_lpm_map.c 	node = tlpm_match(list, key, n_bits);
node               52 tools/testing/selftests/bpf/test_lpm_map.c 	if (node && node->n_bits == n_bits) {
node               53 tools/testing/selftests/bpf/test_lpm_map.c 		memcpy(node->key, key, n);
node               59 tools/testing/selftests/bpf/test_lpm_map.c 	node = malloc(sizeof(*node) + n);
node               60 tools/testing/selftests/bpf/test_lpm_map.c 	assert(node);
node               62 tools/testing/selftests/bpf/test_lpm_map.c 	node->next = list;
node               63 tools/testing/selftests/bpf/test_lpm_map.c 	node->n_bits = n_bits;
node               64 tools/testing/selftests/bpf/test_lpm_map.c 	memcpy(node->key, key, n);
node               66 tools/testing/selftests/bpf/test_lpm_map.c 	return node;
node               71 tools/testing/selftests/bpf/test_lpm_map.c 	struct tlpm_node *node;
node               75 tools/testing/selftests/bpf/test_lpm_map.c 	while ((node = list)) {
node               77 tools/testing/selftests/bpf/test_lpm_map.c 		free(node);
node              115 tools/testing/selftests/bpf/test_lpm_map.c 	struct tlpm_node *node;
node              121 tools/testing/selftests/bpf/test_lpm_map.c 		node = best->next;
node              123 tools/testing/selftests/bpf/test_lpm_map.c 		return node;
node              126 tools/testing/selftests/bpf/test_lpm_map.c 	for (node = list; node; node = node->next) {
node              127 tools/testing/selftests/bpf/test_lpm_map.c 		if (node->next == best) {
node              128 tools/testing/selftests/bpf/test_lpm_map.c 			node->next = best->next;
node              169 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *parent;
node              170 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *left;
node              171 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *right;
node              183 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *root;
node              197 tools/testing/selftests/kvm/lib/sparsebit.c static sparsebit_num_t node_num_set(struct node *nodep)
node              205 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_first(struct sparsebit *s)
node              207 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node              219 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_next(struct sparsebit *s, struct node *np)
node              221 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep = np;
node              247 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_prev(struct sparsebit *s, struct node *np)
node              249 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep = np;
node              258 tools/testing/selftests/kvm/lib/sparsebit.c 		return (struct node *) nodep;
node              268 tools/testing/selftests/kvm/lib/sparsebit.c 	return (struct node *) nodep->parent;
node              276 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_copy_subtree(struct node *subtree)
node              278 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *root;
node              310 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_find(struct sparsebit *s, sparsebit_idx_t idx)
node              312 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node              333 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_add(struct sparsebit *s, sparsebit_idx_t idx)
node              335 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *parentp, *prev;
node              409 tools/testing/selftests/kvm/lib/sparsebit.c static void node_rm(struct sparsebit *s, struct node *nodep)
node              411 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *tmp;
node              498 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_split(struct sparsebit *s, sparsebit_idx_t idx)
node              500 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep1, *nodep2;
node              599 tools/testing/selftests/kvm/lib/sparsebit.c static void node_reduce(struct sparsebit *s, struct node *nodep)
node              605 tools/testing/selftests/kvm/lib/sparsebit.c 		struct node *prev, *next, *tmp;
node              781 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node              807 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node              834 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node              872 tools/testing/selftests/kvm/lib/sparsebit.c static void dump_nodes(FILE *stream, struct node *nodep,
node              901 tools/testing/selftests/kvm/lib/sparsebit.c static inline sparsebit_idx_t node_first_set(struct node *nodep, int start)
node              909 tools/testing/selftests/kvm/lib/sparsebit.c static inline sparsebit_idx_t node_first_clear(struct node *nodep, int start)
node             1089 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node             1103 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep1, *nodep2;
node             1160 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node             1170 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *candidate = NULL;
node             1253 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep1, *nodep2;
node             1374 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *next;
node             1456 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *next;
node             1592 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep;
node             1688 tools/testing/selftests/kvm/lib/sparsebit.c 	struct node *nodep, *prev = NULL;
node              197 tools/testing/selftests/net/reuseport_bpf_numa.c 	int epfd, node;
node              205 tools/testing/selftests/net/reuseport_bpf_numa.c 	for (node = 0; node < len; ++node) {
node              207 tools/testing/selftests/net/reuseport_bpf_numa.c 		ev.data.fd = rcv_fd[node];
node              208 tools/testing/selftests/net/reuseport_bpf_numa.c 		if (epoll_ctl(epfd, EPOLL_CTL_ADD, rcv_fd[node], &ev))
node              213 tools/testing/selftests/net/reuseport_bpf_numa.c 	for (node = 0; node < len; ++node) {
node              214 tools/testing/selftests/net/reuseport_bpf_numa.c 		send_from_node(node, family, proto);
node              215 tools/testing/selftests/net/reuseport_bpf_numa.c 		receive_on_node(rcv_fd, len, epfd, node, proto);
node              219 tools/testing/selftests/net/reuseport_bpf_numa.c 	for (node = len - 1; node >= 0; --node) {
node              220 tools/testing/selftests/net/reuseport_bpf_numa.c 		send_from_node(node, family, proto);
node              221 tools/testing/selftests/net/reuseport_bpf_numa.c 		receive_on_node(rcv_fd, len, epfd, node, proto);
node              225 tools/testing/selftests/net/reuseport_bpf_numa.c 	for (node = 0; node < len; ++node)
node              226 tools/testing/selftests/net/reuseport_bpf_numa.c 		close(rcv_fd[node]);
node              136 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			struct percpu_list_node *node,
node              148 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		newval = (intptr_t)node;
node              150 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		node->next = (struct percpu_list_node *)expect;
node              198 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	struct percpu_list_node *node;
node              200 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	node = list->c[cpu].head;
node              201 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	if (!node)
node              203 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	list->c[cpu].head = node->next;
node              204 tools/testing/selftests/rseq/basic_percpu_ops_test.c 	return node;
node              219 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		struct percpu_list_node *node;
node              221 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		node = this_cpu_list_pop(list, NULL);
node              223 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		if (node)
node              224 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			this_cpu_list_push(list, node, NULL);
node              253 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			struct percpu_list_node *node;
node              257 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			node = malloc(sizeof(*node));
node              258 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			assert(node);
node              259 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			node->data = j;
node              260 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			node->next = list.c[i].head;
node              261 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			list.c[i].head = node;
node              273 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		struct percpu_list_node *node;
node              278 tools/testing/selftests/rseq/basic_percpu_ops_test.c 		while ((node = __percpu_list_pop(&list, i))) {
node              279 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			sum += node->data;
node              280 tools/testing/selftests/rseq/basic_percpu_ops_test.c 			free(node);
node              514 tools/testing/selftests/rseq/param_test.c 			struct percpu_list_node *node,
node              526 tools/testing/selftests/rseq/param_test.c 		newval = (intptr_t)node;
node              528 tools/testing/selftests/rseq/param_test.c 		node->next = (struct percpu_list_node *)expect;
node              546 tools/testing/selftests/rseq/param_test.c 	struct percpu_list_node *node = NULL;
node              563 tools/testing/selftests/rseq/param_test.c 			node = head;
node              572 tools/testing/selftests/rseq/param_test.c 	return node;
node              581 tools/testing/selftests/rseq/param_test.c 	struct percpu_list_node *node;
node              583 tools/testing/selftests/rseq/param_test.c 	node = list->c[cpu].head;
node              584 tools/testing/selftests/rseq/param_test.c 	if (!node)
node              586 tools/testing/selftests/rseq/param_test.c 	list->c[cpu].head = node->next;
node              587 tools/testing/selftests/rseq/param_test.c 	return node;
node              600 tools/testing/selftests/rseq/param_test.c 		struct percpu_list_node *node;
node              602 tools/testing/selftests/rseq/param_test.c 		node = this_cpu_list_pop(list, NULL);
node              605 tools/testing/selftests/rseq/param_test.c 		if (node)
node              606 tools/testing/selftests/rseq/param_test.c 			this_cpu_list_push(list, node, NULL);
node              635 tools/testing/selftests/rseq/param_test.c 			struct percpu_list_node *node;
node              639 tools/testing/selftests/rseq/param_test.c 			node = malloc(sizeof(*node));
node              640 tools/testing/selftests/rseq/param_test.c 			assert(node);
node              641 tools/testing/selftests/rseq/param_test.c 			node->data = j;
node              642 tools/testing/selftests/rseq/param_test.c 			node->next = list.c[i].head;
node              643 tools/testing/selftests/rseq/param_test.c 			list.c[i].head = node;
node              667 tools/testing/selftests/rseq/param_test.c 		struct percpu_list_node *node;
node              672 tools/testing/selftests/rseq/param_test.c 		while ((node = __percpu_list_pop(&list, i))) {
node              673 tools/testing/selftests/rseq/param_test.c 			sum += node->data;
node              674 tools/testing/selftests/rseq/param_test.c 			free(node);
node              687 tools/testing/selftests/rseq/param_test.c 			  struct percpu_buffer_node *node,
node              703 tools/testing/selftests/rseq/param_test.c 		newval_spec = (intptr_t)node;
node              787 tools/testing/selftests/rseq/param_test.c 		struct percpu_buffer_node *node;
node              789 tools/testing/selftests/rseq/param_test.c 		node = this_cpu_buffer_pop(buffer, NULL);
node              792 tools/testing/selftests/rseq/param_test.c 		if (node) {
node              793 tools/testing/selftests/rseq/param_test.c 			if (!this_cpu_buffer_push(buffer, node, NULL)) {
node              832 tools/testing/selftests/rseq/param_test.c 			struct percpu_buffer_node *node;
node              843 tools/testing/selftests/rseq/param_test.c 			node = malloc(sizeof(*node));
node              844 tools/testing/selftests/rseq/param_test.c 			assert(node);
node              845 tools/testing/selftests/rseq/param_test.c 			node->data = j;
node              846 tools/testing/selftests/rseq/param_test.c 			buffer.c[i].array[j - 1] = node;
node              871 tools/testing/selftests/rseq/param_test.c 		struct percpu_buffer_node *node;
node              876 tools/testing/selftests/rseq/param_test.c 		while ((node = __percpu_buffer_pop(&buffer, i))) {
node              877 tools/testing/selftests/rseq/param_test.c 			sum += node->data;
node              878 tools/testing/selftests/rseq/param_test.c 			free(node);
node              118 tools/testing/selftests/x86/test_vdso.c static long sys_getcpu(unsigned * cpu, unsigned * node,
node              121 tools/testing/selftests/x86/test_vdso.c 	return syscall(__NR_getcpu, cpu, node, cache);
node              148 tools/testing/selftests/x86/test_vdso.c 		unsigned node;
node              157 tools/testing/selftests/x86/test_vdso.c 			node = node_sys;
node              159 tools/testing/selftests/x86/test_vdso.c 			node = node_vdso;
node              161 tools/testing/selftests/x86/test_vdso.c 			node = node_vsys;
node              164 tools/testing/selftests/x86/test_vdso.c 		if (!ret_sys && (cpu_sys != cpu || node_sys != node))
node              166 tools/testing/selftests/x86/test_vdso.c 		if (!ret_vdso && (cpu_vdso != cpu || node_vdso != node))
node              168 tools/testing/selftests/x86/test_vdso.c 		if (!ret_vsys && (cpu_vsys != cpu || node_vsys != node))
node              174 tools/testing/selftests/x86/test_vsyscall.c static inline long sys_getcpu(unsigned * cpu, unsigned * node,
node              177 tools/testing/selftests/x86/test_vsyscall.c 	return syscall(SYS_getcpu, cpu, node, cache);
node              327 tools/testing/selftests/x86/test_vsyscall.c 	unsigned node = 0;
node              342 tools/testing/selftests/x86/test_vsyscall.c 		node = node_sys;
node              352 tools/testing/selftests/x86/test_vsyscall.c 				node = node_vdso;
node              362 tools/testing/selftests/x86/test_vsyscall.c 			if (node_vdso != node) {
node              363 tools/testing/selftests/x86/test_vsyscall.c 				printf("[FAIL]\tvDSO reported node %hu but should be %hu\n", node_vdso, node);
node              378 tools/testing/selftests/x86/test_vsyscall.c 				node = node_vsys;
node              388 tools/testing/selftests/x86/test_vsyscall.c 			if (node_vsys != node) {
node              389 tools/testing/selftests/x86/test_vsyscall.c 				printf("[FAIL]\tvsyscall reported node %hu but should be %hu\n", node_vsys, node);
node              151 tools/testing/vsock/vsock_diag_test.c 	struct list_head *node;
node              154 tools/testing/vsock/vsock_diag_test.c 	list_for_each(node, head)
node              147 tools/usb/usbip/libsrc/usbip_host_common.c 			list_add(&edev->node, &hdriver->edev_list);
node              161 tools/usb/usbip/libsrc/usbip_host_common.c 		edev = list_entry(i, struct usbip_exported_device, node);
node              275 tools/usb/usbip/libsrc/usbip_host_common.c 		edev = list_entry(i, struct usbip_exported_device, node);
node               50 tools/usb/usbip/libsrc/usbip_host_common.h 	struct list_head node;
node              111 tools/usb/usbip/src/usbipd.c 		edev = list_entry(i, struct usbip_exported_device, node);
node              178 tools/usb/usbip/src/usbipd.c 		edev = list_entry(j, struct usbip_exported_device, node);
node              198 tools/usb/usbip/src/usbipd.c 		edev = list_entry(j, struct usbip_exported_device, node);
node              276 tools/vm/slabinfo.c 	int node;
node              286 tools/vm/slabinfo.c 		node = strtoul(t, &t, 10);
node              290 tools/vm/slabinfo.c 			numa[node] = nr;
node              291 tools/vm/slabinfo.c 			if (node > highest_node)
node              292 tools/vm/slabinfo.c 				highest_node = node;
node              365 tools/vm/slabinfo.c 	int node;
node              380 tools/vm/slabinfo.c 		for(node = 0; node <= highest_node; node++)
node              381 tools/vm/slabinfo.c 			printf(" %4d", node);
node              383 tools/vm/slabinfo.c 		for(node = 0; node <= highest_node; node++)
node              388 tools/vm/slabinfo.c 	for(node = 0; node <= highest_node; node++) {
node              391 tools/vm/slabinfo.c 		store_size(b, s->numa[node]);
node              397 tools/vm/slabinfo.c 		for(node = 0; node <= highest_node; node++) {
node              400 tools/vm/slabinfo.c 			store_size(b, s->numa_partial[node]);
node               25 virt/kvm/vfio.c 	struct list_head node;
node              165 virt/kvm/vfio.c 	list_for_each_entry(kvg, &kv->group_list, node) {
node              211 virt/kvm/vfio.c 		list_for_each_entry(kvg, &kv->group_list, node) {
node              226 virt/kvm/vfio.c 		list_add_tail(&kvg->node, &kv->group_list);
node              251 virt/kvm/vfio.c 		list_for_each_entry(kvg, &kv->group_list, node) {
node              256 virt/kvm/vfio.c 			list_del(&kvg->node);
node              310 virt/kvm/vfio.c 		list_for_each_entry(kvg, &kv->group_list, node) {
node              368 virt/kvm/vfio.c 	list_for_each_entry_safe(kvg, tmp, &kv->group_list, node) {
node              374 virt/kvm/vfio.c 		list_del(&kvg->node);
node               99 virt/lib/irqbypass.c 	list_for_each_entry(tmp, &producers, node) {
node              107 virt/lib/irqbypass.c 	list_for_each_entry(consumer, &consumers, node) {
node              119 virt/lib/irqbypass.c 	list_add(&producer->node, &producers);
node              149 virt/lib/irqbypass.c 	list_for_each_entry(tmp, &producers, node) {
node              153 virt/lib/irqbypass.c 		list_for_each_entry(consumer, &consumers, node) {
node              160 virt/lib/irqbypass.c 		list_del(&producer->node);
node              194 virt/lib/irqbypass.c 	list_for_each_entry(tmp, &consumers, node) {
node              202 virt/lib/irqbypass.c 	list_for_each_entry(producer, &producers, node) {
node              214 virt/lib/irqbypass.c 	list_add(&consumer->node, &consumers);
node              244 virt/lib/irqbypass.c 	list_for_each_entry(tmp, &consumers, node) {
node              248 virt/lib/irqbypass.c 		list_for_each_entry(producer, &producers, node) {
node              255 virt/lib/irqbypass.c 		list_del(&consumer->node);