curr_node         204 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	struct hinic_api_cmd_cell *cell = chain->curr_node;
curr_node         248 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	struct hinic_api_cmd_cell *curr_node = chain->curr_node;
curr_node         251 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	prepare_cell_ctrl(&curr_node->ctrl, data_size);
curr_node         378 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	chain->curr_node = ctxt->cell_vaddr;
curr_node         778 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c 	chain->curr_node = chain->head_node;
curr_node         188 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.h 	struct hinic_api_cmd_cell       *curr_node;
curr_node         342 drivers/net/ethernet/mellanox/mlx4/alloc.c 	struct mlx4_zone_entry *curr_node;
curr_node         353 drivers/net/ethernet/mellanox/mlx4/alloc.c 	list_for_each_entry(curr_node, &zone_alloc->prios, prio_list) {
curr_node         354 drivers/net/ethernet/mellanox/mlx4/alloc.c 		if (unlikely(curr_node->priority == zone->priority))
curr_node         359 drivers/net/ethernet/mellanox/mlx4/alloc.c 		struct mlx4_zone_entry *it = curr_node;
curr_node         373 drivers/net/ethernet/mellanox/mlx4/alloc.c 		struct mlx4_zone_entry *it = curr_node;
curr_node         379 drivers/net/ethernet/mellanox/mlx4/alloc.c 			if (unlikely(it->priority != curr_node->priority))
curr_node         393 drivers/net/ethernet/mellanox/mlx4/alloc.c 		if (list_is_last(&curr_node->prio_list, &zone_alloc->prios))
curr_node         396 drivers/net/ethernet/mellanox/mlx4/alloc.c 		curr_node = list_first_entry(&curr_node->prio_list,
curr_node         397 drivers/net/ethernet/mellanox/mlx4/alloc.c 					     typeof(*curr_node),
curr_node         400 drivers/net/ethernet/mellanox/mlx4/alloc.c 		list_for_each_entry_from(curr_node, &zone_alloc->entries, list) {
curr_node         401 drivers/net/ethernet/mellanox/mlx4/alloc.c 			res = mlx4_bitmap_alloc_range(curr_node->bitmap, count,
curr_node         404 drivers/net/ethernet/mellanox/mlx4/alloc.c 				res += curr_node->offset;
curr_node         405 drivers/net/ethernet/mellanox/mlx4/alloc.c 				uid = curr_node->uid;
curr_node         781 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	struct fs_node *curr_node;
curr_node         785 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 	curr_node = &prio->node;
curr_node         787 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		ft = find_closest_ft_recursive(parent, &curr_node->list, reverse);
curr_node         788 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		curr_node = parent;
curr_node         789 drivers/net/ethernet/mellanox/mlx5/core/fs_core.c 		parent = curr_node->parent;
curr_node        1148 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *curr_node, *prev_node;
curr_node        1167 fs/btrfs/delayed-inode.c 	curr_node = btrfs_first_delayed_node(delayed_root);
curr_node        1168 fs/btrfs/delayed-inode.c 	while (curr_node && (!count || (count && nr--))) {
curr_node        1170 fs/btrfs/delayed-inode.c 							 curr_node);
curr_node        1172 fs/btrfs/delayed-inode.c 			btrfs_release_delayed_node(curr_node);
curr_node        1173 fs/btrfs/delayed-inode.c 			curr_node = NULL;
curr_node        1178 fs/btrfs/delayed-inode.c 		prev_node = curr_node;
curr_node        1179 fs/btrfs/delayed-inode.c 		curr_node = btrfs_next_delayed_node(curr_node);
curr_node        1183 fs/btrfs/delayed-inode.c 	if (curr_node)
curr_node        1184 fs/btrfs/delayed-inode.c 		btrfs_release_delayed_node(curr_node);
curr_node        1986 fs/btrfs/delayed-inode.c 	struct btrfs_delayed_node *curr_node, *prev_node;
curr_node        1988 fs/btrfs/delayed-inode.c 	curr_node = btrfs_first_delayed_node(fs_info->delayed_root);
curr_node        1989 fs/btrfs/delayed-inode.c 	while (curr_node) {
curr_node        1990 fs/btrfs/delayed-inode.c 		__btrfs_kill_delayed_node(curr_node);
curr_node        1992 fs/btrfs/delayed-inode.c 		prev_node = curr_node;
curr_node        1993 fs/btrfs/delayed-inode.c 		curr_node = btrfs_next_delayed_node(curr_node);
curr_node         554 fs/ext4/dir.c  		info->curr_node = NULL;
curr_node         569 fs/ext4/dir.c  	} else if (!info->curr_node)
curr_node         570 fs/ext4/dir.c  		info->curr_node = rb_first(&info->root);
curr_node         578 fs/ext4/dir.c  		if ((!info->curr_node) ||
curr_node         580 fs/ext4/dir.c  			info->curr_node = NULL;
curr_node         592 fs/ext4/dir.c  			info->curr_node = rb_first(&info->root);
curr_node         595 fs/ext4/dir.c  		fname = rb_entry(info->curr_node, struct fname, rb_hash);
curr_node         601 fs/ext4/dir.c  		info->curr_node = rb_next(info->curr_node);
curr_node         602 fs/ext4/dir.c  		if (info->curr_node) {
curr_node         603 fs/ext4/dir.c  			fname = rb_entry(info->curr_node, struct fname,
curr_node        2175 fs/ext4/ext4.h 	struct rb_node	*curr_node;