new_wb            335 fs/fs-writeback.c 	struct bdi_writeback	*new_wb;
new_wb            359 fs/fs-writeback.c 	struct bdi_writeback *new_wb = isw->new_wb;
new_wb            380 fs/fs-writeback.c 	if (old_wb < new_wb) {
new_wb            382 fs/fs-writeback.c 		spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING);
new_wb            384 fs/fs-writeback.c 		spin_lock(&new_wb->list_lock);
new_wb            397 fs/fs-writeback.c 	trace_inode_switch_wbs(inode, old_wb, new_wb);
new_wb            407 fs/fs-writeback.c 			inc_wb_stat(new_wb, WB_RECLAIMABLE);
new_wb            415 fs/fs-writeback.c 		inc_wb_stat(new_wb, WB_WRITEBACK);
new_wb            418 fs/fs-writeback.c 	wb_get(new_wb);
new_wb            430 fs/fs-writeback.c 		inode->i_wb = new_wb;
new_wb            431 fs/fs-writeback.c 		list_for_each_entry(pos, &new_wb->b_dirty, i_io_list)
new_wb            435 fs/fs-writeback.c 		inode_io_list_move_locked(inode, new_wb, pos->i_io_list.prev);
new_wb            437 fs/fs-writeback.c 		inode->i_wb = new_wb;
new_wb            454 fs/fs-writeback.c 	spin_unlock(&new_wb->list_lock);
new_wb            460 fs/fs-writeback.c 		wb_wakeup(new_wb);
new_wb            463 fs/fs-writeback.c 	wb_put(new_wb);
new_wb            511 fs/fs-writeback.c 		isw->new_wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC);
new_wb            513 fs/fs-writeback.c 	if (!isw->new_wb)
new_wb            520 fs/fs-writeback.c 	    inode_to_wb(inode) == isw->new_wb) {
new_wb            542 fs/fs-writeback.c 	if (isw->new_wb)
new_wb            543 fs/fs-writeback.c 		wb_put(isw->new_wb);
new_wb            212 include/trace/events/writeback.h 		 struct bdi_writeback *new_wb),
new_wb            214 include/trace/events/writeback.h 	TP_ARGS(inode, old_wb, new_wb),
new_wb            227 include/trace/events/writeback.h 		__entry->new_cgroup_ino	= __trace_wb_assign_cgroup(new_wb);