Lines Matching refs:sbi

60 	struct f2fs_sb_info *sbi = bio->bi_private;  in f2fs_write_end_io()  local
72 f2fs_stop_checkpoint(sbi); in f2fs_write_end_io()
75 dec_page_count(sbi, F2FS_WRITEBACK); in f2fs_write_end_io()
78 if (!get_pages(sbi, F2FS_WRITEBACK) && in f2fs_write_end_io()
79 !list_empty(&sbi->cp_wait.task_list)) in f2fs_write_end_io()
80 wake_up(&sbi->cp_wait); in f2fs_write_end_io()
88 static struct bio *__bio_alloc(struct f2fs_sb_info *sbi, block_t blk_addr, in __bio_alloc() argument
95 bio->bi_bdev = sbi->sb->s_bdev; in __bio_alloc()
98 bio->bi_private = is_read ? NULL : sbi; in __bio_alloc()
111 trace_f2fs_submit_read_bio(io->sbi->sb, fio, io->bio); in __submit_merged_bio()
113 trace_f2fs_submit_write_bio(io->sbi->sb, fio, io->bio); in __submit_merged_bio()
119 void f2fs_submit_merged_bio(struct f2fs_sb_info *sbi, in f2fs_submit_merged_bio() argument
125 io = is_read_io(rw) ? &sbi->read_io : &sbi->write_io[btype]; in f2fs_submit_merged_bio()
132 if (test_opt(sbi, NOBARRIER)) in f2fs_submit_merged_bio()
154 bio = __bio_alloc(fio->sbi, fio->blk_addr, 1, is_read_io(fio->rw)); in f2fs_submit_page_bio()
167 struct f2fs_sb_info *sbi = fio->sbi; in f2fs_submit_page_mbio() local
173 io = is_read ? &sbi->read_io : &sbi->write_io[btype]; in f2fs_submit_page_mbio()
175 verify_block_addr(sbi, fio->blk_addr); in f2fs_submit_page_mbio()
180 inc_page_count(sbi, F2FS_WRITEBACK); in f2fs_submit_page_mbio()
187 int bio_blocks = MAX_BIO_BLOCKS(sbi); in f2fs_submit_page_mbio()
189 io->bio = __bio_alloc(sbi, fio->blk_addr, bio_blocks, is_read); in f2fs_submit_page_mbio()
233 struct f2fs_sb_info *sbi = F2FS_I_SB(dn->inode); in reserve_new_block() local
237 if (unlikely(!inc_valid_block_count(sbi, dn->inode, 1))) in reserve_new_block()
287 .sbi = F2FS_I_SB(inode), in get_read_data_page()
463 struct f2fs_sb_info *sbi = F2FS_I_SB(dn->inode); in __allocate_data_block() local
477 if (unlikely(!inc_valid_block_count(sbi, dn->inode, 1))) in __allocate_data_block()
481 get_node_info(sbi, dn->nid, &ni); in __allocate_data_block()
487 allocate_data_block(sbi, NULL, dn->data_blkaddr, &dn->data_blkaddr, in __allocate_data_block()
507 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in __allocate_data_blocks() local
515 f2fs_balance_fs(sbi); in __allocate_data_blocks()
516 f2fs_lock_op(sbi); in __allocate_data_blocks()
529 if (unlikely(f2fs_cp_error(sbi))) in __allocate_data_blocks()
547 f2fs_unlock_op(sbi); in __allocate_data_blocks()
556 f2fs_unlock_op(sbi); in __allocate_data_blocks()
574 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_map_blocks() local
608 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_map_blocks()
665 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_map_blocks()
1108 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_write_data_page() local
1116 .sbi = sbi, in f2fs_write_data_page()
1138 if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING))) in f2fs_write_data_page()
1143 available_free_memory(sbi, BASE_CHECK)) in f2fs_write_data_page()
1148 if (unlikely(f2fs_cp_error(sbi))) in f2fs_write_data_page()
1155 if (unlikely(f2fs_cp_error(sbi))) { in f2fs_write_data_page()
1162 else if (has_not_enough_free_secs(sbi, 0)) in f2fs_write_data_page()
1166 f2fs_lock_op(sbi); in f2fs_write_data_page()
1171 f2fs_unlock_op(sbi); in f2fs_write_data_page()
1183 f2fs_balance_fs(sbi); in f2fs_write_data_page()
1185 f2fs_submit_merged_bio(sbi, DATA, WRITE); in f2fs_write_data_page()
1337 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_write_data_pages() local
1353 get_dirty_pages(inode) < nr_pages_to_skip(sbi, DATA) && in f2fs_write_data_pages()
1354 available_free_memory(sbi, DIRTY_DENTS)) in f2fs_write_data_pages()
1358 if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING))) in f2fs_write_data_pages()
1361 diff = nr_pages_to_write(sbi, DATA, wbc); in f2fs_write_data_pages()
1364 mutex_lock(&sbi->writepages); in f2fs_write_data_pages()
1368 f2fs_submit_merged_bio(sbi, DATA, WRITE); in f2fs_write_data_pages()
1370 mutex_unlock(&sbi->writepages); in f2fs_write_data_pages()
1397 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_write_begin() local
1406 f2fs_balance_fs(sbi); in f2fs_write_begin()
1427 f2fs_lock_op(sbi); in f2fs_write_begin()
1430 ipage = get_node_page(sbi, inode->i_ino); in f2fs_write_begin()
1455 f2fs_unlock_op(sbi); in f2fs_write_begin()
1461 f2fs_wait_on_encrypted_page_writeback(sbi, dn.data_blkaddr); in f2fs_write_begin()
1481 .sbi = sbi, in f2fs_write_begin()
1518 f2fs_unlock_op(sbi); in f2fs_write_begin()
1607 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_invalidate_page() local
1609 if (inode->i_ino >= F2FS_ROOT_INO(sbi) && in f2fs_invalidate_page()
1614 if (inode->i_ino == F2FS_META_INO(sbi)) in f2fs_invalidate_page()
1615 dec_page_count(sbi, F2FS_DIRTY_META); in f2fs_invalidate_page()
1616 else if (inode->i_ino == F2FS_NODE_INO(sbi)) in f2fs_invalidate_page()
1617 dec_page_count(sbi, F2FS_DIRTY_NODES); in f2fs_invalidate_page()