wbc 234 drivers/gpu/drm/i915/gem/i915_gem_shmem.c struct writeback_control wbc = { wbc 263 drivers/gpu/drm/i915/gem/i915_gem_shmem.c ret = mapping->a_ops->writepage(page, &wbc); wbc 53 drivers/staging/exfat/exfat_super.c struct writeback_control *wbc); wbc 3174 drivers/staging/exfat/exfat_super.c static int exfat_writepage(struct page *page, struct writeback_control *wbc) wbc 3176 drivers/staging/exfat/exfat_super.c return block_write_full_page(page, exfat_get_block, wbc); wbc 3180 drivers/staging/exfat/exfat_super.c struct writeback_control *wbc) wbc 3182 drivers/staging/exfat/exfat_super.c return mpage_writepages(mapping, wbc, exfat_get_block); wbc 3412 drivers/staging/exfat/exfat_super.c static int exfat_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 178 fs/9p/vfs_addr.c static int v9fs_vfs_writepage(struct page *page, struct writeback_control *wbc) wbc 187 fs/9p/vfs_addr.c redirty_page_for_writepage(wbc, page); wbc 609 fs/9p/vfs_file.c struct writeback_control wbc = { wbc 626 fs/9p/vfs_file.c wbc.nr_to_write = 0; wbc 629 fs/9p/vfs_file.c sync_inode(inode, &wbc); wbc 288 fs/9p/vfs_super.c struct writeback_control *wbc) wbc 312 fs/9p/vfs_super.c struct writeback_control *wbc) wbc 143 fs/adfs/adfs.h int adfs_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 35 fs/adfs/inode.c static int adfs_writepage(struct page *page, struct writeback_control *wbc) wbc 37 fs/adfs/inode.c return block_write_full_page(page, adfs_get_block, wbc); wbc 350 fs/adfs/inode.c int adfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 364 fs/adfs/inode.c ret = adfs_dir_update(sb, &obj, wbc->sync_mode == WB_SYNC_ALL); wbc 189 fs/affs/affs.h struct writeback_control *wbc); wbc 372 fs/affs/file.c static int affs_writepage(struct page *page, struct writeback_control *wbc) wbc 374 fs/affs/file.c return block_write_full_page(page, affs_get_block, wbc); wbc 170 fs/affs/inode.c affs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 274 fs/afs/write.c static void afs_redirty_pages(struct writeback_control *wbc, wbc 299 fs/afs/write.c redirty_page_for_writepage(wbc, page); wbc 452 fs/afs/write.c struct writeback_control *wbc, wbc 578 fs/afs/write.c afs_redirty_pages(wbc, mapping, first, last); wbc 584 fs/afs/write.c afs_redirty_pages(wbc, mapping, first, last); wbc 609 fs/afs/write.c int afs_writepage(struct page *page, struct writeback_control *wbc) wbc 615 fs/afs/write.c ret = afs_write_back_from_locked_page(page->mapping, wbc, page, wbc 616 fs/afs/write.c wbc->range_end >> PAGE_SHIFT); wbc 622 fs/afs/write.c wbc->nr_to_write -= ret; wbc 632 fs/afs/write.c struct writeback_control *wbc, wbc 669 fs/afs/write.c if (wbc->sync_mode != WB_SYNC_NONE) wbc 677 fs/afs/write.c ret = afs_write_back_from_locked_page(mapping, wbc, page, end); wbc 684 fs/afs/write.c wbc->nr_to_write -= ret; wbc 687 fs/afs/write.c } while (index < end && wbc->nr_to_write > 0); wbc 698 fs/afs/write.c struct writeback_control *wbc) wbc 705 fs/afs/write.c if (wbc->range_cyclic) { wbc 708 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, start, end, &next); wbc 709 fs/afs/write.c if (start > 0 && wbc->nr_to_write > 0 && ret == 0) wbc 710 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, 0, start, wbc 713 fs/afs/write.c } else if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) { wbc 715 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, 0, end, &next); wbc 716 fs/afs/write.c if (wbc->nr_to_write > 0) wbc 719 fs/afs/write.c start = wbc->range_start >> PAGE_SHIFT; wbc 720 fs/afs/write.c end = wbc->range_end >> PAGE_SHIFT; wbc 721 fs/afs/write.c ret = afs_writepages_region(mapping, wbc, start, end, &next); wbc 153 fs/bfs/file.c static int bfs_writepage(struct page *page, struct writeback_control *wbc) wbc 155 fs/bfs/file.c return block_write_full_page(page, bfs_get_block, wbc); wbc 117 fs/bfs/inode.c static int bfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 153 fs/bfs/inode.c if (wbc->sync_mode == WB_SYNC_ALL) { wbc 608 fs/block_dev.c static int blkdev_writepage(struct page *page, struct writeback_control *wbc) wbc 610 fs/block_dev.c return block_write_full_page(page, blkdev_get_block, wbc); wbc 739 fs/block_dev.c struct page *page, struct writeback_control *wbc) wbc 2034 fs/block_dev.c struct writeback_control *wbc) wbc 2036 fs/block_dev.c return generic_writepages(mapping, wbc); wbc 2863 fs/btrfs/ctree.h int btrfs_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 2899 fs/btrfs/ctree.h struct writeback_control *wbc); wbc 944 fs/btrfs/disk-io.c struct writeback_control *wbc) wbc 949 fs/btrfs/disk-io.c if (wbc->sync_mode == WB_SYNC_NONE) { wbc 951 fs/btrfs/disk-io.c if (wbc->for_kupdate) wbc 962 fs/btrfs/disk-io.c return btree_write_cache_pages(mapping, wbc); wbc 2935 fs/btrfs/extent_io.c struct writeback_control *wbc, wbc 2977 fs/btrfs/extent_io.c if (wbc) wbc 2978 fs/btrfs/extent_io.c wbc_account_cgroup_owner(wbc, page, page_size); wbc 2990 fs/btrfs/extent_io.c if (wbc) { wbc 2991 fs/btrfs/extent_io.c wbc_init_bio(wbc, bio); wbc 2992 fs/btrfs/extent_io.c wbc_account_cgroup_owner(wbc, page, page_size); wbc 3317 fs/btrfs/extent_io.c static void update_nr_written(struct writeback_control *wbc, wbc 3320 fs/btrfs/extent_io.c wbc->nr_to_write -= nr_written; wbc 3334 fs/btrfs/extent_io.c struct page *page, struct writeback_control *wbc, wbc 3354 fs/btrfs/extent_io.c delalloc_end, &page_started, nr_written, wbc); wbc 3374 fs/btrfs/extent_io.c if (wbc->nr_to_write < delalloc_to_write) { wbc 3379 fs/btrfs/extent_io.c wbc->nr_to_write = min_t(u64, delalloc_to_write, wbc 3392 fs/btrfs/extent_io.c wbc->nr_to_write -= *nr_written; wbc 3412 fs/btrfs/extent_io.c struct writeback_control *wbc, wbc 3438 fs/btrfs/extent_io.c wbc->pages_skipped++; wbc 3440 fs/btrfs/extent_io.c redirty_page_for_writepage(wbc, page); wbc 3442 fs/btrfs/extent_io.c update_nr_written(wbc, nr_written); wbc 3451 fs/btrfs/extent_io.c update_nr_written(wbc, nr_written + 1); wbc 3525 fs/btrfs/extent_io.c ret = submit_extent_page(REQ_OP_WRITE | write_flags, tree, wbc, wbc 3554 fs/btrfs/extent_io.c static int __extent_writepage(struct page *page, struct writeback_control *wbc, wbc 3568 fs/btrfs/extent_io.c write_flags = wbc_to_write_flags(wbc); wbc 3570 fs/btrfs/extent_io.c trace___extent_writepage(page, inode, wbc); wbc 3599 fs/btrfs/extent_io.c ret = writepage_delalloc(inode, page, wbc, start, &nr_written); wbc 3606 fs/btrfs/extent_io.c ret = __extent_writepage_io(inode, page, wbc, epd, wbc 3852 fs/btrfs/extent_io.c struct writeback_control *wbc, wbc 3862 fs/btrfs/extent_io.c unsigned int write_flags = wbc_to_write_flags(wbc) | REQ_META; wbc 3890 fs/btrfs/extent_io.c ret = submit_extent_page(REQ_OP_WRITE | write_flags, tree, wbc, wbc 3905 fs/btrfs/extent_io.c update_nr_written(wbc, 1); wbc 3921 fs/btrfs/extent_io.c struct writeback_control *wbc) wbc 3929 fs/btrfs/extent_io.c .sync_io = wbc->sync_mode == WB_SYNC_ALL, wbc 3943 fs/btrfs/extent_io.c if (wbc->range_cyclic) { wbc 3952 fs/btrfs/extent_io.c index = wbc->range_start >> PAGE_SHIFT; wbc 3953 fs/btrfs/extent_io.c end = wbc->range_end >> PAGE_SHIFT; wbc 3956 fs/btrfs/extent_io.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 3961 fs/btrfs/extent_io.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 4013 fs/btrfs/extent_io.c ret = write_one_eb(eb, wbc, &epd); wbc 4026 fs/btrfs/extent_io.c nr_to_write_done = wbc->nr_to_write <= 0; wbc 4096 fs/btrfs/extent_io.c struct writeback_control *wbc, wbc 4125 fs/btrfs/extent_io.c if (wbc->range_cyclic) { wbc 4134 fs/btrfs/extent_io.c index = wbc->range_start >> PAGE_SHIFT; wbc 4135 fs/btrfs/extent_io.c end = wbc->range_end >> PAGE_SHIFT; wbc 4136 fs/btrfs/extent_io.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 4148 fs/btrfs/extent_io.c if (range_whole && wbc->nr_to_write == LONG_MAX && wbc 4151 fs/btrfs/extent_io.c wbc->tagged_writepages = 1; wbc 4153 fs/btrfs/extent_io.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 4158 fs/btrfs/extent_io.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 4188 fs/btrfs/extent_io.c if (wbc->sync_mode != WB_SYNC_NONE) { wbc 4202 fs/btrfs/extent_io.c ret = __extent_writepage(page, wbc, epd); wbc 4213 fs/btrfs/extent_io.c nr_to_write_done = wbc->nr_to_write <= 0; wbc 4237 fs/btrfs/extent_io.c if (wbc->range_cyclic || (wbc->nr_to_write > 0 && range_whole)) wbc 4244 fs/btrfs/extent_io.c int extent_write_full_page(struct page *page, struct writeback_control *wbc) wbc 4251 fs/btrfs/extent_io.c .sync_io = wbc->sync_mode == WB_SYNC_ALL, wbc 4254 fs/btrfs/extent_io.c ret = __extent_writepage(page, wbc, &epd); wbc 4312 fs/btrfs/extent_io.c struct writeback_control *wbc) wbc 4319 fs/btrfs/extent_io.c .sync_io = wbc->sync_mode == WB_SYNC_ALL, wbc 4322 fs/btrfs/extent_io.c ret = extent_write_cache_pages(mapping, wbc, &epd); wbc 408 fs/btrfs/extent_io.h int extent_write_full_page(struct page *page, struct writeback_control *wbc); wbc 412 fs/btrfs/extent_io.h struct writeback_control *wbc); wbc 414 fs/btrfs/extent_io.h struct writeback_control *wbc); wbc 1711 fs/btrfs/inode.c struct writeback_control *wbc) wbc 1715 fs/btrfs/inode.c unsigned int write_flags = wbc_to_write_flags(wbc); wbc 8862 fs/btrfs/inode.c static int btrfs_writepage(struct page *page, struct writeback_control *wbc) wbc 8868 fs/btrfs/inode.c redirty_page_for_writepage(wbc, page); wbc 8879 fs/btrfs/inode.c redirty_page_for_writepage(wbc, page); wbc 8882 fs/btrfs/inode.c ret = extent_write_full_page(page, wbc); wbc 8888 fs/btrfs/inode.c struct writeback_control *wbc) wbc 8890 fs/btrfs/inode.c return extent_writepages(mapping, wbc); wbc 53 fs/buffer.c enum rw_hint hint, struct writeback_control *wbc); wbc 1681 fs/buffer.c get_block_t *get_block, struct writeback_control *wbc, wbc 1690 fs/buffer.c int write_flags = wbc_to_write_flags(wbc); wbc 1755 fs/buffer.c if (wbc->sync_mode != WB_SYNC_NONE) { wbc 1758 fs/buffer.c redirty_page_for_writepage(wbc, page); wbc 1779 fs/buffer.c inode->i_write_hint, wbc); wbc 1834 fs/buffer.c inode->i_write_hint, wbc); wbc 2727 fs/buffer.c struct writeback_control *wbc) wbc 2765 fs/buffer.c ret = mpage_writepage(page, get_block, wbc); wbc 2767 fs/buffer.c ret = __block_write_full_page(inode, page, get_block, wbc, wbc 2931 fs/buffer.c struct writeback_control *wbc) wbc 2940 fs/buffer.c return __block_write_full_page(inode, page, get_block, wbc, wbc 2964 fs/buffer.c return __block_write_full_page(inode, page, get_block, wbc, wbc 3037 fs/buffer.c enum rw_hint write_hint, struct writeback_control *wbc) wbc 3078 fs/buffer.c if (wbc) { wbc 3079 fs/buffer.c wbc_init_bio(wbc, bio); wbc 3080 fs/buffer.c wbc_account_cgroup_owner(wbc, bh->b_page, bh->b_size); wbc 579 fs/ceph/addr.c static int writepage_nounlock(struct page *page, struct writeback_control *wbc) wbc 608 fs/ceph/addr.c redirty_page_for_writepage(wbc, page); wbc 638 fs/ceph/addr.c if (!wbc) wbc 639 fs/ceph/addr.c wbc = &tmp_wbc; wbc 643 fs/ceph/addr.c redirty_page_for_writepage(wbc, page); wbc 652 fs/ceph/addr.c wbc->pages_skipped++; wbc 670 fs/ceph/addr.c static int ceph_writepage(struct page *page, struct writeback_control *wbc) wbc 676 fs/ceph/addr.c err = writepage_nounlock(page, wbc); wbc 780 fs/ceph/addr.c struct writeback_control *wbc) wbc 797 fs/ceph/addr.c wbc->sync_mode == WB_SYNC_NONE ? "NONE" : wbc 798 fs/ceph/addr.c (wbc->sync_mode == WB_SYNC_ALL ? "ALL" : "HOLD")); wbc 814 fs/ceph/addr.c start_index = wbc->range_cyclic ? mapping->writeback_index : 0; wbc 832 fs/ceph/addr.c if (wbc->range_cyclic) { wbc 839 fs/ceph/addr.c index = wbc->range_start >> PAGE_SHIFT; wbc 840 fs/ceph/addr.c end = wbc->range_end >> PAGE_SHIFT; wbc 841 fs/ceph/addr.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 898 fs/ceph/addr.c wbc->sync_mode != WB_SYNC_NONE) wbc 920 fs/ceph/addr.c if (wbc->sync_mode == WB_SYNC_NONE) { wbc 974 fs/ceph/addr.c redirty_page_for_writepage(wbc, page); wbc 1131 fs/ceph/addr.c wbc->nr_to_write -= i; wbc 1141 fs/ceph/addr.c if (wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) wbc 1157 fs/ceph/addr.c if (wbc->sync_mode != WB_SYNC_NONE && wbc 1182 fs/ceph/addr.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 2280 fs/ceph/caps.c int ceph_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 2286 fs/ceph/caps.c int wait = (wbc->sync_mode == WB_SYNC_ALL && !wbc->for_sync); wbc 1049 fs/ceph/super.h extern int ceph_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 2172 fs/cifs/file.c struct writeback_control *wbc, wbc 2197 fs/cifs/file.c if (!wbc->range_cyclic && page->index > end) { wbc 2209 fs/cifs/file.c if (wbc->sync_mode != WB_SYNC_NONE) wbc 2250 fs/cifs/file.c struct address_space *mapping, struct writeback_control *wbc) wbc 2256 fs/cifs/file.c wdata->sync_mode = wbc->sync_mode; wbc 2279 fs/cifs/file.c struct writeback_control *wbc) wbc 2297 fs/cifs/file.c return generic_writepages(mapping, wbc); wbc 2300 fs/cifs/file.c if (wbc->range_cyclic) { wbc 2304 fs/cifs/file.c index = wbc->range_start >> PAGE_SHIFT; wbc 2305 fs/cifs/file.c end = wbc->range_end >> PAGE_SHIFT; wbc 2306 fs/cifs/file.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 2352 fs/cifs/file.c nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc, wbc 2374 fs/cifs/file.c rc = wdata_send_pages(wdata, nr_pages, mapping, wbc); wbc 2384 fs/cifs/file.c redirty_page_for_writepage(wbc, wbc 2396 fs/cifs/file.c if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN) { wbc 2410 fs/cifs/file.c wbc->nr_to_write -= nr_pages; wbc 2411 fs/cifs/file.c if (wbc->nr_to_write <= 0) wbc 2430 fs/cifs/file.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 2440 fs/cifs/file.c cifs_writepage_locked(struct page *page, struct writeback_control *wbc) wbc 2465 fs/cifs/file.c if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN) wbc 2467 fs/cifs/file.c redirty_page_for_writepage(wbc, page); wbc 2480 fs/cifs/file.c static int cifs_writepage(struct page *page, struct writeback_control *wbc) wbc 2482 fs/cifs/file.c int rc = cifs_writepage_locked(page, wbc); wbc 4653 fs/cifs/file.c struct writeback_control wbc = { wbc 4663 fs/cifs/file.c rc = cifs_writepage_locked(page, &wbc); wbc 940 fs/dax.c struct block_device *bdev, struct writeback_control *wbc) wbc 942 fs/dax.c XA_STATE(xas, &mapping->i_pages, wbc->range_start >> PAGE_SHIFT); wbc 944 fs/dax.c pgoff_t end_index = wbc->range_end >> PAGE_SHIFT; wbc 953 fs/dax.c if (!mapping->nrexceptional || wbc->sync_mode != WB_SYNC_ALL) wbc 72 fs/ecryptfs/ecryptfs_kernel.h struct writeback_control *wbc; wbc 51 fs/ecryptfs/mmap.c static int ecryptfs_writepage(struct page *page, struct writeback_control *wbc) wbc 870 fs/ext2/inode.c static int ext2_writepage(struct page *page, struct writeback_control *wbc) wbc 872 fs/ext2/inode.c return block_write_full_page(page, ext2_get_block, wbc); wbc 928 fs/ext2/inode.c struct writeback_control *wbc) wbc 930 fs/ext2/inode.c return nobh_writepage(page, ext2_get_block, wbc); wbc 955 fs/ext2/inode.c ext2_writepages(struct address_space *mapping, struct writeback_control *wbc) wbc 957 fs/ext2/inode.c return mpage_writepages(mapping, wbc, ext2_get_block); wbc 961 fs/ext2/inode.c ext2_dax_writepages(struct address_space *mapping, struct writeback_control *wbc) wbc 964 fs/ext2/inode.c mapping->host->i_sb->s_bdev, wbc); wbc 1639 fs/ext2/inode.c int ext2_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1641 fs/ext2/inode.c return __ext2_write_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 3341 fs/ext4/ext4.h struct writeback_control *wbc); wbc 3347 fs/ext4/ext4.h struct writeback_control *wbc, wbc 1674 fs/ext4/inode.c struct writeback_control *wbc; wbc 2123 fs/ext4/inode.c struct writeback_control *wbc) wbc 2167 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); wbc 2190 fs/ext4/inode.c ext4_io_submit_init(&io_submit, wbc); wbc 2193 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); wbc 2197 fs/ext4/inode.c ret = ext4_bio_write_page(&io_submit, page, len, wbc, keep_towrite); wbc 2231 fs/ext4/inode.c err = ext4_bio_write_page(&mpd->io_submit, page, len, mpd->wbc, false); wbc 2233 fs/ext4/inode.c mpd->wbc->nr_to_write--; wbc 2636 fs/ext4/inode.c long left = mpd->wbc->nr_to_write; wbc 2645 fs/ext4/inode.c if (mpd->wbc->sync_mode == WB_SYNC_ALL || mpd->wbc->tagged_writepages) wbc 2670 fs/ext4/inode.c if (mpd->wbc->sync_mode == WB_SYNC_NONE && left <= 0) wbc 2687 fs/ext4/inode.c (mpd->wbc->sync_mode == WB_SYNC_NONE)) || wbc 2719 fs/ext4/inode.c struct writeback_control *wbc) wbc 2722 fs/ext4/inode.c long nr_to_write = wbc->nr_to_write; wbc 2738 fs/ext4/inode.c trace_ext4_writepages(inode, wbc); wbc 2749 fs/ext4/inode.c ret = generic_writepages(mapping, wbc); wbc 2796 fs/ext4/inode.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 2799 fs/ext4/inode.c if (wbc->range_cyclic) { wbc 2806 fs/ext4/inode.c mpd.first_page = wbc->range_start >> PAGE_SHIFT; wbc 2807 fs/ext4/inode.c mpd.last_page = wbc->range_end >> PAGE_SHIFT; wbc 2811 fs/ext4/inode.c mpd.wbc = wbc; wbc 2812 fs/ext4/inode.c ext4_io_submit_init(&mpd.io_submit, wbc); wbc 2814 fs/ext4/inode.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 2866 fs/ext4/inode.c wbc->nr_to_write, inode->i_ino, ret); wbc 2874 fs/ext4/inode.c trace_ext4_da_write_pages(inode, mpd.first_page, mpd.wbc); wbc 2940 fs/ext4/inode.c if (!ret && !cycled && wbc->nr_to_write > 0) { wbc 2948 fs/ext4/inode.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 2956 fs/ext4/inode.c trace_ext4_writepages_result(inode, wbc, ret, wbc 2957 fs/ext4/inode.c nr_to_write - wbc->nr_to_write); wbc 2963 fs/ext4/inode.c struct writeback_control *wbc) wbc 2966 fs/ext4/inode.c long nr_to_write = wbc->nr_to_write; wbc 2974 fs/ext4/inode.c trace_ext4_writepages(inode, wbc); wbc 2976 fs/ext4/inode.c ret = dax_writeback_mapping_range(mapping, inode->i_sb->s_bdev, wbc); wbc 2977 fs/ext4/inode.c trace_ext4_writepages_result(inode, wbc, ret, wbc 2978 fs/ext4/inode.c nr_to_write - wbc->nr_to_write); wbc 5421 fs/ext4/inode.c int ext4_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 5444 fs/ext4/inode.c if (wbc->sync_mode != WB_SYNC_ALL || wbc->for_sync) wbc 5459 fs/ext4/inode.c if (wbc->sync_mode == WB_SYNC_ALL && !wbc->for_sync) wbc 354 fs/ext4/page-io.c struct writeback_control *wbc) wbc 356 fs/ext4/page-io.c io->io_wbc = wbc; wbc 407 fs/ext4/page-io.c struct writeback_control *wbc, wbc 494 fs/ext4/page-io.c (io->io_bio || wbc->sync_mode == WB_SYNC_ALL)) { wbc 530 fs/ext4/page-io.c redirty_page_for_writepage(wbc, page); wbc 1235 fs/ext4/super.c struct writeback_control wbc = { wbc 1240 fs/ext4/super.c return ext4_write_inode(inode, &wbc); wbc 292 fs/f2fs/checkpoint.c struct writeback_control *wbc, wbc 303 fs/f2fs/checkpoint.c if (wbc->for_reclaim && page->index < GET_SUM_BLOCK(sbi, 0)) wbc 309 fs/f2fs/checkpoint.c if (wbc->for_reclaim) wbc 320 fs/f2fs/checkpoint.c redirty_page_for_writepage(wbc, page); wbc 325 fs/f2fs/checkpoint.c struct writeback_control *wbc) wbc 327 fs/f2fs/checkpoint.c return __f2fs_write_meta_page(page, wbc, FS_META_IO); wbc 331 fs/f2fs/checkpoint.c struct writeback_control *wbc) wbc 340 fs/f2fs/checkpoint.c if (wbc->sync_mode != WB_SYNC_ALL && wbc 349 fs/f2fs/checkpoint.c trace_f2fs_writepages(mapping->host, wbc, META); wbc 350 fs/f2fs/checkpoint.c diff = nr_pages_to_write(sbi, META, wbc); wbc 351 fs/f2fs/checkpoint.c written = f2fs_sync_meta_pages(sbi, META, wbc->nr_to_write, FS_META_IO); wbc 353 fs/f2fs/checkpoint.c wbc->nr_to_write = max((long)0, wbc->nr_to_write - written - diff); wbc 357 fs/f2fs/checkpoint.c wbc->pages_skipped += get_pages(sbi, F2FS_DIRTY_META); wbc 358 fs/f2fs/checkpoint.c trace_f2fs_writepages(mapping->host, wbc, META); wbc 370 fs/f2fs/checkpoint.c struct writeback_control wbc = { wbc 410 fs/f2fs/checkpoint.c if (__f2fs_write_meta_page(page, &wbc, io_type)) { wbc 1161 fs/f2fs/checkpoint.c struct writeback_control wbc = { wbc 1225 fs/f2fs/checkpoint.c err = f2fs_sync_node_pages(sbi, &wbc, false, FS_CP_NODE_IO); wbc 1340 fs/f2fs/checkpoint.c struct writeback_control wbc = { wbc 1361 fs/f2fs/checkpoint.c err = __f2fs_write_meta_page(page, &wbc, FS_CP_META_IO); wbc 2080 fs/f2fs/data.c struct writeback_control *wbc, wbc 2097 fs/f2fs/data.c .op_flags = wbc_to_write_flags(wbc), wbc 2104 fs/f2fs/data.c .io_wbc = wbc, wbc 2143 fs/f2fs/data.c (!wbc->for_reclaim && wbc 2154 fs/f2fs/data.c if (!wbc->for_reclaim) wbc 2196 fs/f2fs/data.c if (wbc->for_reclaim) { wbc 2222 fs/f2fs/data.c redirty_page_for_writepage(wbc, page); wbc 2229 fs/f2fs/data.c if (!err || wbc->for_reclaim) wbc 2236 fs/f2fs/data.c struct writeback_control *wbc) wbc 2238 fs/f2fs/data.c return __write_data_page(page, NULL, NULL, NULL, wbc, FS_DATA_IO); wbc 2247 fs/f2fs/data.c struct writeback_control *wbc, wbc 2274 fs/f2fs/data.c if (wbc->range_cyclic) { wbc 2283 fs/f2fs/data.c index = wbc->range_start >> PAGE_SHIFT; wbc 2284 fs/f2fs/data.c end = wbc->range_end >> PAGE_SHIFT; wbc 2285 fs/f2fs/data.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 2289 fs/f2fs/data.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 2294 fs/f2fs/data.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 2311 fs/f2fs/data.c wbc->sync_mode == WB_SYNC_NONE) { wbc 2332 fs/f2fs/data.c if (wbc->sync_mode != WB_SYNC_NONE) { wbc 2345 fs/f2fs/data.c &last_block, wbc, io_type); wbc 2357 fs/f2fs/data.c if (wbc->sync_mode == WB_SYNC_ALL) { wbc 2372 fs/f2fs/data.c if (--wbc->nr_to_write <= 0 && wbc 2373 fs/f2fs/data.c wbc->sync_mode == WB_SYNC_NONE) { wbc 2388 fs/f2fs/data.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 2402 fs/f2fs/data.c struct writeback_control *wbc) wbc 2411 fs/f2fs/data.c if (wbc->sync_mode != WB_SYNC_ALL) wbc 2419 fs/f2fs/data.c struct writeback_control *wbc, wbc 2433 fs/f2fs/data.c if (!get_dirty_pages(inode) && wbc->sync_mode == WB_SYNC_NONE) wbc 2441 fs/f2fs/data.c wbc->sync_mode == WB_SYNC_NONE && wbc 2450 fs/f2fs/data.c trace_f2fs_writepages(mapping->host, wbc, DATA); wbc 2453 fs/f2fs/data.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 2458 fs/f2fs/data.c if (__should_serialize_io(inode, wbc)) { wbc 2464 fs/f2fs/data.c ret = f2fs_write_cache_pages(mapping, wbc, io_type); wbc 2470 fs/f2fs/data.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 2481 fs/f2fs/data.c wbc->pages_skipped += get_dirty_pages(inode); wbc 2482 fs/f2fs/data.c trace_f2fs_writepages(mapping->host, wbc, DATA); wbc 2487 fs/f2fs/data.c struct writeback_control *wbc) wbc 2491 fs/f2fs/data.c return __f2fs_write_data_pages(mapping, wbc, wbc 2951 fs/f2fs/f2fs.h int f2fs_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 3064 fs/f2fs/f2fs.h struct writeback_control *wbc, bool atomic, wbc 3067 fs/f2fs/f2fs.h struct writeback_control *wbc, wbc 215 fs/f2fs/file.c struct writeback_control wbc = { wbc 287 fs/f2fs/file.c ret = f2fs_fsync_node_pages(sbi, inode, &wbc, atomic, &seq_id); wbc 610 fs/f2fs/inode.c int f2fs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 633 fs/f2fs/inode.c if (wbc && wbc->nr_to_write) wbc 1499 fs/f2fs/node.c struct writeback_control *wbc, bool do_balance, wbc 1510 fs/f2fs/node.c .op_flags = wbc_to_write_flags(wbc), wbc 1515 fs/f2fs/node.c .io_wbc = wbc, wbc 1528 fs/f2fs/node.c wbc->sync_mode == WB_SYNC_NONE && wbc 1539 fs/f2fs/node.c if (wbc->for_reclaim) { wbc 1581 fs/f2fs/node.c if (wbc->for_reclaim) { wbc 1600 fs/f2fs/node.c redirty_page_for_writepage(wbc, page); wbc 1609 fs/f2fs/node.c struct writeback_control wbc = { wbc 1625 fs/f2fs/node.c &wbc, false, FS_GC_NODE_IO, NULL)) { wbc 1643 fs/f2fs/node.c struct writeback_control *wbc) wbc 1645 fs/f2fs/node.c return __write_node_page(page, false, NULL, wbc, false, wbc 1650 fs/f2fs/node.c struct writeback_control *wbc, bool atomic, wbc 1730 fs/f2fs/node.c &submitted, wbc, true, wbc 1809 fs/f2fs/node.c struct writeback_control *wbc, wbc 1835 fs/f2fs/node.c wbc->sync_mode == WB_SYNC_NONE) { wbc 1855 fs/f2fs/node.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 1895 fs/f2fs/node.c wbc, do_balance, io_type, NULL); wbc 1901 fs/f2fs/node.c if (--wbc->nr_to_write == 0) wbc 1907 fs/f2fs/node.c if (wbc->nr_to_write == 0) { wbc 1915 fs/f2fs/node.c wbc->sync_mode == WB_SYNC_NONE && step == 1) wbc 1973 fs/f2fs/node.c struct writeback_control *wbc) wbc 1986 fs/f2fs/node.c if (wbc->sync_mode != WB_SYNC_ALL && wbc 1991 fs/f2fs/node.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 1996 fs/f2fs/node.c trace_f2fs_writepages(mapping->host, wbc, NODE); wbc 1998 fs/f2fs/node.c diff = nr_pages_to_write(sbi, NODE, wbc); wbc 2000 fs/f2fs/node.c f2fs_sync_node_pages(sbi, wbc, true, FS_NODE_IO); wbc 2002 fs/f2fs/node.c wbc->nr_to_write = max((long)0, wbc->nr_to_write - diff); wbc 2004 fs/f2fs/node.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 2009 fs/f2fs/node.c wbc->pages_skipped += get_pages(sbi, F2FS_DIRTY_NODES); wbc 2010 fs/f2fs/node.c trace_f2fs_writepages(mapping->host, wbc, NODE); wbc 835 fs/f2fs/segment.h struct writeback_control *wbc) wbc 839 fs/f2fs/segment.h if (wbc->sync_mode != WB_SYNC_NONE) wbc 842 fs/f2fs/segment.h nr_to_write = wbc->nr_to_write; wbc 847 fs/f2fs/segment.h wbc->nr_to_write = desired; wbc 196 fs/fat/inode.c static int fat_writepage(struct page *page, struct writeback_control *wbc) wbc 198 fs/fat/inode.c return block_write_full_page(page, fat_get_block, wbc); wbc 202 fs/fat/inode.c struct writeback_control *wbc) wbc 204 fs/fat/inode.c return mpage_writepages(mapping, wbc, fat_get_block); wbc 904 fs/fat/inode.c static int fat_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 915 fs/fat/inode.c err = __fat_write_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 557 fs/fs-writeback.c void wbc_attach_and_unlock_inode(struct writeback_control *wbc, wbc 565 fs/fs-writeback.c wbc->wb = inode_to_wb(inode); wbc 566 fs/fs-writeback.c wbc->inode = inode; wbc 568 fs/fs-writeback.c wbc->wb_id = wbc->wb->memcg_css->id; wbc 569 fs/fs-writeback.c wbc->wb_lcand_id = inode->i_wb_frn_winner; wbc 570 fs/fs-writeback.c wbc->wb_tcand_id = 0; wbc 571 fs/fs-writeback.c wbc->wb_bytes = 0; wbc 572 fs/fs-writeback.c wbc->wb_lcand_bytes = 0; wbc 573 fs/fs-writeback.c wbc->wb_tcand_bytes = 0; wbc 575 fs/fs-writeback.c wb_get(wbc->wb); wbc 585 fs/fs-writeback.c if (unlikely(wb_dying(wbc->wb) && !css_is_dying(wbc->wb->memcg_css))) wbc 586 fs/fs-writeback.c inode_switch_wbs(inode, wbc->wb_id); wbc 627 fs/fs-writeback.c void wbc_detach_inode(struct writeback_control *wbc) wbc 629 fs/fs-writeback.c struct bdi_writeback *wb = wbc->wb; wbc 630 fs/fs-writeback.c struct inode *inode = wbc->inode; wbc 642 fs/fs-writeback.c if (wbc->wb_bytes >= wbc->wb_lcand_bytes && wbc 643 fs/fs-writeback.c wbc->wb_bytes >= wbc->wb_tcand_bytes) { wbc 644 fs/fs-writeback.c max_id = wbc->wb_id; wbc 645 fs/fs-writeback.c max_bytes = wbc->wb_bytes; wbc 646 fs/fs-writeback.c } else if (wbc->wb_lcand_bytes >= wbc->wb_tcand_bytes) { wbc 647 fs/fs-writeback.c max_id = wbc->wb_lcand_id; wbc 648 fs/fs-writeback.c max_bytes = wbc->wb_lcand_bytes; wbc 650 fs/fs-writeback.c max_id = wbc->wb_tcand_id; wbc 651 fs/fs-writeback.c max_bytes = wbc->wb_tcand_bytes; wbc 683 fs/fs-writeback.c if (wbc->wb_id != max_id) wbc 687 fs/fs-writeback.c trace_inode_foreign_history(inode, wbc, history); wbc 708 fs/fs-writeback.c wb_put(wbc->wb); wbc 709 fs/fs-writeback.c wbc->wb = NULL; wbc 723 fs/fs-writeback.c void wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page, wbc 735 fs/fs-writeback.c if (!wbc->wb || wbc->no_cgroup_owner) wbc 745 fs/fs-writeback.c if (id == wbc->wb_id) { wbc 746 fs/fs-writeback.c wbc->wb_bytes += bytes; wbc 750 fs/fs-writeback.c if (id == wbc->wb_lcand_id) wbc 751 fs/fs-writeback.c wbc->wb_lcand_bytes += bytes; wbc 754 fs/fs-writeback.c if (!wbc->wb_tcand_bytes) wbc 755 fs/fs-writeback.c wbc->wb_tcand_id = id; wbc 756 fs/fs-writeback.c if (id == wbc->wb_tcand_id) wbc 757 fs/fs-writeback.c wbc->wb_tcand_bytes += bytes; wbc 759 fs/fs-writeback.c wbc->wb_tcand_bytes -= min(bytes, wbc->wb_tcand_bytes); wbc 1306 fs/fs-writeback.c static int write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1311 fs/fs-writeback.c trace_writeback_write_inode_start(inode, wbc); wbc 1312 fs/fs-writeback.c ret = inode->i_sb->s_op->write_inode(inode, wbc); wbc 1313 fs/fs-writeback.c trace_writeback_write_inode(inode, wbc); wbc 1378 fs/fs-writeback.c struct writeback_control *wbc) wbc 1389 fs/fs-writeback.c (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages)) wbc 1392 fs/fs-writeback.c if (wbc->pages_skipped) { wbc 1406 fs/fs-writeback.c if (wbc->nr_to_write <= 0) { wbc 1441 fs/fs-writeback.c __writeback_single_inode(struct inode *inode, struct writeback_control *wbc) wbc 1444 fs/fs-writeback.c long nr_to_write = wbc->nr_to_write; wbc 1450 fs/fs-writeback.c trace_writeback_single_inode_start(inode, wbc, nr_to_write); wbc 1452 fs/fs-writeback.c ret = do_writepages(mapping, wbc); wbc 1461 fs/fs-writeback.c if (wbc->sync_mode == WB_SYNC_ALL && !wbc->for_sync) { wbc 1477 fs/fs-writeback.c wbc->sync_mode == WB_SYNC_ALL || wbc 1511 fs/fs-writeback.c int err = write_inode(inode, wbc); wbc 1515 fs/fs-writeback.c trace_writeback_single_inode(inode, wbc, nr_to_write); wbc 1528 fs/fs-writeback.c struct writeback_control *wbc) wbc 1540 fs/fs-writeback.c if (wbc->sync_mode != WB_SYNC_ALL) wbc 1559 fs/fs-writeback.c (wbc->sync_mode != WB_SYNC_ALL || wbc 1563 fs/fs-writeback.c wbc_attach_and_unlock_inode(wbc, inode); wbc 1565 fs/fs-writeback.c ret = __writeback_single_inode(inode, wbc); wbc 1567 fs/fs-writeback.c wbc_detach_inode(wbc); wbc 1628 fs/fs-writeback.c struct writeback_control wbc = { wbc 1676 fs/fs-writeback.c if ((inode->i_state & I_SYNC) && wbc.sync_mode != WB_SYNC_ALL) { wbc 1706 fs/fs-writeback.c wbc_attach_and_unlock_inode(&wbc, inode); wbc 1709 fs/fs-writeback.c wbc.nr_to_write = write_chunk; wbc 1710 fs/fs-writeback.c wbc.pages_skipped = 0; wbc 1716 fs/fs-writeback.c __writeback_single_inode(inode, &wbc); wbc 1718 fs/fs-writeback.c wbc_detach_inode(&wbc); wbc 1719 fs/fs-writeback.c work->nr_pages -= write_chunk - wbc.nr_to_write; wbc 1720 fs/fs-writeback.c wrote += write_chunk - wbc.nr_to_write; wbc 1743 fs/fs-writeback.c requeue_inode(inode, tmp_wb, &wbc); wbc 2574 fs/fs-writeback.c struct writeback_control wbc = { wbc 2582 fs/fs-writeback.c wbc.nr_to_write = 0; wbc 2585 fs/fs-writeback.c return writeback_single_inode(inode, &wbc); wbc 2600 fs/fs-writeback.c int sync_inode(struct inode *inode, struct writeback_control *wbc) wbc 2602 fs/fs-writeback.c return writeback_single_inode(inode, wbc); wbc 2617 fs/fs-writeback.c struct writeback_control wbc = { wbc 2622 fs/fs-writeback.c return sync_inode(inode, &wbc); wbc 1780 fs/fuse/file.c int fuse_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1877 fs/fuse/file.c static int fuse_writepage(struct page *page, struct writeback_control *wbc) wbc 1887 fs/fuse/file.c WARN_ON(wbc->sync_mode == WB_SYNC_ALL); wbc 1889 fs/fuse/file.c redirty_page_for_writepage(wbc, page); wbc 2011 fs/fuse/file.c struct writeback_control *wbc, void *_data) wbc 2126 fs/fuse/file.c struct writeback_control *wbc) wbc 2148 fs/fuse/file.c err = write_cache_pages(mapping, wbc, fuse_writepages_fill, &data); wbc 1056 fs/fuse/fuse_i.h int fuse_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 89 fs/gfs2/aops.c static int gfs2_writepage(struct page *page, struct writeback_control *wbc) wbc 109 fs/gfs2/aops.c return nobh_writepage(page, gfs2_get_block_noalloc, wbc); wbc 112 fs/gfs2/aops.c redirty_page_for_writepage(wbc, page); wbc 122 fs/gfs2/aops.c struct writeback_control *wbc) wbc 140 fs/gfs2/aops.c return __block_write_full_page(inode, page, get_block, wbc, wbc 155 fs/gfs2/aops.c static int __gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) wbc 169 fs/gfs2/aops.c return gfs2_write_full_page(page, gfs2_get_block_noalloc, wbc); wbc 181 fs/gfs2/aops.c static int gfs2_jdata_writepage(struct page *page, struct writeback_control *wbc) wbc 192 fs/gfs2/aops.c ret = __gfs2_jdata_writepage(page, wbc); wbc 196 fs/gfs2/aops.c redirty_page_for_writepage(wbc, page); wbc 210 fs/gfs2/aops.c struct writeback_control *wbc) wbc 213 fs/gfs2/aops.c int ret = mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); wbc 239 fs/gfs2/aops.c struct writeback_control *wbc, wbc 273 fs/gfs2/aops.c if (wbc->sync_mode != WB_SYNC_NONE) wbc 283 fs/gfs2/aops.c trace_wbc_writepage(wbc, inode_to_bdi(inode)); wbc 285 fs/gfs2/aops.c ret = __gfs2_jdata_writepage(page, wbc); wbc 313 fs/gfs2/aops.c if (--wbc->nr_to_write <= 0 && wbc->sync_mode == WB_SYNC_NONE) { wbc 334 fs/gfs2/aops.c struct writeback_control *wbc) wbc 349 fs/gfs2/aops.c if (wbc->range_cyclic) { wbc 358 fs/gfs2/aops.c index = wbc->range_start >> PAGE_SHIFT; wbc 359 fs/gfs2/aops.c end = wbc->range_end >> PAGE_SHIFT; wbc 360 fs/gfs2/aops.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 364 fs/gfs2/aops.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 370 fs/gfs2/aops.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 379 fs/gfs2/aops.c ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); wbc 400 fs/gfs2/aops.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 415 fs/gfs2/aops.c struct writeback_control *wbc) wbc 421 fs/gfs2/aops.c ret = gfs2_write_cache_jdata(mapping, wbc); wbc 422 fs/gfs2/aops.c if (ret == 0 && wbc->sync_mode == WB_SYNC_ALL) { wbc 425 fs/gfs2/aops.c ret = gfs2_write_cache_jdata(mapping, wbc); wbc 91 fs/gfs2/log.c struct writeback_control *wbc, wbc 128 fs/gfs2/log.c generic_writepages(mapping, wbc); wbc 130 fs/gfs2/log.c if (wbc->nr_to_write <= 0) wbc 148 fs/gfs2/log.c void gfs2_ail1_flush(struct gfs2_sbd *sdp, struct writeback_control *wbc) wbc 155 fs/gfs2/log.c trace_gfs2_ail_flush(sdp, wbc, 1); wbc 160 fs/gfs2/log.c if (wbc->nr_to_write <= 0) wbc 162 fs/gfs2/log.c if (gfs2_ail1_start_one(sdp, wbc, tr, &withdraw)) wbc 169 fs/gfs2/log.c trace_gfs2_ail_flush(sdp, wbc, 0); wbc 179 fs/gfs2/log.c struct writeback_control wbc = { wbc 186 fs/gfs2/log.c return gfs2_ail1_flush(sdp, &wbc); wbc 75 fs/gfs2/log.h extern void gfs2_ail1_flush(struct gfs2_sbd *sdp, struct writeback_control *wbc); wbc 33 fs/gfs2/meta_io.c static int gfs2_aspace_writepage(struct page *page, struct writeback_control *wbc) wbc 37 fs/gfs2/meta_io.c int write_flags = REQ_META | REQ_PRIO | wbc_to_write_flags(wbc); wbc 55 fs/gfs2/meta_io.c if (wbc->sync_mode != WB_SYNC_NONE) { wbc 58 fs/gfs2/meta_io.c redirty_page_for_writepage(wbc, page); wbc 501 fs/gfs2/super.c static int gfs2_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 508 fs/gfs2/super.c bool flush_all = (wbc->sync_mode == WB_SYNC_ALL || gfs2_is_jdata(ip)); wbc 515 fs/gfs2/super.c gfs2_ail1_flush(sdp, wbc); wbc 405 fs/gfs2/trace_gfs2.h TP_PROTO(const struct gfs2_sbd *sdp, const struct writeback_control *wbc, int start), wbc 407 fs/gfs2/trace_gfs2.h TP_ARGS(sdp, wbc, start), wbc 419 fs/gfs2/trace_gfs2.h __entry->sync_mode = wbc->sync_mode; wbc 420 fs/gfs2/trace_gfs2.h __entry->nr_to_write = wbc->nr_to_write; wbc 31 fs/hfs/inode.c static int hfs_writepage(struct page *page, struct writeback_control *wbc) wbc 33 fs/hfs/inode.c return block_write_full_page(page, hfs_get_block, wbc); wbc 155 fs/hfs/inode.c struct writeback_control *wbc) wbc 157 fs/hfs/inode.c return mpage_writepages(mapping, wbc, hfs_get_block); wbc 418 fs/hfs/inode.c int hfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 30 fs/hfsplus/inode.c static int hfsplus_writepage(struct page *page, struct writeback_control *wbc) wbc 32 fs/hfsplus/inode.c return block_write_full_page(page, hfsplus_get_block, wbc); wbc 152 fs/hfsplus/inode.c struct writeback_control *wbc) wbc 154 fs/hfsplus/inode.c return mpage_writepages(mapping, wbc, hfsplus_get_block); wbc 148 fs/hfsplus/super.c struct writeback_control *wbc) wbc 401 fs/hostfs/hostfs_kern.c static int hostfs_writepage(struct page *page, struct writeback_control *wbc) wbc 123 fs/hpfs/file.c static int hpfs_writepage(struct page *page, struct writeback_control *wbc) wbc 125 fs/hpfs/file.c return block_write_full_page(page, hpfs_get_block, wbc); wbc 135 fs/hpfs/file.c struct writeback_control *wbc) wbc 137 fs/hpfs/file.c return mpage_writepages(mapping, wbc, hpfs_get_block); wbc 194 fs/jbd2/commit.c struct writeback_control wbc = { wbc 201 fs/jbd2/commit.c ret = generic_writepages(mapping, &wbc); wbc 115 fs/jfs/inode.c int jfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 117 fs/jfs/inode.c int wait = wbc->sync_mode == WB_SYNC_ALL; wbc 283 fs/jfs/inode.c static int jfs_writepage(struct page *page, struct writeback_control *wbc) wbc 285 fs/jfs/inode.c return block_write_full_page(page, jfs_get_block, wbc); wbc 289 fs/jfs/inode.c struct writeback_control *wbc) wbc 291 fs/jfs/inode.c return mpage_writepages(mapping, wbc, jfs_get_block); wbc 335 fs/jfs/jfs_metapage.c static int metapage_writepage(struct page *page, struct writeback_control *wbc) wbc 443 fs/jfs/jfs_metapage.c redirty_page_for_writepage(wbc, page); wbc 24 fs/minix/inode.c struct writeback_control *wbc); wbc 381 fs/minix/inode.c static int minix_writepage(struct page *page, struct writeback_control *wbc) wbc 383 fs/minix/inode.c return block_write_full_page(page, minix_get_block, wbc); wbc 599 fs/minix/inode.c static int minix_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 610 fs/minix/inode.c if (wbc->sync_mode == WB_SYNC_ALL && buffer_dirty(bh)) { wbc 496 fs/mpage.c static int __mpage_writepage(struct page *page, struct writeback_control *wbc, wbc 519 fs/mpage.c int op_flags = wbc_to_write_flags(wbc); wbc 633 fs/mpage.c page, wbc)) wbc 641 fs/mpage.c wbc_init_bio(wbc, bio); wbc 650 fs/mpage.c wbc_account_cgroup_owner(wbc, page, PAGE_SIZE); wbc 678 fs/mpage.c ret = mapping->a_ops->writepage(page, wbc); wbc 713 fs/mpage.c struct writeback_control *wbc, get_block_t get_block) wbc 721 fs/mpage.c ret = generic_writepages(mapping, wbc); wbc 730 fs/mpage.c ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd); wbc 732 fs/mpage.c int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? wbc 743 fs/mpage.c struct writeback_control *wbc) wbc 751 fs/mpage.c int ret = __mpage_writepage(page, wbc, &mpd); wbc 753 fs/mpage.c int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? wbc 19 fs/nfs/nfs4super.c static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 76 fs/nfs/nfs4super.c static int nfs4_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 78 fs/nfs/nfs4super.c int ret = nfs_write_inode(inode, wbc); wbc 82 fs/nfs/nfs4super.c wbc->sync_mode == WB_SYNC_ALL); wbc 331 fs/nfs/write.c static int wb_priority(struct writeback_control *wbc) wbc 335 fs/nfs/write.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 659 fs/nfs/write.c static int nfs_do_writepage(struct page *page, struct writeback_control *wbc, wbc 667 fs/nfs/write.c redirty_page_for_writepage(wbc, page); wbc 677 fs/nfs/write.c struct writeback_control *wbc) wbc 686 fs/nfs/write.c err = nfs_do_writepage(page, wbc, &pgio); wbc 696 fs/nfs/write.c int nfs_writepage(struct page *page, struct writeback_control *wbc) wbc 700 fs/nfs/write.c ret = nfs_writepage_locked(page, wbc); wbc 706 fs/nfs/write.c static int nfs_writepages_callback(struct page *page, struct writeback_control *wbc, void *data) wbc 710 fs/nfs/write.c ret = nfs_do_writepage(page, wbc, data); wbc 721 fs/nfs/write.c int nfs_writepages(struct address_space *mapping, struct writeback_control *wbc) wbc 734 fs/nfs/write.c nfs_pageio_init_write(&pgio, inode, wb_priority(wbc), false, wbc 737 fs/nfs/write.c err = write_cache_pages(mapping, wbc, nfs_writepages_callback, &pgio); wbc 1925 fs/nfs/write.c struct writeback_control *wbc) wbc 1942 fs/nfs/write.c if (wbc && wbc->sync_mode == WB_SYNC_NONE) { wbc 1943 fs/nfs/write.c if (nscan < wbc->nr_to_write) wbc 1944 fs/nfs/write.c wbc->nr_to_write -= nscan; wbc 1946 fs/nfs/write.c wbc->nr_to_write = 0; wbc 1964 fs/nfs/write.c int nfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1970 fs/nfs/write.c if (wbc->sync_mode == WB_SYNC_NONE) { wbc 1985 fs/nfs/write.c ret = __nfs_commit_inode(inode, flags, wbc); wbc 2075 fs/nfs/write.c struct writeback_control wbc = { wbc 2088 fs/nfs/write.c ret = nfs_writepage_locked(page, &wbc); wbc 163 fs/nilfs2/inode.c struct writeback_control *wbc) wbc 173 fs/nilfs2/inode.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 175 fs/nilfs2/inode.c wbc->range_start, wbc 176 fs/nilfs2/inode.c wbc->range_end); wbc 180 fs/nilfs2/inode.c static int nilfs_writepage(struct page *page, struct writeback_control *wbc) wbc 197 fs/nilfs2/inode.c redirty_page_for_writepage(wbc, page); wbc 200 fs/nilfs2/inode.c if (wbc->sync_mode == WB_SYNC_ALL) { wbc 204 fs/nilfs2/inode.c } else if (wbc->for_reclaim) wbc 401 fs/nilfs2/mdt.c nilfs_mdt_write_page(struct page *page, struct writeback_control *wbc) wbc 419 fs/nilfs2/mdt.c redirty_page_for_writepage(wbc, page); wbc 427 fs/nilfs2/mdt.c if (wbc->sync_mode == WB_SYNC_ALL) wbc 429 fs/nilfs2/mdt.c else if (wbc->for_reclaim) wbc 533 fs/ntfs/aops.c static int ntfs_write_block(struct page *page, struct writeback_control *wbc) wbc 574 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); wbc 849 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); wbc 903 fs/ntfs/aops.c struct writeback_control *wbc) wbc 943 fs/ntfs/aops.c sync = (wbc->sync_mode == WB_SYNC_ALL); wbc 1298 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); wbc 1339 fs/ntfs/aops.c static int ntfs_writepage(struct page *page, struct writeback_control *wbc) wbc 1410 fs/ntfs/aops.c return ntfs_write_mst_block(page, wbc); wbc 1412 fs/ntfs/aops.c return ntfs_write_block(page, wbc); wbc 1508 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); wbc 2654 fs/ntfs/super.c static int ntfs_write_inode(struct inode *vi, struct writeback_control *wbc) wbc 2656 fs/ntfs/super.c return __ntfs_write_inode(vi, wbc->sync_mode == WB_SYNC_ALL); wbc 412 fs/ocfs2/aops.c static int ocfs2_writepage(struct page *page, struct writeback_control *wbc) wbc 418 fs/ocfs2/aops.c return block_write_full_page(page, ocfs2_get_block, wbc); wbc 298 fs/omfs/file.c static int omfs_writepage(struct page *page, struct writeback_control *wbc) wbc 300 fs/omfs/file.c return block_write_full_page(page, omfs_get_block, wbc); wbc 304 fs/omfs/file.c omfs_writepages(struct address_space *mapping, struct writeback_control *wbc) wbc 306 fs/omfs/file.c return mpage_writepages(mapping, wbc, omfs_get_block); wbc 172 fs/omfs/inode.c static int omfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 174 fs/omfs/inode.c return __omfs_write_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 19 fs/orangefs/inode.c struct writeback_control *wbc) wbc 74 fs/orangefs/inode.c static int orangefs_writepage(struct page *page, struct writeback_control *wbc) wbc 77 fs/orangefs/inode.c ret = orangefs_writepage_locked(page, wbc); wbc 95 fs/orangefs/inode.c struct writeback_control *wbc) wbc 162 fs/orangefs/inode.c struct writeback_control *wbc, void *data) wbc 188 fs/orangefs/inode.c orangefs_writepages_work(ow, wbc); wbc 202 fs/orangefs/inode.c orangefs_writepages_work(ow, wbc); wbc 205 fs/orangefs/inode.c ret = orangefs_writepage_locked(page, wbc); wbc 211 fs/orangefs/inode.c orangefs_writepages_work(ow, wbc); wbc 219 fs/orangefs/inode.c struct writeback_control *wbc) wbc 240 fs/orangefs/inode.c ret = write_cache_pages(mapping, wbc, orangefs_writepages_callback, ow); wbc 242 fs/orangefs/inode.c ret = orangefs_writepages_work(ow, wbc); wbc 576 fs/orangefs/inode.c struct writeback_control wbc = { wbc 582 fs/orangefs/inode.c r = orangefs_writepage_locked(page, &wbc); wbc 153 fs/orangefs/super.c struct writeback_control *wbc) wbc 1775 fs/reiserfs/inode.c int reiserfs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1789 fs/reiserfs/inode.c if (wbc->sync_mode == WB_SYNC_ALL && !(current->flags & PF_MEMALLOC)) { wbc 2528 fs/reiserfs/inode.c struct writeback_control *wbc) wbc 2546 fs/reiserfs/inode.c redirty_page_for_writepage(wbc, page); wbc 2641 fs/reiserfs/inode.c if (wbc->sync_mode != WB_SYNC_NONE) { wbc 2645 fs/reiserfs/inode.c redirty_page_for_writepage(wbc, page); wbc 2748 fs/reiserfs/inode.c static int reiserfs_writepage(struct page *page, struct writeback_control *wbc) wbc 2752 fs/reiserfs/inode.c return reiserfs_write_full_page(page, wbc); wbc 3070 fs/reiserfs/reiserfs.h int reiserfs_write_inode(struct inode *inode, struct writeback_control *wbc); wbc 143 fs/sysv/ialloc.c struct writeback_control wbc = { wbc 175 fs/sysv/ialloc.c sysv_write_inode(inode, &wbc); /* ensure inode not allocated again */ wbc 281 fs/sysv/inode.c int sysv_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 283 fs/sysv/inode.c return __sysv_write_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 454 fs/sysv/itree.c static int sysv_writepage(struct page *page, struct writeback_control *wbc) wbc 456 fs/sysv/itree.c return block_write_full_page(page,get_block,wbc); wbc 141 fs/sysv/sysv.h extern int sysv_write_inode(struct inode *, struct writeback_control *wbc); wbc 1005 fs/ubifs/file.c static int ubifs_writepage(struct page *page, struct writeback_control *wbc) wbc 277 fs/ubifs/super.c static int ubifs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 70 fs/udf/file.c struct writeback_control *wbc) wbc 182 fs/udf/inode.c static int udf_writepage(struct page *page, struct writeback_control *wbc) wbc 184 fs/udf/inode.c return block_write_full_page(page, udf_get_block, wbc); wbc 188 fs/udf/inode.c struct writeback_control *wbc) wbc 190 fs/udf/inode.c return mpage_writepages(mapping, wbc, udf_get_block); wbc 1660 fs/udf/inode.c int udf_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 1662 fs/udf/inode.c return udf_update_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 156 fs/udf/udfdecl.h extern int udf_write_inode(struct inode *, struct writeback_control *wbc); wbc 470 fs/ufs/inode.c static int ufs_writepage(struct page *page, struct writeback_control *wbc) wbc 472 fs/ufs/inode.c return block_write_full_page(page,ufs_getfrag_block,wbc); wbc 836 fs/ufs/inode.c int ufs_write_inode(struct inode *inode, struct writeback_control *wbc) wbc 838 fs/ufs/inode.c return ufs_update_inode(inode, wbc->sync_mode == WB_SYNC_ALL); wbc 645 fs/xfs/xfs_aops.c struct writeback_control *wbc, wbc 701 fs/xfs/xfs_aops.c struct writeback_control *wbc) wbc 709 fs/xfs/xfs_aops.c bio->bi_opf = REQ_OP_WRITE | wbc_to_write_flags(wbc); wbc 711 fs/xfs/xfs_aops.c wbc_init_bio(wbc, bio); wbc 761 fs/xfs/xfs_aops.c struct writeback_control *wbc, wbc 783 fs/xfs/xfs_aops.c wpc->imap.br_state, offset, bdev, sector, wbc); wbc 799 fs/xfs/xfs_aops.c wbc_account_cgroup_owner(wbc, page, len); wbc 868 fs/xfs/xfs_aops.c struct writeback_control *wbc, wbc 899 fs/xfs/xfs_aops.c xfs_add_to_ioend(inode, file_offset, page, iop, wpc, wbc, wbc 955 fs/xfs/xfs_aops.c error2 = xfs_submit_ioend(wbc, ioend, error); wbc 981 fs/xfs/xfs_aops.c struct writeback_control *wbc, wbc 1079 fs/xfs/xfs_aops.c return xfs_writepage_map(wpc, wbc, inode, page, end_offset); wbc 1082 fs/xfs/xfs_aops.c redirty_page_for_writepage(wbc, page); wbc 1090 fs/xfs/xfs_aops.c struct writeback_control *wbc) wbc 1095 fs/xfs/xfs_aops.c ret = xfs_do_writepage(page, wbc, &wpc); wbc 1097 fs/xfs/xfs_aops.c ret = xfs_submit_ioend(wbc, wpc.ioend, ret); wbc 1104 fs/xfs/xfs_aops.c struct writeback_control *wbc) wbc 1110 fs/xfs/xfs_aops.c ret = write_cache_pages(mapping, wbc, xfs_do_writepage, &wpc); wbc 1112 fs/xfs/xfs_aops.c ret = xfs_submit_ioend(wbc, wpc.ioend, ret); wbc 1119 fs/xfs/xfs_aops.c struct writeback_control *wbc) wbc 1123 fs/xfs/xfs_aops.c xfs_find_bdev_for_inode(mapping->host), wbc); wbc 220 include/linux/buffer_head.h struct writeback_control *wbc); wbc 222 include/linux/buffer_head.h get_block_t *get_block, struct writeback_control *wbc, wbc 267 include/linux/buffer_head.h struct writeback_control *wbc); wbc 144 include/linux/dax.h struct block_device *bdev, struct writeback_control *wbc); wbc 183 include/linux/dax.h struct block_device *bdev, struct writeback_control *wbc) wbc 363 include/linux/fs.h int (*writepage)(struct page *page, struct writeback_control *wbc); wbc 1949 include/linux/fs.h int (*write_inode) (struct inode *, struct writeback_control *wbc); wbc 2215 include/linux/fs.h int sync_inode(struct inode *inode, struct writeback_control *wbc); wbc 1606 include/linux/mm.h int redirty_page_for_writepage(struct writeback_control *wbc, wbc 21 include/linux/mpage.h struct writeback_control *wbc, get_block_t get_block); wbc 23 include/linux/mpage.h struct writeback_control *wbc); wbc 534 include/linux/nfs_fs.h extern int nfs_writepage(struct page *page, struct writeback_control *wbc); wbc 389 include/linux/swap.h extern int swap_writepage(struct page *page, struct writeback_control *wbc); wbc 391 include/linux/swap.h extern int __swap_writepage(struct page *page, struct writeback_control *wbc, wbc 550 include/linux/swap.h static inline int swap_writepage(struct page *p, struct writeback_control *wbc) wbc 97 include/linux/writeback.h static inline int wbc_to_write_flags(struct writeback_control *wbc) wbc 101 include/linux/writeback.h if (wbc->punt_to_cgroup) wbc 104 include/linux/writeback.h if (wbc->sync_mode == WB_SYNC_ALL) wbc 106 include/linux/writeback.h else if (wbc->for_kupdate || wbc->for_background) wbc 113 include/linux/writeback.h wbc_blkcg_css(struct writeback_control *wbc) wbc 116 include/linux/writeback.h if (wbc->wb) wbc 117 include/linux/writeback.h return wbc->wb->blkcg_css; wbc 214 include/linux/writeback.h void wbc_attach_and_unlock_inode(struct writeback_control *wbc, wbc 217 include/linux/writeback.h void wbc_detach_inode(struct writeback_control *wbc); wbc 218 include/linux/writeback.h void wbc_account_cgroup_owner(struct writeback_control *wbc, struct page *page, wbc 263 include/linux/writeback.h static inline void wbc_attach_fdatawrite_inode(struct writeback_control *wbc, wbc 268 include/linux/writeback.h wbc_attach_and_unlock_inode(wbc, inode); wbc 281 include/linux/writeback.h static inline void wbc_init_bio(struct writeback_control *wbc, struct bio *bio) wbc 289 include/linux/writeback.h if (wbc->wb) wbc 290 include/linux/writeback.h bio_associate_blkg_from_css(bio, wbc->wb->blkcg_css); wbc 303 include/linux/writeback.h static inline void wbc_attach_and_unlock_inode(struct writeback_control *wbc, wbc 310 include/linux/writeback.h static inline void wbc_attach_fdatawrite_inode(struct writeback_control *wbc, wbc 315 include/linux/writeback.h static inline void wbc_detach_inode(struct writeback_control *wbc) wbc 319 include/linux/writeback.h static inline void wbc_init_bio(struct writeback_control *wbc, struct bio *bio) wbc 323 include/linux/writeback.h static inline void wbc_account_cgroup_owner(struct writeback_control *wbc, wbc 391 include/linux/writeback.h typedef int (*writepage_t)(struct page *page, struct writeback_control *wbc, wbc 395 include/linux/writeback.h struct writeback_control *wbc); wbc 399 include/linux/writeback.h struct writeback_control *wbc, writepage_t writepage, wbc 401 include/linux/writeback.h int do_writepages(struct address_space *mapping, struct writeback_control *wbc); wbc 563 include/trace/events/btrfs.h const struct writeback_control *wbc), wbc 565 include/trace/events/btrfs.h TP_ARGS(page, inode, wbc), wbc 584 include/trace/events/btrfs.h __entry->nr_to_write = wbc->nr_to_write; wbc 585 include/trace/events/btrfs.h __entry->pages_skipped = wbc->pages_skipped; wbc 586 include/trace/events/btrfs.h __entry->range_start = wbc->range_start; wbc 587 include/trace/events/btrfs.h __entry->range_end = wbc->range_end; wbc 588 include/trace/events/btrfs.h __entry->for_kupdate = wbc->for_kupdate; wbc 589 include/trace/events/btrfs.h __entry->for_reclaim = wbc->for_reclaim; wbc 590 include/trace/events/btrfs.h __entry->range_cyclic = wbc->range_cyclic; wbc 612 include/trace/events/btrfs.h const struct writeback_control *wbc), wbc 614 include/trace/events/btrfs.h TP_ARGS(page, inode, wbc) wbc 392 include/trace/events/ext4.h TP_PROTO(struct inode *inode, struct writeback_control *wbc), wbc 394 include/trace/events/ext4.h TP_ARGS(inode, wbc), wbc 412 include/trace/events/ext4.h __entry->nr_to_write = wbc->nr_to_write; wbc 413 include/trace/events/ext4.h __entry->pages_skipped = wbc->pages_skipped; wbc 414 include/trace/events/ext4.h __entry->range_start = wbc->range_start; wbc 415 include/trace/events/ext4.h __entry->range_end = wbc->range_end; wbc 417 include/trace/events/ext4.h __entry->sync_mode = wbc->sync_mode; wbc 418 include/trace/events/ext4.h __entry->for_kupdate = wbc->for_kupdate; wbc 419 include/trace/events/ext4.h __entry->range_cyclic = wbc->range_cyclic; wbc 435 include/trace/events/ext4.h struct writeback_control *wbc), wbc 437 include/trace/events/ext4.h TP_ARGS(inode, first_page, wbc), wbc 451 include/trace/events/ext4.h __entry->nr_to_write = wbc->nr_to_write; wbc 452 include/trace/events/ext4.h __entry->sync_mode = wbc->sync_mode; wbc 490 include/trace/events/ext4.h TP_PROTO(struct inode *inode, struct writeback_control *wbc, wbc 493 include/trace/events/ext4.h TP_ARGS(inode, wbc, ret, pages_written), wbc 510 include/trace/events/ext4.h __entry->pages_skipped = wbc->pages_skipped; wbc 512 include/trace/events/ext4.h __entry->sync_mode = wbc->sync_mode; wbc 1307 include/trace/events/f2fs.h TP_PROTO(struct inode *inode, struct writeback_control *wbc, int type), wbc 1309 include/trace/events/f2fs.h TP_ARGS(inode, wbc, type), wbc 1335 include/trace/events/f2fs.h __entry->nr_to_write = wbc->nr_to_write; wbc 1336 include/trace/events/f2fs.h __entry->pages_skipped = wbc->pages_skipped; wbc 1337 include/trace/events/f2fs.h __entry->range_start = wbc->range_start; wbc 1338 include/trace/events/f2fs.h __entry->range_end = wbc->range_end; wbc 1340 include/trace/events/f2fs.h __entry->sync_mode = wbc->sync_mode; wbc 1341 include/trace/events/f2fs.h __entry->for_kupdate = wbc->for_kupdate; wbc 1342 include/trace/events/f2fs.h __entry->for_background = wbc->for_background; wbc 1343 include/trace/events/f2fs.h __entry->tagged_writepages = wbc->tagged_writepages; wbc 1344 include/trace/events/f2fs.h __entry->for_reclaim = wbc->for_reclaim; wbc 1345 include/trace/events/f2fs.h __entry->range_cyclic = wbc->range_cyclic; wbc 1346 include/trace/events/f2fs.h __entry->for_sync = wbc->for_sync; wbc 157 include/trace/events/writeback.h static inline unsigned int __trace_wbc_assign_cgroup(struct writeback_control *wbc) wbc 159 include/trace/events/writeback.h if (wbc->wb) wbc 160 include/trace/events/writeback.h return __trace_wb_assign_cgroup(wbc->wb); wbc 171 include/trace/events/writeback.h static inline unsigned int __trace_wbc_assign_cgroup(struct writeback_control *wbc) wbc 182 include/trace/events/writeback.h TP_PROTO(struct inode *inode, struct writeback_control *wbc, wbc 185 include/trace/events/writeback.h TP_ARGS(inode, wbc, history), wbc 197 include/trace/events/writeback.h __entry->cgroup_ino = __trace_wbc_assign_cgroup(wbc); wbc 307 include/trace/events/writeback.h TP_PROTO(struct inode *inode, struct writeback_control *wbc), wbc 309 include/trace/events/writeback.h TP_ARGS(inode, wbc), wbc 322 include/trace/events/writeback.h __entry->sync_mode = wbc->sync_mode; wbc 323 include/trace/events/writeback.h __entry->cgroup_ino = __trace_wbc_assign_cgroup(wbc); wbc 336 include/trace/events/writeback.h TP_PROTO(struct inode *inode, struct writeback_control *wbc), wbc 338 include/trace/events/writeback.h TP_ARGS(inode, wbc) wbc 343 include/trace/events/writeback.h TP_PROTO(struct inode *inode, struct writeback_control *wbc), wbc 345 include/trace/events/writeback.h TP_ARGS(inode, wbc) wbc 446 include/trace/events/writeback.h TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi), wbc 447 include/trace/events/writeback.h TP_ARGS(wbc, bdi), wbc 464 include/trace/events/writeback.h __entry->nr_to_write = wbc->nr_to_write; wbc 465 include/trace/events/writeback.h __entry->pages_skipped = wbc->pages_skipped; wbc 466 include/trace/events/writeback.h __entry->sync_mode = wbc->sync_mode; wbc 467 include/trace/events/writeback.h __entry->for_kupdate = wbc->for_kupdate; wbc 468 include/trace/events/writeback.h __entry->for_background = wbc->for_background; wbc 469 include/trace/events/writeback.h __entry->for_reclaim = wbc->for_reclaim; wbc 470 include/trace/events/writeback.h __entry->range_cyclic = wbc->range_cyclic; wbc 471 include/trace/events/writeback.h __entry->range_start = (long)wbc->range_start; wbc 472 include/trace/events/writeback.h __entry->range_end = (long)wbc->range_end; wbc 473 include/trace/events/writeback.h __entry->cgroup_ino = __trace_wbc_assign_cgroup(wbc); wbc 495 include/trace/events/writeback.h TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi), \ wbc 496 include/trace/events/writeback.h TP_ARGS(wbc, bdi)) wbc 781 include/trace/events/writeback.h struct writeback_control *wbc, wbc 785 include/trace/events/writeback.h TP_ARGS(inode, wbc, nr_to_write), wbc 806 include/trace/events/writeback.h __entry->wrote = nr_to_write - wbc->nr_to_write; wbc 807 include/trace/events/writeback.h __entry->cgroup_ino = __trace_wbc_assign_cgroup(wbc); wbc 826 include/trace/events/writeback.h struct writeback_control *wbc, wbc 828 include/trace/events/writeback.h TP_ARGS(inode, wbc, nr_to_write) wbc 833 include/trace/events/writeback.h struct writeback_control *wbc, wbc 835 include/trace/events/writeback.h TP_ARGS(inode, wbc, nr_to_write) wbc 409 mm/filemap.c struct writeback_control wbc = { wbc 420 mm/filemap.c wbc_attach_fdatawrite_inode(&wbc, mapping->host); wbc 421 mm/filemap.c ret = do_writepages(mapping, &wbc); wbc 422 mm/filemap.c wbc_detach_inode(&wbc); wbc 851 mm/migrate.c struct writeback_control wbc = { wbc 878 mm/migrate.c rc = mapping->a_ops->writepage(page, &wbc); wbc 2159 mm/page-writeback.c struct writeback_control *wbc, writepage_t writepage, wbc 2175 mm/page-writeback.c if (wbc->range_cyclic) { wbc 2180 mm/page-writeback.c index = wbc->range_start >> PAGE_SHIFT; wbc 2181 mm/page-writeback.c end = wbc->range_end >> PAGE_SHIFT; wbc 2182 mm/page-writeback.c if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) wbc 2185 mm/page-writeback.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 2189 mm/page-writeback.c if (wbc->sync_mode == WB_SYNC_ALL || wbc->tagged_writepages) wbc 2227 mm/page-writeback.c if (wbc->sync_mode != WB_SYNC_NONE) wbc 2237 mm/page-writeback.c trace_wbc_writepage(wbc, inode_to_bdi(mapping->host)); wbc 2238 mm/page-writeback.c error = (*writepage)(page, wbc, data); wbc 2255 mm/page-writeback.c } else if (wbc->sync_mode != WB_SYNC_ALL) { wbc 2271 mm/page-writeback.c if (--wbc->nr_to_write <= 0 && wbc 2272 mm/page-writeback.c wbc->sync_mode == WB_SYNC_NONE) { wbc 2286 mm/page-writeback.c if (wbc->range_cyclic && !done) wbc 2288 mm/page-writeback.c if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0)) wbc 2299 mm/page-writeback.c static int __writepage(struct page *page, struct writeback_control *wbc, wbc 2303 mm/page-writeback.c int ret = mapping->a_ops->writepage(page, wbc); wbc 2319 mm/page-writeback.c struct writeback_control *wbc) wbc 2329 mm/page-writeback.c ret = write_cache_pages(mapping, wbc, __writepage, mapping); wbc 2336 mm/page-writeback.c int do_writepages(struct address_space *mapping, struct writeback_control *wbc) wbc 2340 mm/page-writeback.c if (wbc->nr_to_write <= 0) wbc 2344 mm/page-writeback.c ret = mapping->a_ops->writepages(mapping, wbc); wbc 2346 mm/page-writeback.c ret = generic_writepages(mapping, wbc); wbc 2347 mm/page-writeback.c if ((ret != -ENOMEM) || (wbc->sync_mode != WB_SYNC_ALL)) wbc 2370 mm/page-writeback.c struct writeback_control wbc = { wbc 2381 mm/page-writeback.c ret = mapping->a_ops->writepage(page, &wbc); wbc 2527 mm/page-writeback.c int redirty_page_for_writepage(struct writeback_control *wbc, struct page *page) wbc 2531 mm/page-writeback.c wbc->pages_skipped++; wbc 244 mm/page_io.c int swap_writepage(struct page *page, struct writeback_control *wbc) wbc 258 mm/page_io.c ret = __swap_writepage(page, wbc, end_swap_bio_write); wbc 277 mm/page_io.c int __swap_writepage(struct page *page, struct writeback_control *wbc, wbc 326 mm/page_io.c ret = bdev_write_page(sis->bdev, swap_page_sector(page), page, wbc); wbc 340 mm/page_io.c bio->bi_opf = REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc); wbc 1293 mm/shmem.c static int shmem_writepage(struct page *page, struct writeback_control *wbc) wbc 1319 mm/shmem.c if (!wbc->for_reclaim) { wbc 1383 mm/shmem.c swap_writepage(page, wbc); wbc 1391 mm/shmem.c if (wbc->for_reclaim) wbc 868 mm/vmscan.c struct writeback_control wbc = { wbc 877 mm/vmscan.c res = mapping->a_ops->writepage(page, &wbc); wbc 852 mm/zswap.c struct writeback_control wbc = { wbc 907 mm/zswap.c __swap_writepage(page, &wbc, end_swap_bio_write);