Loading fs/ext4/page-io.c +3 −4 Original line number Diff line number Diff line Loading @@ -371,10 +371,9 @@ void ext4_io_submit(struct ext4_io_submit *io) struct bio *bio = io->io_bio; if (bio) { int io_op_flags = io->io_wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0; if (io->io_wbc->sync_mode == WB_SYNC_ALL) io->io_bio->bi_opf |= REQ_SYNC; io->io_bio->bi_write_hint = io->io_end->inode->i_write_hint; bio_set_op_attrs(io->io_bio, REQ_OP_WRITE, io_op_flags); submit_bio(io->io_bio); } io->io_bio = NULL; Loading @@ -397,7 +396,7 @@ static void io_submit_init_bio(struct ext4_io_submit *io, * bio_alloc will _always_ be able to allocate a bio if * __GFP_DIRECT_RECLAIM is set, see comments for bio_alloc_bioset(). */ bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, 0, GFP_NOIO); bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, REQ_OP_WRITE, GFP_NOIO); fscrypt_set_bio_crypt_ctx_bh(bio, bh, GFP_NOIO); bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); bio->bi_end_io = ext4_end_bio; Loading fs/mpage.c +21 −29 Original line number Diff line number Diff line Loading @@ -57,10 +57,9 @@ static void mpage_end_io(struct bio *bio) bio_put(bio); } static struct bio *mpage_bio_submit(int op, int op_flags, struct bio *bio) static struct bio *mpage_bio_submit(struct bio *bio) { bio->bi_end_io = mpage_end_io; bio_set_op_attrs(bio, op, op_flags); guard_bio_eod(bio); submit_bio(bio); return NULL; Loading Loading @@ -146,16 +145,15 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) struct block_device *bdev = NULL; int length; int fully_mapped = 1; int op_flags; int op = REQ_OP_READ; unsigned nblocks; unsigned relative_block; gfp_t gfp; if (args->is_readahead) { op_flags = REQ_RAHEAD; op |= REQ_RAHEAD; gfp = readahead_gfp_mask(page->mapping); } else { op_flags = 0; gfp = mapping_gfp_constraint(page->mapping, GFP_KERNEL); } Loading Loading @@ -264,7 +262,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) * This page will go to BIO. Do we need to send this BIO off first? */ if (args->bio && (args->last_block_in_bio != blocks[0] - 1)) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); alloc_new: if (args->bio == NULL) { Loading @@ -273,7 +271,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) page)) goto out; } args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), 0, args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), op, gfp); if (args->bio == NULL) goto confused; Loading @@ -282,7 +280,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) length = first_hole << blkbits; if (bio_add_page(args->bio, page, length, 0) < length) { args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); goto alloc_new; } Loading @@ -290,7 +288,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) nblocks = map_bh->b_size >> blkbits; if ((buffer_boundary(map_bh) && relative_block == nblocks) || (first_hole != blocks_per_page)) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); else args->last_block_in_bio = blocks[blocks_per_page - 1]; out: Loading @@ -298,7 +296,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) confused: if (args->bio) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); if (!PageUptodate(page)) block_read_full_page(page, args->get_block); else Loading Loading @@ -361,7 +359,7 @@ void mpage_readahead(struct readahead_control *rac, get_block_t get_block) put_page(page); } if (args.bio) mpage_bio_submit(REQ_OP_READ, REQ_RAHEAD, args.bio); mpage_bio_submit(args.bio); } EXPORT_SYMBOL(mpage_readahead); Loading @@ -378,7 +376,7 @@ int mpage_readpage(struct page *page, get_block_t get_block) args.bio = do_mpage_readpage(&args); if (args.bio) mpage_bio_submit(REQ_OP_READ, 0, args.bio); mpage_bio_submit(args.bio); return 0; } EXPORT_SYMBOL(mpage_readpage); Loading Loading @@ -469,7 +467,6 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, struct buffer_head map_bh; loff_t i_size = i_size_read(inode); int ret = 0; int op_flags = wbc_to_write_flags(wbc); if (page_has_buffers(page)) { struct buffer_head *head = page_buffers(page); Loading Loading @@ -577,7 +574,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, * This page will go to BIO. Do we need to send this BIO off first? */ if (bio && mpd->last_block_in_bio != blocks[0] - 1) bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); alloc_new: if (bio == NULL) { Loading @@ -586,9 +583,10 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, page, wbc)) goto out; } bio = bio_alloc(bdev, BIO_MAX_VECS, 0, GFP_NOFS); bio = bio_alloc(bdev, BIO_MAX_VECS, REQ_OP_WRITE | wbc_to_write_flags(wbc), GFP_NOFS); bio->bi_iter.bi_sector = blocks[0] << (blkbits - 9); wbc_init_bio(wbc, bio); bio->bi_write_hint = inode->i_write_hint; } Loading @@ -601,7 +599,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, wbc_account_cgroup_owner(wbc, page, PAGE_SIZE); length = first_unmapped << blkbits; if (bio_add_page(bio, page, length, 0) < length) { bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); goto alloc_new; } Loading @@ -611,7 +609,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, set_page_writeback(page); unlock_page(page); if (boundary || (first_unmapped != blocks_per_page)) { bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); if (boundary_block) { write_boundary_block(boundary_bdev, boundary_block, 1 << blkbits); Loading @@ -623,7 +621,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, confused: if (bio) bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); if (mpd->use_writepage) { ret = mapping->a_ops->writepage(page, wbc); Loading Loading @@ -679,11 +677,8 @@ mpage_writepages(struct address_space *mapping, }; ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd); if (mpd.bio) { int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0); mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio); } if (mpd.bio) mpage_bio_submit(mpd.bio); } blk_finish_plug(&plug); return ret; Loading @@ -700,11 +695,8 @@ int mpage_writepage(struct page *page, get_block_t get_block, .use_writepage = 0, }; int ret = __mpage_writepage(page, wbc, &mpd); if (mpd.bio) { int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0); mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio); } if (mpd.bio) mpage_bio_submit(mpd.bio); return ret; } EXPORT_SYMBOL(mpage_writepage); fs/nilfs2/segbuf.c +9 −11 Original line number Diff line number Diff line Loading @@ -337,8 +337,7 @@ static void nilfs_end_bio_write(struct bio *bio) } static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf, struct nilfs_write_info *wi, int mode, int mode_flags) struct nilfs_write_info *wi) { struct bio *bio = wi->bio; int err; Loading @@ -356,7 +355,6 @@ static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf, bio->bi_end_io = nilfs_end_bio_write; bio->bi_private = segbuf; bio_set_op_attrs(bio, mode, mode_flags); submit_bio(bio); segbuf->sb_nbio++; Loading Loading @@ -384,15 +382,15 @@ static void nilfs_segbuf_prepare_write(struct nilfs_segment_buffer *segbuf, static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf, struct nilfs_write_info *wi, struct buffer_head *bh, int mode) struct buffer_head *bh) { int len, err; BUG_ON(wi->nr_vecs <= 0); repeat: if (!wi->bio) { wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs, 0, GFP_NOIO); wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs, REQ_OP_WRITE, GFP_NOIO); wi->bio->bi_iter.bi_sector = (wi->blocknr + wi->end) << (wi->nilfs->ns_blocksize_bits - 9); } Loading @@ -403,7 +401,7 @@ static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf, return 0; } /* bio is FULL */ err = nilfs_segbuf_submit_bio(segbuf, wi, mode, 0); err = nilfs_segbuf_submit_bio(segbuf, wi); /* never submit current bh */ if (likely(!err)) goto repeat; Loading Loading @@ -433,13 +431,13 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf, nilfs_segbuf_prepare_write(segbuf, &wi); list_for_each_entry(bh, &segbuf->sb_segsum_buffers, b_assoc_buffers) { res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); res = nilfs_segbuf_submit_bh(segbuf, &wi, bh); if (unlikely(res)) goto failed_bio; } list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) { res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); res = nilfs_segbuf_submit_bh(segbuf, &wi, bh); if (unlikely(res)) goto failed_bio; } Loading @@ -449,8 +447,8 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf, * Last BIO is always sent through the following * submission. */ res = nilfs_segbuf_submit_bio(segbuf, &wi, REQ_OP_WRITE, REQ_SYNC); wi.bio->bi_opf |= REQ_SYNC; res = nilfs_segbuf_submit_bio(segbuf, &wi); } failed_bio: Loading Loading
fs/ext4/page-io.c +3 −4 Original line number Diff line number Diff line Loading @@ -371,10 +371,9 @@ void ext4_io_submit(struct ext4_io_submit *io) struct bio *bio = io->io_bio; if (bio) { int io_op_flags = io->io_wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0; if (io->io_wbc->sync_mode == WB_SYNC_ALL) io->io_bio->bi_opf |= REQ_SYNC; io->io_bio->bi_write_hint = io->io_end->inode->i_write_hint; bio_set_op_attrs(io->io_bio, REQ_OP_WRITE, io_op_flags); submit_bio(io->io_bio); } io->io_bio = NULL; Loading @@ -397,7 +396,7 @@ static void io_submit_init_bio(struct ext4_io_submit *io, * bio_alloc will _always_ be able to allocate a bio if * __GFP_DIRECT_RECLAIM is set, see comments for bio_alloc_bioset(). */ bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, 0, GFP_NOIO); bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, REQ_OP_WRITE, GFP_NOIO); fscrypt_set_bio_crypt_ctx_bh(bio, bh, GFP_NOIO); bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); bio->bi_end_io = ext4_end_bio; Loading
fs/mpage.c +21 −29 Original line number Diff line number Diff line Loading @@ -57,10 +57,9 @@ static void mpage_end_io(struct bio *bio) bio_put(bio); } static struct bio *mpage_bio_submit(int op, int op_flags, struct bio *bio) static struct bio *mpage_bio_submit(struct bio *bio) { bio->bi_end_io = mpage_end_io; bio_set_op_attrs(bio, op, op_flags); guard_bio_eod(bio); submit_bio(bio); return NULL; Loading Loading @@ -146,16 +145,15 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) struct block_device *bdev = NULL; int length; int fully_mapped = 1; int op_flags; int op = REQ_OP_READ; unsigned nblocks; unsigned relative_block; gfp_t gfp; if (args->is_readahead) { op_flags = REQ_RAHEAD; op |= REQ_RAHEAD; gfp = readahead_gfp_mask(page->mapping); } else { op_flags = 0; gfp = mapping_gfp_constraint(page->mapping, GFP_KERNEL); } Loading Loading @@ -264,7 +262,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) * This page will go to BIO. Do we need to send this BIO off first? */ if (args->bio && (args->last_block_in_bio != blocks[0] - 1)) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); alloc_new: if (args->bio == NULL) { Loading @@ -273,7 +271,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) page)) goto out; } args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), 0, args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), op, gfp); if (args->bio == NULL) goto confused; Loading @@ -282,7 +280,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) length = first_hole << blkbits; if (bio_add_page(args->bio, page, length, 0) < length) { args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); goto alloc_new; } Loading @@ -290,7 +288,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) nblocks = map_bh->b_size >> blkbits; if ((buffer_boundary(map_bh) && relative_block == nblocks) || (first_hole != blocks_per_page)) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); else args->last_block_in_bio = blocks[blocks_per_page - 1]; out: Loading @@ -298,7 +296,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args) confused: if (args->bio) args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio); args->bio = mpage_bio_submit(args->bio); if (!PageUptodate(page)) block_read_full_page(page, args->get_block); else Loading Loading @@ -361,7 +359,7 @@ void mpage_readahead(struct readahead_control *rac, get_block_t get_block) put_page(page); } if (args.bio) mpage_bio_submit(REQ_OP_READ, REQ_RAHEAD, args.bio); mpage_bio_submit(args.bio); } EXPORT_SYMBOL(mpage_readahead); Loading @@ -378,7 +376,7 @@ int mpage_readpage(struct page *page, get_block_t get_block) args.bio = do_mpage_readpage(&args); if (args.bio) mpage_bio_submit(REQ_OP_READ, 0, args.bio); mpage_bio_submit(args.bio); return 0; } EXPORT_SYMBOL(mpage_readpage); Loading Loading @@ -469,7 +467,6 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, struct buffer_head map_bh; loff_t i_size = i_size_read(inode); int ret = 0; int op_flags = wbc_to_write_flags(wbc); if (page_has_buffers(page)) { struct buffer_head *head = page_buffers(page); Loading Loading @@ -577,7 +574,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, * This page will go to BIO. Do we need to send this BIO off first? */ if (bio && mpd->last_block_in_bio != blocks[0] - 1) bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); alloc_new: if (bio == NULL) { Loading @@ -586,9 +583,10 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, page, wbc)) goto out; } bio = bio_alloc(bdev, BIO_MAX_VECS, 0, GFP_NOFS); bio = bio_alloc(bdev, BIO_MAX_VECS, REQ_OP_WRITE | wbc_to_write_flags(wbc), GFP_NOFS); bio->bi_iter.bi_sector = blocks[0] << (blkbits - 9); wbc_init_bio(wbc, bio); bio->bi_write_hint = inode->i_write_hint; } Loading @@ -601,7 +599,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, wbc_account_cgroup_owner(wbc, page, PAGE_SIZE); length = first_unmapped << blkbits; if (bio_add_page(bio, page, length, 0) < length) { bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); goto alloc_new; } Loading @@ -611,7 +609,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, set_page_writeback(page); unlock_page(page); if (boundary || (first_unmapped != blocks_per_page)) { bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); if (boundary_block) { write_boundary_block(boundary_bdev, boundary_block, 1 << blkbits); Loading @@ -623,7 +621,7 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc, confused: if (bio) bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio); bio = mpage_bio_submit(bio); if (mpd->use_writepage) { ret = mapping->a_ops->writepage(page, wbc); Loading Loading @@ -679,11 +677,8 @@ mpage_writepages(struct address_space *mapping, }; ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd); if (mpd.bio) { int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0); mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio); } if (mpd.bio) mpage_bio_submit(mpd.bio); } blk_finish_plug(&plug); return ret; Loading @@ -700,11 +695,8 @@ int mpage_writepage(struct page *page, get_block_t get_block, .use_writepage = 0, }; int ret = __mpage_writepage(page, wbc, &mpd); if (mpd.bio) { int op_flags = (wbc->sync_mode == WB_SYNC_ALL ? REQ_SYNC : 0); mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio); } if (mpd.bio) mpage_bio_submit(mpd.bio); return ret; } EXPORT_SYMBOL(mpage_writepage);
fs/nilfs2/segbuf.c +9 −11 Original line number Diff line number Diff line Loading @@ -337,8 +337,7 @@ static void nilfs_end_bio_write(struct bio *bio) } static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf, struct nilfs_write_info *wi, int mode, int mode_flags) struct nilfs_write_info *wi) { struct bio *bio = wi->bio; int err; Loading @@ -356,7 +355,6 @@ static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf, bio->bi_end_io = nilfs_end_bio_write; bio->bi_private = segbuf; bio_set_op_attrs(bio, mode, mode_flags); submit_bio(bio); segbuf->sb_nbio++; Loading Loading @@ -384,15 +382,15 @@ static void nilfs_segbuf_prepare_write(struct nilfs_segment_buffer *segbuf, static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf, struct nilfs_write_info *wi, struct buffer_head *bh, int mode) struct buffer_head *bh) { int len, err; BUG_ON(wi->nr_vecs <= 0); repeat: if (!wi->bio) { wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs, 0, GFP_NOIO); wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs, REQ_OP_WRITE, GFP_NOIO); wi->bio->bi_iter.bi_sector = (wi->blocknr + wi->end) << (wi->nilfs->ns_blocksize_bits - 9); } Loading @@ -403,7 +401,7 @@ static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf, return 0; } /* bio is FULL */ err = nilfs_segbuf_submit_bio(segbuf, wi, mode, 0); err = nilfs_segbuf_submit_bio(segbuf, wi); /* never submit current bh */ if (likely(!err)) goto repeat; Loading Loading @@ -433,13 +431,13 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf, nilfs_segbuf_prepare_write(segbuf, &wi); list_for_each_entry(bh, &segbuf->sb_segsum_buffers, b_assoc_buffers) { res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); res = nilfs_segbuf_submit_bh(segbuf, &wi, bh); if (unlikely(res)) goto failed_bio; } list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) { res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE); res = nilfs_segbuf_submit_bh(segbuf, &wi, bh); if (unlikely(res)) goto failed_bio; } Loading @@ -449,8 +447,8 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf, * Last BIO is always sent through the following * submission. */ res = nilfs_segbuf_submit_bio(segbuf, &wi, REQ_OP_WRITE, REQ_SYNC); wi.bio->bi_opf |= REQ_SYNC; res = nilfs_segbuf_submit_bio(segbuf, &wi); } failed_bio: Loading