mirror of
https://github.com/torvalds/linux.git
synced 2024-12-05 10:32:35 +00:00
0431d926b3
The new checkpoint cache flush mechanism requires us to issue an unconditional cache flush before we start a new checkpoint. We don't want to block for this if we can help it, and we have a fair chunk of CPU work to do between starting the checkpoint and issuing the first journal IO. Hence it makes sense to amortise the latency cost of the cache flush by issuing it asynchronously and then waiting for it only when we need to issue the first IO in the transaction. To do this, we need async cache flush primitives to submit the cache flush bio and to wait on it. The block layer has no such primitives for filesystems, so roll our own for the moment. Signed-off-by: Dave Chinner <dchinner@redhat.com> Reviewed-by: Brian Foster <bfoster@redhat.com> Reviewed-by: Darrick J. Wong <djwong@kernel.org> Reviewed-by: Allison Henderson <allison.henderson@oracle.com> Signed-off-by: Darrick J. Wong <djwong@kernel.org>
97 lines
2.2 KiB
C
97 lines
2.2 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2019 Christoph Hellwig.
|
|
*/
|
|
#include "xfs.h"
|
|
|
|
static inline unsigned int bio_max_vecs(unsigned int count)
|
|
{
|
|
return bio_max_segs(howmany(count, PAGE_SIZE));
|
|
}
|
|
|
|
static void
|
|
xfs_flush_bdev_async_endio(
|
|
struct bio *bio)
|
|
{
|
|
complete(bio->bi_private);
|
|
}
|
|
|
|
/*
|
|
* Submit a request for an async cache flush to run. If the request queue does
|
|
* not require flush operations, just skip it altogether. If the caller needs
|
|
* to wait for the flush completion at a later point in time, they must supply a
|
|
* valid completion. This will be signalled when the flush completes. The
|
|
* caller never sees the bio that is issued here.
|
|
*/
|
|
void
|
|
xfs_flush_bdev_async(
|
|
struct bio *bio,
|
|
struct block_device *bdev,
|
|
struct completion *done)
|
|
{
|
|
struct request_queue *q = bdev->bd_disk->queue;
|
|
|
|
if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags)) {
|
|
complete(done);
|
|
return;
|
|
}
|
|
|
|
bio_init(bio, NULL, 0);
|
|
bio_set_dev(bio, bdev);
|
|
bio->bi_opf = REQ_OP_WRITE | REQ_PREFLUSH | REQ_SYNC;
|
|
bio->bi_private = done;
|
|
bio->bi_end_io = xfs_flush_bdev_async_endio;
|
|
|
|
submit_bio(bio);
|
|
}
|
|
int
|
|
xfs_rw_bdev(
|
|
struct block_device *bdev,
|
|
sector_t sector,
|
|
unsigned int count,
|
|
char *data,
|
|
unsigned int op)
|
|
|
|
{
|
|
unsigned int is_vmalloc = is_vmalloc_addr(data);
|
|
unsigned int left = count;
|
|
int error;
|
|
struct bio *bio;
|
|
|
|
if (is_vmalloc && op == REQ_OP_WRITE)
|
|
flush_kernel_vmap_range(data, count);
|
|
|
|
bio = bio_alloc(GFP_KERNEL, bio_max_vecs(left));
|
|
bio_set_dev(bio, bdev);
|
|
bio->bi_iter.bi_sector = sector;
|
|
bio->bi_opf = op | REQ_META | REQ_SYNC;
|
|
|
|
do {
|
|
struct page *page = kmem_to_page(data);
|
|
unsigned int off = offset_in_page(data);
|
|
unsigned int len = min_t(unsigned, left, PAGE_SIZE - off);
|
|
|
|
while (bio_add_page(bio, page, len, off) != len) {
|
|
struct bio *prev = bio;
|
|
|
|
bio = bio_alloc(GFP_KERNEL, bio_max_vecs(left));
|
|
bio_copy_dev(bio, prev);
|
|
bio->bi_iter.bi_sector = bio_end_sector(prev);
|
|
bio->bi_opf = prev->bi_opf;
|
|
bio_chain(prev, bio);
|
|
|
|
submit_bio(prev);
|
|
}
|
|
|
|
data += len;
|
|
left -= len;
|
|
} while (left > 0);
|
|
|
|
error = submit_bio_wait(bio);
|
|
bio_put(bio);
|
|
|
|
if (is_vmalloc && op == REQ_OP_READ)
|
|
invalidate_kernel_vmap_range(data, count);
|
|
return error;
|
|
}
|