mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 06:31:52 +00:00
md: pass mddev to make_request functions rather than request_queue
We used to pass the personality make_request function direct to the block layer so the first argument had to be a queue. But now we have the intermediary md_make_request so it makes at lot more sense to pass a struct mddev_s. It makes it possible to have an mddev without its own queue too. Signed-off-by: NeilBrown <neilb@suse.de>
This commit is contained in:
parent
cca9cf90c5
commit
21a52c6d05
@ -168,9 +168,8 @@ static void add_sector(conf_t *conf, sector_t start, int mode)
|
||||
conf->nfaults = n+1;
|
||||
}
|
||||
|
||||
static int make_request(struct request_queue *q, struct bio *bio)
|
||||
static int make_request(mddev_t *mddev, struct bio *bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
conf_t *conf = mddev->private;
|
||||
int failit = 0;
|
||||
|
||||
|
@ -286,9 +286,8 @@ static int linear_stop (mddev_t *mddev)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int linear_make_request (struct request_queue *q, struct bio *bio)
|
||||
static int linear_make_request (mddev_t *mddev, struct bio *bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
dev_info_t *tmp_dev;
|
||||
sector_t start_sector;
|
||||
|
||||
@ -328,9 +327,9 @@ static int linear_make_request (struct request_queue *q, struct bio *bio)
|
||||
|
||||
bp = bio_split(bio, end_sector - bio->bi_sector);
|
||||
|
||||
if (linear_make_request(q, &bp->bio1))
|
||||
if (linear_make_request(mddev, &bp->bio1))
|
||||
generic_make_request(&bp->bio1);
|
||||
if (linear_make_request(q, &bp->bio2))
|
||||
if (linear_make_request(mddev, &bp->bio2))
|
||||
generic_make_request(&bp->bio2);
|
||||
bio_pair_release(bp);
|
||||
return 0;
|
||||
|
@ -240,7 +240,7 @@ static int md_make_request(struct request_queue *q, struct bio *bio)
|
||||
atomic_inc(&mddev->active_io);
|
||||
rcu_read_unlock();
|
||||
|
||||
rv = mddev->pers->make_request(q, bio);
|
||||
rv = mddev->pers->make_request(mddev, bio);
|
||||
|
||||
cpu = part_stat_lock();
|
||||
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
|
||||
@ -354,7 +354,7 @@ static void md_submit_barrier(struct work_struct *ws)
|
||||
bio_endio(bio, 0);
|
||||
else {
|
||||
bio->bi_rw &= ~(1<<BIO_RW_BARRIER);
|
||||
if (mddev->pers->make_request(mddev->queue, bio))
|
||||
if (mddev->pers->make_request(mddev, bio))
|
||||
generic_make_request(bio);
|
||||
mddev->barrier = POST_REQUEST_BARRIER;
|
||||
submit_barriers(mddev);
|
||||
|
@ -330,7 +330,7 @@ struct mdk_personality
|
||||
int level;
|
||||
struct list_head list;
|
||||
struct module *owner;
|
||||
int (*make_request)(struct request_queue *q, struct bio *bio);
|
||||
int (*make_request)(mddev_t *mddev, struct bio *bio);
|
||||
int (*run)(mddev_t *mddev);
|
||||
int (*stop)(mddev_t *mddev);
|
||||
void (*status)(struct seq_file *seq, mddev_t *mddev);
|
||||
|
@ -135,9 +135,8 @@ static void multipath_unplug(struct request_queue *q)
|
||||
}
|
||||
|
||||
|
||||
static int multipath_make_request (struct request_queue *q, struct bio * bio)
|
||||
static int multipath_make_request(mddev_t *mddev, struct bio * bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
multipath_conf_t *conf = mddev->private;
|
||||
struct multipath_bh * mp_bh;
|
||||
struct multipath_info *multipath;
|
||||
|
@ -465,9 +465,8 @@ static inline int is_io_in_chunk_boundary(mddev_t *mddev,
|
||||
}
|
||||
}
|
||||
|
||||
static int raid0_make_request(struct request_queue *q, struct bio *bio)
|
||||
static int raid0_make_request(mddev_t *mddev, struct bio *bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
unsigned int chunk_sects;
|
||||
sector_t sector_offset;
|
||||
struct strip_zone *zone;
|
||||
@ -495,9 +494,9 @@ static int raid0_make_request(struct request_queue *q, struct bio *bio)
|
||||
else
|
||||
bp = bio_split(bio, chunk_sects -
|
||||
sector_div(sector, chunk_sects));
|
||||
if (raid0_make_request(q, &bp->bio1))
|
||||
if (raid0_make_request(mddev, &bp->bio1))
|
||||
generic_make_request(&bp->bio1);
|
||||
if (raid0_make_request(q, &bp->bio2))
|
||||
if (raid0_make_request(mddev, &bp->bio2))
|
||||
generic_make_request(&bp->bio2);
|
||||
|
||||
bio_pair_release(bp);
|
||||
|
@ -773,9 +773,8 @@ do_sync_io:
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int make_request(struct request_queue *q, struct bio * bio)
|
||||
static int make_request(mddev_t *mddev, struct bio * bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
conf_t *conf = mddev->private;
|
||||
mirror_info_t *mirror;
|
||||
r1bio_t *r1_bio;
|
||||
|
@ -788,9 +788,8 @@ static void unfreeze_array(conf_t *conf)
|
||||
spin_unlock_irq(&conf->resync_lock);
|
||||
}
|
||||
|
||||
static int make_request(struct request_queue *q, struct bio * bio)
|
||||
static int make_request(mddev_t *mddev, struct bio * bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
conf_t *conf = mddev->private;
|
||||
mirror_info_t *mirror;
|
||||
r10bio_t *r10_bio;
|
||||
@ -824,9 +823,9 @@ static int make_request(struct request_queue *q, struct bio * bio)
|
||||
*/
|
||||
bp = bio_split(bio,
|
||||
chunk_sects - (bio->bi_sector & (chunk_sects - 1)) );
|
||||
if (make_request(q, &bp->bio1))
|
||||
if (make_request(mddev, &bp->bio1))
|
||||
generic_make_request(&bp->bio1);
|
||||
if (make_request(q, &bp->bio2))
|
||||
if (make_request(mddev, &bp->bio2))
|
||||
generic_make_request(&bp->bio2);
|
||||
|
||||
bio_pair_release(bp);
|
||||
|
@ -3753,9 +3753,8 @@ static int bio_fits_rdev(struct bio *bi)
|
||||
}
|
||||
|
||||
|
||||
static int chunk_aligned_read(struct request_queue *q, struct bio * raid_bio)
|
||||
static int chunk_aligned_read(mddev_t *mddev, struct bio * raid_bio)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
raid5_conf_t *conf = mddev->private;
|
||||
int dd_idx;
|
||||
struct bio* align_bi;
|
||||
@ -3870,9 +3869,8 @@ static struct stripe_head *__get_priority_stripe(raid5_conf_t *conf)
|
||||
return sh;
|
||||
}
|
||||
|
||||
static int make_request(struct request_queue *q, struct bio * bi)
|
||||
static int make_request(mddev_t *mddev, struct bio * bi)
|
||||
{
|
||||
mddev_t *mddev = q->queuedata;
|
||||
raid5_conf_t *conf = mddev->private;
|
||||
int dd_idx;
|
||||
sector_t new_sector;
|
||||
@ -3896,7 +3894,7 @@ static int make_request(struct request_queue *q, struct bio * bi)
|
||||
|
||||
if (rw == READ &&
|
||||
mddev->reshape_position == MaxSector &&
|
||||
chunk_aligned_read(q,bi))
|
||||
chunk_aligned_read(mddev,bi))
|
||||
return 0;
|
||||
|
||||
logical_sector = bi->bi_sector & ~((sector_t)STRIPE_SECTORS-1);
|
||||
|
Loading…
Reference in New Issue
Block a user