mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00
md: pass mddev to make_request functions rather than request_queue
We used to pass the personality make_request function direct to the block layer so the first argument had to be a queue. But now we have the intermediary md_make_request so it makes at lot more sense to pass a struct mddev_s. It makes it possible to have an mddev without its own queue too. Signed-off-by: NeilBrown <neilb@suse.de>
This commit is contained in:
parent
cca9cf90c5
commit
21a52c6d05
9 changed files with 18 additions and 26 deletions
|
@ -168,9 +168,8 @@ static void add_sector(conf_t *conf, sector_t start, int mode)
|
||||||
conf->nfaults = n+1;
|
conf->nfaults = n+1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int make_request(struct request_queue *q, struct bio *bio)
|
static int make_request(mddev_t *mddev, struct bio *bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
conf_t *conf = mddev->private;
|
conf_t *conf = mddev->private;
|
||||||
int failit = 0;
|
int failit = 0;
|
||||||
|
|
||||||
|
|
|
@ -286,9 +286,8 @@ static int linear_stop (mddev_t *mddev)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int linear_make_request (struct request_queue *q, struct bio *bio)
|
static int linear_make_request (mddev_t *mddev, struct bio *bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
dev_info_t *tmp_dev;
|
dev_info_t *tmp_dev;
|
||||||
sector_t start_sector;
|
sector_t start_sector;
|
||||||
|
|
||||||
|
@ -328,9 +327,9 @@ static int linear_make_request (struct request_queue *q, struct bio *bio)
|
||||||
|
|
||||||
bp = bio_split(bio, end_sector - bio->bi_sector);
|
bp = bio_split(bio, end_sector - bio->bi_sector);
|
||||||
|
|
||||||
if (linear_make_request(q, &bp->bio1))
|
if (linear_make_request(mddev, &bp->bio1))
|
||||||
generic_make_request(&bp->bio1);
|
generic_make_request(&bp->bio1);
|
||||||
if (linear_make_request(q, &bp->bio2))
|
if (linear_make_request(mddev, &bp->bio2))
|
||||||
generic_make_request(&bp->bio2);
|
generic_make_request(&bp->bio2);
|
||||||
bio_pair_release(bp);
|
bio_pair_release(bp);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -240,7 +240,7 @@ static int md_make_request(struct request_queue *q, struct bio *bio)
|
||||||
atomic_inc(&mddev->active_io);
|
atomic_inc(&mddev->active_io);
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
|
||||||
rv = mddev->pers->make_request(q, bio);
|
rv = mddev->pers->make_request(mddev, bio);
|
||||||
|
|
||||||
cpu = part_stat_lock();
|
cpu = part_stat_lock();
|
||||||
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
|
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
|
||||||
|
@ -354,7 +354,7 @@ static void md_submit_barrier(struct work_struct *ws)
|
||||||
bio_endio(bio, 0);
|
bio_endio(bio, 0);
|
||||||
else {
|
else {
|
||||||
bio->bi_rw &= ~(1<<BIO_RW_BARRIER);
|
bio->bi_rw &= ~(1<<BIO_RW_BARRIER);
|
||||||
if (mddev->pers->make_request(mddev->queue, bio))
|
if (mddev->pers->make_request(mddev, bio))
|
||||||
generic_make_request(bio);
|
generic_make_request(bio);
|
||||||
mddev->barrier = POST_REQUEST_BARRIER;
|
mddev->barrier = POST_REQUEST_BARRIER;
|
||||||
submit_barriers(mddev);
|
submit_barriers(mddev);
|
||||||
|
|
|
@ -330,7 +330,7 @@ struct mdk_personality
|
||||||
int level;
|
int level;
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
struct module *owner;
|
struct module *owner;
|
||||||
int (*make_request)(struct request_queue *q, struct bio *bio);
|
int (*make_request)(mddev_t *mddev, struct bio *bio);
|
||||||
int (*run)(mddev_t *mddev);
|
int (*run)(mddev_t *mddev);
|
||||||
int (*stop)(mddev_t *mddev);
|
int (*stop)(mddev_t *mddev);
|
||||||
void (*status)(struct seq_file *seq, mddev_t *mddev);
|
void (*status)(struct seq_file *seq, mddev_t *mddev);
|
||||||
|
|
|
@ -135,9 +135,8 @@ static void multipath_unplug(struct request_queue *q)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static int multipath_make_request (struct request_queue *q, struct bio * bio)
|
static int multipath_make_request(mddev_t *mddev, struct bio * bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
multipath_conf_t *conf = mddev->private;
|
multipath_conf_t *conf = mddev->private;
|
||||||
struct multipath_bh * mp_bh;
|
struct multipath_bh * mp_bh;
|
||||||
struct multipath_info *multipath;
|
struct multipath_info *multipath;
|
||||||
|
|
|
@ -465,9 +465,8 @@ static inline int is_io_in_chunk_boundary(mddev_t *mddev,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int raid0_make_request(struct request_queue *q, struct bio *bio)
|
static int raid0_make_request(mddev_t *mddev, struct bio *bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
unsigned int chunk_sects;
|
unsigned int chunk_sects;
|
||||||
sector_t sector_offset;
|
sector_t sector_offset;
|
||||||
struct strip_zone *zone;
|
struct strip_zone *zone;
|
||||||
|
@ -495,9 +494,9 @@ static int raid0_make_request(struct request_queue *q, struct bio *bio)
|
||||||
else
|
else
|
||||||
bp = bio_split(bio, chunk_sects -
|
bp = bio_split(bio, chunk_sects -
|
||||||
sector_div(sector, chunk_sects));
|
sector_div(sector, chunk_sects));
|
||||||
if (raid0_make_request(q, &bp->bio1))
|
if (raid0_make_request(mddev, &bp->bio1))
|
||||||
generic_make_request(&bp->bio1);
|
generic_make_request(&bp->bio1);
|
||||||
if (raid0_make_request(q, &bp->bio2))
|
if (raid0_make_request(mddev, &bp->bio2))
|
||||||
generic_make_request(&bp->bio2);
|
generic_make_request(&bp->bio2);
|
||||||
|
|
||||||
bio_pair_release(bp);
|
bio_pair_release(bp);
|
||||||
|
|
|
@ -773,9 +773,8 @@ do_sync_io:
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int make_request(struct request_queue *q, struct bio * bio)
|
static int make_request(mddev_t *mddev, struct bio * bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
conf_t *conf = mddev->private;
|
conf_t *conf = mddev->private;
|
||||||
mirror_info_t *mirror;
|
mirror_info_t *mirror;
|
||||||
r1bio_t *r1_bio;
|
r1bio_t *r1_bio;
|
||||||
|
|
|
@ -788,9 +788,8 @@ static void unfreeze_array(conf_t *conf)
|
||||||
spin_unlock_irq(&conf->resync_lock);
|
spin_unlock_irq(&conf->resync_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int make_request(struct request_queue *q, struct bio * bio)
|
static int make_request(mddev_t *mddev, struct bio * bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
conf_t *conf = mddev->private;
|
conf_t *conf = mddev->private;
|
||||||
mirror_info_t *mirror;
|
mirror_info_t *mirror;
|
||||||
r10bio_t *r10_bio;
|
r10bio_t *r10_bio;
|
||||||
|
@ -824,9 +823,9 @@ static int make_request(struct request_queue *q, struct bio * bio)
|
||||||
*/
|
*/
|
||||||
bp = bio_split(bio,
|
bp = bio_split(bio,
|
||||||
chunk_sects - (bio->bi_sector & (chunk_sects - 1)) );
|
chunk_sects - (bio->bi_sector & (chunk_sects - 1)) );
|
||||||
if (make_request(q, &bp->bio1))
|
if (make_request(mddev, &bp->bio1))
|
||||||
generic_make_request(&bp->bio1);
|
generic_make_request(&bp->bio1);
|
||||||
if (make_request(q, &bp->bio2))
|
if (make_request(mddev, &bp->bio2))
|
||||||
generic_make_request(&bp->bio2);
|
generic_make_request(&bp->bio2);
|
||||||
|
|
||||||
bio_pair_release(bp);
|
bio_pair_release(bp);
|
||||||
|
|
|
@ -3753,9 +3753,8 @@ static int bio_fits_rdev(struct bio *bi)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static int chunk_aligned_read(struct request_queue *q, struct bio * raid_bio)
|
static int chunk_aligned_read(mddev_t *mddev, struct bio * raid_bio)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
raid5_conf_t *conf = mddev->private;
|
raid5_conf_t *conf = mddev->private;
|
||||||
int dd_idx;
|
int dd_idx;
|
||||||
struct bio* align_bi;
|
struct bio* align_bi;
|
||||||
|
@ -3870,9 +3869,8 @@ static struct stripe_head *__get_priority_stripe(raid5_conf_t *conf)
|
||||||
return sh;
|
return sh;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int make_request(struct request_queue *q, struct bio * bi)
|
static int make_request(mddev_t *mddev, struct bio * bi)
|
||||||
{
|
{
|
||||||
mddev_t *mddev = q->queuedata;
|
|
||||||
raid5_conf_t *conf = mddev->private;
|
raid5_conf_t *conf = mddev->private;
|
||||||
int dd_idx;
|
int dd_idx;
|
||||||
sector_t new_sector;
|
sector_t new_sector;
|
||||||
|
@ -3896,7 +3894,7 @@ static int make_request(struct request_queue *q, struct bio * bi)
|
||||||
|
|
||||||
if (rw == READ &&
|
if (rw == READ &&
|
||||||
mddev->reshape_position == MaxSector &&
|
mddev->reshape_position == MaxSector &&
|
||||||
chunk_aligned_read(q,bi))
|
chunk_aligned_read(mddev,bi))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
logical_sector = bi->bi_sector & ~((sector_t)STRIPE_SECTORS-1);
|
logical_sector = bi->bi_sector & ~((sector_t)STRIPE_SECTORS-1);
|
||||||
|
|
Loading…
Add table
Reference in a new issue