Peter T. Breuer wrote: []
The patch was originally developed for 2.4, then ported to 2.6.3, and then to 2.6.8.1. Neil has recently been doing stuff, so I don't think it applies cleanly to 2.6.10, but somebody WAS porting it for me until they found that 2.6.10 didn't support their hardware ... and I recall discussing with him what to do about the change of map() to read_balance() in the code (essentially, put map() back). And finding that the spinlocks have changed too.
Well, it turns out current code is easier to modify, including spinlocks.
But now, with read_balance() in place, which can pick a "random" disk to read from, we have to keep some sort of bitmap to mark disks which we tried to read from.
For the hack below I've added r1_bio->tried_disk of type unsigned long which has one bit per disk, so current scheme is limited to 32 disks in array. This is really a hack for now -- I don't know much about kernel programming rules... ;)
@@ -266,9 +368,19 @@
/*
* this branch is our 'one mirror IO has finished' event handler:
*/
- if (!uptodate)
- md_error(r1_bio->mddev, conf->mirrors[mirror].rdev);
- else
+ if (!uptodate) {
+#ifdef CONFIG_MD_RAID1_ROBUST_READ
+ /*
+ * Only fault disk out of array on write error, not read.
+ */
+ if (0)
+#endif /* CONFIG_MD_RAID1_ROBUST_READ */
+ md_error(r1_bio->mddev, conf->mirrors[mirror].rdev);
+#ifdef DO_ADD_READ_WRITE_CORRECT
What's this? Was it an attempt (incomplete) to do rewrite-after-failed-read?
+ else /* tell next time we're here that we're a retry */
+ set_bit(R1BIO_ReadRetry, &r1_bio->state);
+#endif /* DO_ADD_READ_WRITE_CORRECT */
+ } else
/*
* Set R1BIO_Uptodate in our master bio, so that
* we will return a good error code for to the higher
@@ -285,8 +397,20 @@
/*
* we have only one bio on the read side
*/
- if (uptodate)
- raid_end_bio_io(r1_bio);
+ if (uptodate
+#ifdef CONFIG_MD_RAID1_ROBUST_READ
+ /* Give up and error if we're last */
+ || (atomic_dec_and_test(&r1_bio->remaining))
+#endif /* CONFIG_MD_RAID1_ROBUST_READ */
+ )
+#ifdef DO_ADD_READ_WRITE_CORRECT
+ if (uptodate && test_bit(R1BIO_ReadRetry, &r1_bio->state)) {
+ /* Success at last - rewrite failed reads */
+ set_bit(R1BIO_IsSync, &r1_bio->state);
+ reschedule_retry(r1_bio);
+ } else
+#endif /* DO_ADD_READ_WRITE_CORRECT */
+ raid_end_bio_io(r1_bio);
Hmm. Should we do actual write here, instead of rescheduling a successeful read further, after finishing the original request?
/mjt
--- ./include/linux/raid/raid1.h.orig Wed Mar 2 10:38:10 2005
+++ ./include/linux/raid/raid1.h Sat Mar 19 18:53:42 2005
@@ -83,6 +83,7 @@ struct r1bio_s {
* if the IO is in READ direction, then this is where we read
*/
int read_disk;
+ unsigned long tried_disks; /* bitmap, disks we had tried
to read from */
struct list_head retry_list;
/*
--- ./drivers/md/raid1.c.orig Wed Mar 2 10:38:10 2005
+++ ./drivers/md/raid1.c Sat Mar 19 18:57:16 2005
@@ -234,9 +234,13 @@ static int raid1_end_read_request(struct
/*
* this branch is our 'one mirror IO has finished' event handler:
*/
- if (!uptodate)
- md_error(r1_bio->mddev, conf->mirrors[mirror].rdev);
- else
+
+ update_head_pos(mirror, r1_bio);
+
+ /*
+ * we have only one bio on the read side
+ */
+ if (uptodate) {
/*
* Set R1BIO_Uptodate in our master bio, so that
* we will return a good error code for to the higher
@@ -247,14 +251,8 @@ static int raid1_end_read_request(struct
* wait for the 'master' bio.
*/
set_bit(R1BIO_Uptodate, &r1_bio->state);
-
- update_head_pos(mirror, r1_bio);
-
- /*
- * we have only one bio on the read side
- */
- if (uptodate)
raid_end_bio_io(r1_bio);
+ }
else {
/*
* oops, read error:
@@ -332,6 +330,10 @@ static int raid1_end_write_request(struc
*
* The rdev for the device selected will have nr_pending incremented.
*/
+
+#define disk_tried_before(r1_bio, disk) ((r1_bio)->tried_disks &
(1<<(disk)))
+#define mark_disk_tried(r1_bio, disk) ((r1_bio)->tried_disks |= 1<<(disk))
+
static int read_balance(conf_t *conf, r1bio_t *r1_bio)
{
const unsigned long this_sector = r1_bio->sector;
@@ -351,7 +353,8 @@ static int read_balance(conf_t *conf, r1
new_disk = 0;
while (!conf->mirrors[new_disk].rdev ||
- !conf->mirrors[new_disk].rdev->in_sync) {
+ !conf->mirrors[new_disk].rdev->in_sync ||
+ disk_tried_before(r1_bio, new_disk)) {
new_disk++;
if (new_disk == conf->raid_disks) {
new_disk = -1;
@@ -364,7 +367,8 @@ static int read_balance(conf_t *conf, r1
/* make sure the disk is operational */
while (!conf->mirrors[new_disk].rdev ||
- !conf->mirrors[new_disk].rdev->in_sync) {
+ !conf->mirrors[new_disk].rdev->in_sync ||
+ disk_tried_before(r1_bio, new_disk)) {
if (new_disk <= 0)
new_disk = conf->raid_disks;
new_disk--;
@@ -394,7 +398,8 @@ static int read_balance(conf_t *conf, r1
disk--;
if (!conf->mirrors[disk].rdev ||
- !conf->mirrors[disk].rdev->in_sync)
+ !conf->mirrors[disk].rdev->in_sync ||
+ disk_tried_before(r1_bio, new_disk))
continue;
if (!atomic_read(&conf->mirrors[disk].rdev->nr_pending)) {
@@ -415,6 +420,7 @@ rb_out:
conf->next_seq_sect = this_sector + sectors;
conf->last_used = new_disk;
atomic_inc(&conf->mirrors[new_disk].rdev->nr_pending);
+ mark_disk_tried(r1_bio, new_disk);
}
rcu_read_unlock();
@@ -545,6 +551,7 @@ static int make_request(request_queue_t
r1_bio->sector = bio->bi_sector;
r1_bio->state = 0;
+ r1_bio->tried_disks = 0;
if (bio_data_dir(bio) == READ) {
/*
