|
@@ -2049,9 +2049,12 @@ handle_requests_to_failed_array(raid5_conf_t *conf, struct stripe_head *sh,
|
|
bi = bi2;
|
|
bi = bi2;
|
|
}
|
|
}
|
|
|
|
|
|
- /* fail any reads if this device is non-operational */
|
|
|
|
- if (!test_bit(R5_Insync, &sh->dev[i].flags) ||
|
|
|
|
- test_bit(R5_ReadError, &sh->dev[i].flags)) {
|
|
|
|
|
|
+ /* fail any reads if this device is non-operational and
|
|
|
|
+ * the data has not reached the cache yet.
|
|
|
|
+ */
|
|
|
|
+ if (!test_bit(R5_Wantfill, &sh->dev[i].flags) &&
|
|
|
|
+ (!test_bit(R5_Insync, &sh->dev[i].flags) ||
|
|
|
|
+ test_bit(R5_ReadError, &sh->dev[i].flags))) {
|
|
bi = sh->dev[i].toread;
|
|
bi = sh->dev[i].toread;
|
|
sh->dev[i].toread = NULL;
|
|
sh->dev[i].toread = NULL;
|
|
if (test_and_clear_bit(R5_Overlap, &sh->dev[i].flags))
|
|
if (test_and_clear_bit(R5_Overlap, &sh->dev[i].flags))
|
|
@@ -2740,37 +2743,27 @@ static void handle_stripe5(struct stripe_head *sh)
|
|
struct r5dev *dev = &sh->dev[i];
|
|
struct r5dev *dev = &sh->dev[i];
|
|
clear_bit(R5_Insync, &dev->flags);
|
|
clear_bit(R5_Insync, &dev->flags);
|
|
|
|
|
|
- pr_debug("check %d: state 0x%lx read %p write %p written %p\n",
|
|
|
|
- i, dev->flags, dev->toread, dev->towrite, dev->written);
|
|
|
|
- /* maybe we can reply to a read */
|
|
|
|
- if (test_bit(R5_UPTODATE, &dev->flags) && dev->toread) {
|
|
|
|
- struct bio *rbi, *rbi2;
|
|
|
|
- pr_debug("Return read for disc %d\n", i);
|
|
|
|
- spin_lock_irq(&conf->device_lock);
|
|
|
|
- rbi = dev->toread;
|
|
|
|
- dev->toread = NULL;
|
|
|
|
- if (test_and_clear_bit(R5_Overlap, &dev->flags))
|
|
|
|
- wake_up(&conf->wait_for_overlap);
|
|
|
|
- spin_unlock_irq(&conf->device_lock);
|
|
|
|
- while (rbi && rbi->bi_sector < dev->sector + STRIPE_SECTORS) {
|
|
|
|
- copy_data(0, rbi, dev->page, dev->sector);
|
|
|
|
- rbi2 = r5_next_bio(rbi, dev->sector);
|
|
|
|
- spin_lock_irq(&conf->device_lock);
|
|
|
|
- if (--rbi->bi_phys_segments == 0) {
|
|
|
|
- rbi->bi_next = return_bi;
|
|
|
|
- return_bi = rbi;
|
|
|
|
- }
|
|
|
|
- spin_unlock_irq(&conf->device_lock);
|
|
|
|
- rbi = rbi2;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
|
|
+ pr_debug("check %d: state 0x%lx toread %p read %p write %p "
|
|
|
|
+ "written %p\n", i, dev->flags, dev->toread, dev->read,
|
|
|
|
+ dev->towrite, dev->written);
|
|
|
|
+
|
|
|
|
+ /* maybe we can request a biofill operation
|
|
|
|
+ *
|
|
|
|
+ * new wantfill requests are only permitted while
|
|
|
|
+ * STRIPE_OP_BIOFILL is clear
|
|
|
|
+ */
|
|
|
|
+ if (test_bit(R5_UPTODATE, &dev->flags) && dev->toread &&
|
|
|
|
+ !test_bit(STRIPE_OP_BIOFILL, &sh->ops.pending))
|
|
|
|
+ set_bit(R5_Wantfill, &dev->flags);
|
|
|
|
|
|
/* now count some things */
|
|
/* now count some things */
|
|
if (test_bit(R5_LOCKED, &dev->flags)) s.locked++;
|
|
if (test_bit(R5_LOCKED, &dev->flags)) s.locked++;
|
|
if (test_bit(R5_UPTODATE, &dev->flags)) s.uptodate++;
|
|
if (test_bit(R5_UPTODATE, &dev->flags)) s.uptodate++;
|
|
if (test_bit(R5_Wantcompute, &dev->flags)) s.compute++;
|
|
if (test_bit(R5_Wantcompute, &dev->flags)) s.compute++;
|
|
|
|
|
|
- if (dev->toread)
|
|
|
|
|
|
+ if (test_bit(R5_Wantfill, &dev->flags))
|
|
|
|
+ s.to_fill++;
|
|
|
|
+ else if (dev->toread)
|
|
s.to_read++;
|
|
s.to_read++;
|
|
if (dev->towrite) {
|
|
if (dev->towrite) {
|
|
s.to_write++;
|
|
s.to_write++;
|
|
@@ -2793,6 +2786,10 @@ static void handle_stripe5(struct stripe_head *sh)
|
|
set_bit(R5_Insync, &dev->flags);
|
|
set_bit(R5_Insync, &dev->flags);
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
|
|
+
|
|
|
|
+ if (s.to_fill && !test_and_set_bit(STRIPE_OP_BIOFILL, &sh->ops.pending))
|
|
|
|
+ sh->ops.count++;
|
|
|
|
+
|
|
pr_debug("locked=%d uptodate=%d to_read=%d"
|
|
pr_debug("locked=%d uptodate=%d to_read=%d"
|
|
" to_write=%d failed=%d failed_num=%d\n",
|
|
" to_write=%d failed=%d failed_num=%d\n",
|
|
s.locked, s.uptodate, s.to_read, s.to_write,
|
|
s.locked, s.uptodate, s.to_read, s.to_write,
|