|
@@ -4231,8 +4231,6 @@ action_show(struct mddev *mddev, char *page)
|
|
|
return sprintf(page, "%s\n", type);
|
|
|
}
|
|
|
|
|
|
-static void reap_sync_thread(struct mddev *mddev);
|
|
|
-
|
|
|
static ssize_t
|
|
|
action_store(struct mddev *mddev, const char *page, size_t len)
|
|
|
{
|
|
@@ -4247,7 +4245,7 @@ action_store(struct mddev *mddev, const char *page, size_t len)
|
|
|
if (cmd_match(page, "idle") || cmd_match(page, "frozen")) {
|
|
|
if (mddev->sync_thread) {
|
|
|
set_bit(MD_RECOVERY_INTR, &mddev->recovery);
|
|
|
- reap_sync_thread(mddev);
|
|
|
+ md_reap_sync_thread(mddev);
|
|
|
}
|
|
|
} else if (test_bit(MD_RECOVERY_RUNNING, &mddev->recovery) ||
|
|
|
test_bit(MD_RECOVERY_NEEDED, &mddev->recovery))
|
|
@@ -5285,7 +5283,7 @@ static void __md_stop_writes(struct mddev *mddev)
|
|
|
if (mddev->sync_thread) {
|
|
|
set_bit(MD_RECOVERY_FROZEN, &mddev->recovery);
|
|
|
set_bit(MD_RECOVERY_INTR, &mddev->recovery);
|
|
|
- reap_sync_thread(mddev);
|
|
|
+ md_reap_sync_thread(mddev);
|
|
|
}
|
|
|
|
|
|
del_timer_sync(&mddev->safemode_timer);
|
|
@@ -7742,51 +7740,6 @@ no_add:
|
|
|
return spares;
|
|
|
}
|
|
|
|
|
|
-static void reap_sync_thread(struct mddev *mddev)
|
|
|
-{
|
|
|
- struct md_rdev *rdev;
|
|
|
-
|
|
|
- /* resync has finished, collect result */
|
|
|
- md_unregister_thread(&mddev->sync_thread);
|
|
|
- if (!test_bit(MD_RECOVERY_INTR, &mddev->recovery) &&
|
|
|
- !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) {
|
|
|
- /* success...*/
|
|
|
- /* activate any spares */
|
|
|
- if (mddev->pers->spare_active(mddev)) {
|
|
|
- sysfs_notify(&mddev->kobj, NULL,
|
|
|
- "degraded");
|
|
|
- set_bit(MD_CHANGE_DEVS, &mddev->flags);
|
|
|
- }
|
|
|
- }
|
|
|
- if (test_bit(MD_RECOVERY_RESHAPE, &mddev->recovery) &&
|
|
|
- mddev->pers->finish_reshape)
|
|
|
- mddev->pers->finish_reshape(mddev);
|
|
|
-
|
|
|
- /* If array is no-longer degraded, then any saved_raid_disk
|
|
|
- * information must be scrapped. Also if any device is now
|
|
|
- * In_sync we must scrape the saved_raid_disk for that device
|
|
|
- * do the superblock for an incrementally recovered device
|
|
|
- * written out.
|
|
|
- */
|
|
|
- rdev_for_each(rdev, mddev)
|
|
|
- if (!mddev->degraded ||
|
|
|
- test_bit(In_sync, &rdev->flags))
|
|
|
- rdev->saved_raid_disk = -1;
|
|
|
-
|
|
|
- md_update_sb(mddev, 1);
|
|
|
- clear_bit(MD_RECOVERY_RUNNING, &mddev->recovery);
|
|
|
- clear_bit(MD_RECOVERY_SYNC, &mddev->recovery);
|
|
|
- clear_bit(MD_RECOVERY_RESHAPE, &mddev->recovery);
|
|
|
- clear_bit(MD_RECOVERY_REQUESTED, &mddev->recovery);
|
|
|
- clear_bit(MD_RECOVERY_CHECK, &mddev->recovery);
|
|
|
- /* flag recovery needed just to double check */
|
|
|
- set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
|
|
|
- sysfs_notify_dirent_safe(mddev->sysfs_action);
|
|
|
- md_new_event(mddev);
|
|
|
- if (mddev->event_work.func)
|
|
|
- queue_work(md_misc_wq, &mddev->event_work);
|
|
|
-}
|
|
|
-
|
|
|
/*
|
|
|
* This routine is regularly called by all per-raid-array threads to
|
|
|
* deal with generic issues like resync and super-block update.
|
|
@@ -7883,7 +7836,7 @@ void md_check_recovery(struct mddev *mddev)
|
|
|
goto unlock;
|
|
|
}
|
|
|
if (mddev->sync_thread) {
|
|
|
- reap_sync_thread(mddev);
|
|
|
+ md_reap_sync_thread(mddev);
|
|
|
goto unlock;
|
|
|
}
|
|
|
/* Set RUNNING before clearing NEEDED to avoid
|
|
@@ -7964,6 +7917,51 @@ void md_check_recovery(struct mddev *mddev)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
+void md_reap_sync_thread(struct mddev *mddev)
|
|
|
+{
|
|
|
+ struct md_rdev *rdev;
|
|
|
+
|
|
|
+ /* resync has finished, collect result */
|
|
|
+ md_unregister_thread(&mddev->sync_thread);
|
|
|
+ if (!test_bit(MD_RECOVERY_INTR, &mddev->recovery) &&
|
|
|
+ !test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) {
|
|
|
+ /* success...*/
|
|
|
+ /* activate any spares */
|
|
|
+ if (mddev->pers->spare_active(mddev)) {
|
|
|
+ sysfs_notify(&mddev->kobj, NULL,
|
|
|
+ "degraded");
|
|
|
+ set_bit(MD_CHANGE_DEVS, &mddev->flags);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ if (test_bit(MD_RECOVERY_RESHAPE, &mddev->recovery) &&
|
|
|
+ mddev->pers->finish_reshape)
|
|
|
+ mddev->pers->finish_reshape(mddev);
|
|
|
+
|
|
|
+ /* If array is no-longer degraded, then any saved_raid_disk
|
|
|
+ * information must be scrapped. Also if any device is now
|
|
|
+ * In_sync we must scrape the saved_raid_disk for that device
|
|
|
+ * do the superblock for an incrementally recovered device
|
|
|
+ * written out.
|
|
|
+ */
|
|
|
+ rdev_for_each(rdev, mddev)
|
|
|
+ if (!mddev->degraded ||
|
|
|
+ test_bit(In_sync, &rdev->flags))
|
|
|
+ rdev->saved_raid_disk = -1;
|
|
|
+
|
|
|
+ md_update_sb(mddev, 1);
|
|
|
+ clear_bit(MD_RECOVERY_RUNNING, &mddev->recovery);
|
|
|
+ clear_bit(MD_RECOVERY_SYNC, &mddev->recovery);
|
|
|
+ clear_bit(MD_RECOVERY_RESHAPE, &mddev->recovery);
|
|
|
+ clear_bit(MD_RECOVERY_REQUESTED, &mddev->recovery);
|
|
|
+ clear_bit(MD_RECOVERY_CHECK, &mddev->recovery);
|
|
|
+ /* flag recovery needed just to double check */
|
|
|
+ set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
|
|
|
+ sysfs_notify_dirent_safe(mddev->sysfs_action);
|
|
|
+ md_new_event(mddev);
|
|
|
+ if (mddev->event_work.func)
|
|
|
+ queue_work(md_misc_wq, &mddev->event_work);
|
|
|
+}
|
|
|
+
|
|
|
void md_wait_for_blocked_rdev(struct md_rdev *rdev, struct mddev *mddev)
|
|
|
{
|
|
|
sysfs_notify_dirent_safe(rdev->sysfs_state);
|
|
@@ -8689,6 +8687,7 @@ EXPORT_SYMBOL(md_register_thread);
|
|
|
EXPORT_SYMBOL(md_unregister_thread);
|
|
|
EXPORT_SYMBOL(md_wakeup_thread);
|
|
|
EXPORT_SYMBOL(md_check_recovery);
|
|
|
+EXPORT_SYMBOL(md_reap_sync_thread);
|
|
|
MODULE_LICENSE("GPL");
|
|
|
MODULE_DESCRIPTION("MD RAID framework");
|
|
|
MODULE_ALIAS("md");
|