if (!test_bit(MD_RECOVERY_FROZEN, &rs->md.recovery))
md_stop_writes(&rs->md);
- __mddev_suspend(&rs->md, false);
+ mddev_suspend(&rs->md, false);
}
}
}
/* Check for any resize/reshape on @rs and adjust/initiate */
- /* Be prepared for __mddev_resume() in raid_resume() */
+ /* Be prepared for mddev_resume() in raid_resume() */
set_bit(MD_RECOVERY_FROZEN, &mddev->recovery);
if (mddev->recovery_cp && mddev->recovery_cp < MaxSector) {
set_bit(MD_RECOVERY_REQUESTED, &mddev->recovery);
* Make sure no new requests are submitted to the device, and any requests that
* have been submitted are completely handled.
*/
-int __mddev_suspend(struct mddev *mddev, bool interruptible)
+int mddev_suspend(struct mddev *mddev, bool interruptible)
{
int err = 0;
mutex_unlock(&mddev->suspend_mutex);
return 0;
}
-EXPORT_SYMBOL_GPL(__mddev_suspend);
+EXPORT_SYMBOL_GPL(mddev_suspend);
-void __mddev_resume(struct mddev *mddev)
+void mddev_resume(struct mddev *mddev)
{
lockdep_assert_not_held(&mddev->reconfig_mutex);
return;
}
- /* entred the memalloc scope from __mddev_suspend() */
+ /* entred the memalloc scope from mddev_suspend() */
memalloc_noio_restore(mddev->noio_flag);
percpu_ref_resurrect(&mddev->active_io);
mutex_unlock(&mddev->suspend_mutex);
}
-EXPORT_SYMBOL_GPL(__mddev_resume);
+EXPORT_SYMBOL_GPL(mddev_resume);
/*
* Generic flush handling for md
if (new != (sector_t)new)
return -EINVAL;
- err = __mddev_suspend(mddev, true);
+ err = mddev_suspend(mddev, true);
if (err)
return err;
WRITE_ONCE(mddev->suspend_lo, new);
- __mddev_resume(mddev);
+ mddev_resume(mddev);
return len;
}
if (new != (sector_t)new)
return -EINVAL;
- err = __mddev_suspend(mddev, true);
+ err = mddev_suspend(mddev, true);
if (err)
return err;
WRITE_ONCE(mddev->suspend_hi, new);
- __mddev_resume(mddev);
+ mddev_resume(mddev);
return len;
}
extern void md_rdev_clear(struct md_rdev *rdev);
extern void md_handle_request(struct mddev *mddev, struct bio *bio);
-extern int __mddev_suspend(struct mddev *mddev, bool interruptible);
-extern void __mddev_resume(struct mddev *mddev);
+extern int mddev_suspend(struct mddev *mddev, bool interruptible);
+extern void mddev_resume(struct mddev *mddev);
extern void md_reload_sb(struct mddev *mddev, int raid_disk);
extern void md_update_sb(struct mddev *mddev, int force);
{
int ret;
- ret = __mddev_suspend(mddev, true);
+ ret = mddev_suspend(mddev, true);
if (ret)
return ret;
ret = mddev_lock(mddev);
if (ret)
- __mddev_resume(mddev);
+ mddev_resume(mddev);
return ret;
}
static inline void mddev_suspend_and_lock_nointr(struct mddev *mddev)
{
- __mddev_suspend(mddev, false);
+ mddev_suspend(mddev, false);
mutex_lock(&mddev->reconfig_mutex);
}
static inline void mddev_unlock_and_resume(struct mddev *mddev)
{
mddev_unlock(mddev);
- __mddev_resume(mddev);
+ mddev_resume(mddev);
}
struct mdu_array_info_s;
log = READ_ONCE(conf->log);
if (log) {
- __mddev_suspend(mddev, false);
+ mddev_suspend(mddev, false);
log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH;
- __mddev_resume(mddev);
+ mddev_resume(mddev);
}
}