Yahoo Gmail Google Facebook Delicious Twitter Reddit Stumpleupon Myspace Digg

Search queries

www.xxxcon.mvn, www.xxxcon, error expect exactly 2 parameter, procmail msmtp, yxxxcom, mesg: ttyname failed: Unpassender IOCTL (I/O-Control) für das Gerät,, ora-65280, wwww.xxxmo, wwwxxx.cko



#1: Potential data rollback/corruption after drive failure and re-appearance

Posted on 2011-10-17 08:46:16 by Moshe Melnikov


I am testing the following scenario: a simple RAID1 md array with drives A
and B. Assume that drive B fails, but the array remains operational and
services IOs. After a while, machine is rebooted. After reboot drive B comes
back, but now drive A becomes inaccessible. Assembling the array with both
drives results in a degraded array, with a single drive B. However, B's data
is the array's data at the time of drive B failure, not the latest array's
data. So the data kind of rolls back in time.

Testing a similar scenario with RAID5: A,B and C drives, C drive fails,
RAID5 becomes degraded but operational. After reboot B and C are accessible,
but A disappears. Assembling the array fails, unless --force is given.
With --force, the array comes up, but the data, of course, is corrupted.

Is this behavior intentional?

Suppose I want to protect against this by first examining the MD superblocks
(--examine). I want to find the most updated drive, and check what array
state it shows. Which part of "mdadm --examine" output should I use to find
the most updated drive? The "Update Time" or the "Events" counter? Or
perhaps something else?

Moshe Melnikov

To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to
More majordomo info at

Report this message