What the heck happened to my array? (No apparent data loss).

What the heck happened to my array? (No apparent data loss).

am 03.04.2011 15:32:20 von Brad Campbell

2.6.38.2
x86_64
10 x 1TB SATA drives in a single RAID-6

Here is the chain of events.

Saturday morning I started a reshape on a 10 element RAID-6. Simply
changing the Chunk size from 512k to 64k. This was going to take about
4.5 days according to the initial estimates.

I then went away for the weekend and came home to a wedged array.
Here is the chain of events that caused it.

This occurred about 1 minute after my scheduled morning SMART long (it
is Sunday after all) began.

Apr 3 03:19:08 srv kernel: [288180.455339] sd 0:0:12:0: [sdd] Unhandled
error code
Apr 3 03:19:08 srv kernel: [288180.455359] sd 0:0:12:0: [sdd] Result:
hostbyte=0x04 driverbyte=0x00
Apr 3 03:19:08 srv kernel: [288180.455377] sd 0:0:12:0: [sdd] CDB:
cdb[0]=0x2a: 2a 00 00 00 00 08 00 00 02 00
Apr 3 03:19:08 srv kernel: [288180.455415] end_request: I/O error, dev
sdd, sector 8
Apr 3 03:19:08 srv kernel: [288180.455449] end_request: I/O error, dev
sdd, sector 8
Apr 3 03:19:08 srv kernel: [288180.455462] md: super_written gets
error=-5, uptodate=0
Apr 3 03:19:08 srv kernel: [288180.455477] md/raid:md0: Disk failure on
sdd, disabling device.
Apr 3 03:19:08 srv kernel: [288180.455480] md/raid:md0: Operation
continuing on 9 devices.
Apr 3 03:19:08 srv kernel: [288180.472914] md: md0: reshape done.
Apr 3 03:19:08 srv kernel: [288180.472983] md: delaying data-check of
md5 until md3 has finished (they share one or more physical units)
Apr 3 03:19:08 srv kernel: [288180.473002] md: delaying data-check of
md4 until md6 has finished (they share one or more physical units)
Apr 3 03:19:08 srv kernel: [288180.473030] md: delaying data-check of
md6 until md5 has finished (they share one or more physical units)
Apr 3 03:19:08 srv kernel: [288180.473047] md: delaying data-check of
md3 until md1 has finished (they share one or more physical units)
Apr 3 03:19:08 srv kernel: [288180.551450] md: reshape of RAID array md0
Apr 3 03:19:08 srv kernel: [288180.551468] md: minimum _guaranteed_
speed: 200000 KB/sec/disk.
Apr 3 03:19:08 srv kernel: [288180.551483] md: using maximum available
idle IO bandwidth (but not more than 200000 KB/sec) for reshape.
Apr 3 03:19:08 srv kernel: [288180.551514] md: using 128k window, over
a total of 976759808 blocks.
Apr 3 03:19:08 srv kernel: [288180.620089] sd 0:0:12:0: [sdd]
Synchronizing SCSI cache
Apr 3 03:19:08 srv mdadm[4803]: RebuildFinished event detected on md
device /dev/md0
Apr 3 03:19:08 srv mdadm[4803]: Fail event detected on md device
/dev/md0, component device /dev/sdd
Apr 3 03:19:08 srv mdadm[4803]: RebuildStarted event detected on md
device /dev/md0
Apr 3 03:19:10 srv kernel: [288182.614918] scsi 0:0:12:0: Direct-Access
ATA MAXTOR STM310003 MX1A PQ: 0 ANSI: 5
Apr 3 03:19:10 srv kernel: [288182.615312] sd 0:0:12:0: Attached scsi
generic sg3 type 0
Apr 3 03:19:10 srv kernel: [288182.618262] sd 0:0:12:0: [sdq]
1953525168 512-byte logical blocks: (1.00 TB/931 GiB)
Apr 3 03:19:10 srv kernel: [288182.736998] sd 0:0:12:0: [sdq] Write
Protect is off
Apr 3 03:19:10 srv kernel: [288182.737019] sd 0:0:12:0: [sdq] Mode
Sense: 73 00 00 08
Apr 3 03:19:10 srv kernel: [288182.740521] sd 0:0:12:0: [sdq] Write
cache: enabled, read cache: enabled, doesn't support DPO or FUA
Apr 3 03:19:10 srv kernel: [288182.848999] sdq: unknown partition table
Apr 3 03:19:10 srv ata_id[28453]: HDIO_GET_IDENTITY failed for '/dev/sdq'
Apr 3 03:19:10 srv kernel: [288182.970091] sd 0:0:12:0: [sdq] Attached
SCSI disk
Apr 3 03:20:01 srv /USR/SBIN/CRON[28624]: (brad) CMD ([ -z
"`/usr/bin/pgrep -u brad collect`" ] && /usr/bin/screen -X -S brad-bot
screen /home/brad/bin/collect-thermostat)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28625]: (root) CMD ([ -z
`/usr/bin/pgrep -u root keepalive` ] && /home/brad/bin/launch-keepalive)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28626]: (brad) CMD ([ -z "`screen
-list | grep brad-bot`" ] && /home/brad/bin/botstart)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28628]: (root) CMD (if [ -x
/usr/bin/mrtg ] && [ -r /etc/mrtg.cfg ]; then mkdir -p /var/log/mrtg ;
env LANG=C /usr/bin/mrtg /etc/mrtg.cfg 2>&1 | tee -a
/var/log/mrtg/mrtg.log ; fi)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28627]: (brad) CMD
(/home/brad/rrd/rrd-create-graphs)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28590]: (CRON) error (grandchild
#28625 failed with exit status 1)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28589]: (CRON) error (grandchild
#28626 failed with exit status 1)
Apr 3 03:20:01 srv /USR/SBIN/CRON[28587]: (CRON) error (grandchild
#28624 failed with exit status 1)
Apr 3 03:22:10 srv kernel: [288363.070094] INFO: task jbd2/md0-8:2647
blocked for more than 120 seconds.
Apr 3 03:22:10 srv kernel: [288363.070114] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
Apr 3 03:22:10 srv kernel: [288363.070132] jbd2/md0-8 D
ffff88041aa52948 0 2647 2 0x00000000
Apr 3 03:22:10 srv kernel: [288363.070154] ffff88041aa526f0
0000000000000046 0000000000000000 ffff8804196769b0
Apr 3 03:22:10 srv kernel: [288363.070178] 0000000000011180
ffff88041bdc5fd8 0000000000004000 0000000000011180
Apr 3 03:22:10 srv kernel: [288363.070201] ffff88041bdc4010
ffff88041aa52950 ffff88041bdc5fd8 ffff88041aa52948
Apr 3 03:22:10 srv kernel: [288363.070224] Call Trace:
Apr 3 03:22:10 srv kernel: [288363.070246] [] ?
queue_work_on+0x16/0x20
Apr 3 03:22:10 srv kernel: [288363.070266] [] ?
md_write_start+0xad/0x190
Apr 3 03:22:10 srv kernel: [288363.070283] [] ?
autoremove_wake_function+0x0/0x30
Apr 3 03:22:10 srv kernel: [288363.070299] [] ?
make_request+0x35/0x600
Apr 3 03:22:10 srv kernel: [288363.070317] [] ?
__alloc_pages_nodemask+0x10b/0x810
Apr 3 03:22:10 srv kernel: [288363.070335] [] ?
T.1015+0x32/0x90
Apr 3 03:22:10 srv kernel: [288363.070350] [] ?
md_make_request+0xd4/0x200
Apr 3 03:22:10 srv kernel: [288363.070366] [] ?
ext4_map_blocks+0x178/0x210
Apr 3 03:22:10 srv kernel: [288363.070382] [] ?
generic_make_request+0x144/0x2f0
Apr 3 03:22:10 srv kernel: [288363.070397] [] ?
jbd2_journal_file_buffer+0x3d/0x70
Apr 3 03:22:10 srv kernel: [288363.070413] [] ?
submit_bio+0x5c/0xd0
Apr 3 03:22:10 srv kernel: [288363.070430] [] ?
submit_bh+0xe5/0x120
Apr 3 03:22:10 srv kernel: [288363.070445] [] ?
jbd2_journal_commit_transaction+0x441/0x1180
Apr 3 03:22:10 srv kernel: [288363.070466] [] ?
lock_timer_base+0x33/0x70
Apr 3 03:22:10 srv kernel: [288363.070480] [] ?
autoremove_wake_function+0x0/0x30
Apr 3 03:22:10 srv kernel: [288363.070498] [] ?
kjournald2+0xb1/0x1e0
Apr 3 03:22:10 srv kernel: [288363.070511] [] ?
autoremove_wake_function+0x0/0x30
Apr 3 03:22:10 srv kernel: [288363.070527] [] ?
kjournald2+0x0/0x1e0
Apr 3 03:22:10 srv kernel: [288363.070544] [] ?
kjournald2+0x0/0x1e0
Apr 3 03:22:10 srv kernel: [288363.070557] [] ?
kthread+0x96/0xa0
Apr 3 03:22:10 srv kernel: [288363.070573] [] ?
kernel_thread_helper+0x4/0x10
Apr 3 03:22:10 srv kernel: [288363.070588] [] ?
kthread+0x0/0xa0
Apr 3 03:22:10 srv kernel: [288363.070602] [] ?
kernel_thread_helper+0x0/0x10

So apparently sdd suffered an unknown failure (it happens) and the array
kicked it out (as it should). But 120 seconds later all tasks accessing
that array trigger their 120 second hangcheck warning and are all suck
in the D state.

At the time the array was 12.1% of the way through a reshape. I had to
reboot the machine to get it back up and it's now continuing the reshape
on 9 drives.

brad@srv:~$ cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md0 : active raid6 sdc[0] sdh[9] sda[8] sde[7] sdg[5] sdb[4] sdf[3]
sdm[2] sdl[1]
7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2
[10/9] [UUUUUU_UUU]
[===>.................] reshape = 16.5% (162091008/976759808)
finish=5778.6min speed=2349K/sec



To make matters more confusing the other arrays on the machine were in
the middle of their "Debians first Sunday of every month" "check" scrub.

I have the full syslog and can probably procure any other information
that might be useful. I don't think I've lost any data, the machine
continued reshaping and we're all moving along nicely. I just wanted to
report it and offer assistance in diagnosing it should that be requested.

Regards,
Brad
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array? (No apparent data loss).

am 03.04.2011 17:47:34 von Roberto Spadim

what kernel version? more informations about your linux box?

2011/4/3 Brad Campbell :
> 2.6.38.2
> x86_64
> 10 x 1TB SATA drives in a single RAID-6
>
> Here is the chain of events.
>
> Saturday morning I started a reshape on a 10 element RAID-6. Simply c=
hanging
> the Chunk size from 512k to 64k. This was going to take about 4.5 day=
s
> according to the initial estimates.
>
> I then went away for the weekend and came home to a wedged array.
> Here is the chain of events that caused it.
>
> This occurred about 1 minute after my scheduled morning SMART long (i=
t is
> Sunday after all) began.
>
> Apr =A03 03:19:08 srv kernel: [288180.455339] sd 0:0:12:0: [sdd] Unha=
ndled
> error code
> Apr =A03 03:19:08 srv kernel: [288180.455359] sd 0:0:12:0: [sdd] =A0R=
esult:
> hostbyte=3D0x04 driverbyte=3D0x00
> Apr =A03 03:19:08 srv kernel: [288180.455377] sd 0:0:12:0: [sdd] CDB:
> cdb[0]=3D0x2a: 2a 00 00 00 00 08 00 00 02 00
> Apr =A03 03:19:08 srv kernel: [288180.455415] end_request: I/O error,=
dev sdd,
> sector 8
> Apr =A03 03:19:08 srv kernel: [288180.455449] end_request: I/O error,=
dev sdd,
> sector 8
> Apr =A03 03:19:08 srv kernel: [288180.455462] md: super_written gets =
error=3D-5,
> uptodate=3D0
> Apr =A03 03:19:08 srv kernel: [288180.455477] md/raid:md0: Disk failu=
re on
> sdd, disabling device.
> Apr =A03 03:19:08 srv kernel: [288180.455480] md/raid:md0: Operation
> continuing on 9 devices.
> Apr =A03 03:19:08 srv kernel: [288180.472914] md: md0: reshape done.
> Apr =A03 03:19:08 srv kernel: [288180.472983] md: delaying data-check=
of md5
> until md3 has finished (they share one or more physical units)
> Apr =A03 03:19:08 srv kernel: [288180.473002] md: delaying data-check=
of md4
> until md6 has finished (they share one or more physical units)
> Apr =A03 03:19:08 srv kernel: [288180.473030] md: delaying data-check=
of md6
> until md5 has finished (they share one or more physical units)
> Apr =A03 03:19:08 srv kernel: [288180.473047] md: delaying data-check=
of md3
> until md1 has finished (they share one or more physical units)
> Apr =A03 03:19:08 srv kernel: [288180.551450] md: reshape of RAID arr=
ay md0
> Apr =A03 03:19:08 srv kernel: [288180.551468] md: minimum _guaranteed=
_ speed:
> 200000 KB/sec/disk.
> Apr =A03 03:19:08 srv kernel: [288180.551483] md: using maximum avail=
able idle
> IO bandwidth (but not more than 200000 KB/sec) for reshape.
> Apr =A03 03:19:08 srv kernel: [288180.551514] md: using 128k window, =
over a
> total of 976759808 blocks.
> Apr =A03 03:19:08 srv kernel: [288180.620089] sd 0:0:12:0: [sdd] Sync=
hronizing
> SCSI cache
> Apr =A03 03:19:08 srv mdadm[4803]: RebuildFinished event detected on =
md device
> /dev/md0
> Apr =A03 03:19:08 srv mdadm[4803]: Fail event detected on md device /=
dev/md0,
> component device /dev/sdd
> Apr =A03 03:19:08 srv mdadm[4803]: RebuildStarted event detected on m=
d device
> /dev/md0
> Apr =A03 03:19:10 srv kernel: [288182.614918] scsi 0:0:12:0: Direct-A=
ccess
> ATA =A0 =A0 =A0MAXTOR STM310003 MX1A PQ: 0 ANSI: 5
> Apr =A03 03:19:10 srv kernel: [288182.615312] sd 0:0:12:0: Attached s=
csi
> generic sg3 type 0
> Apr =A03 03:19:10 srv kernel: [288182.618262] sd 0:0:12:0: [sdq] 1953=
525168
> 512-byte logical blocks: (1.00 TB/931 GiB)
> Apr =A03 03:19:10 srv kernel: [288182.736998] sd 0:0:12:0: [sdq] Writ=
e Protect
> is off
> Apr =A03 03:19:10 srv kernel: [288182.737019] sd 0:0:12:0: [sdq] Mode=
Sense:
> 73 00 00 08
> Apr =A03 03:19:10 srv kernel: [288182.740521] sd 0:0:12:0: [sdq] Writ=
e cache:
> enabled, read cache: enabled, doesn't support DPO or FUA
> Apr =A03 03:19:10 srv kernel: [288182.848999] =A0sdq: unknown partiti=
on table
> Apr =A03 03:19:10 srv ata_id[28453]: HDIO_GET_IDENTITY failed for '/d=
ev/sdq'
> Apr =A03 03:19:10 srv kernel: [288182.970091] sd 0:0:12:0: [sdq] Atta=
ched SCSI
> disk
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28624]: (brad) CMD ([ -z "`/usr/=
bin/pgrep
> -u brad collect`" ] && /usr/bin/screen -X -S brad-bot screen
> /home/brad/bin/collect-thermostat)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28625]: (root) CMD ([ -z `/usr/b=
in/pgrep
> -u root keepalive` ] && /home/brad/bin/launch-keepalive)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28626]: (brad) CMD ([ -z "`scree=
n -list |
> grep brad-bot`" ] && /home/brad/bin/botstart)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28628]: (root) CMD (if [ -x /usr=
/bin/mrtg
> ] && [ -r /etc/mrtg.cfg ]; then mkdir -p /var/log/mrtg ; env LANG=3DC
> /usr/bin/mrtg /etc/mrtg.cfg 2>&1 | tee -a /var/log/mrtg/mrtg.log ; fi=
)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28627]: (brad) CMD
> (/home/brad/rrd/rrd-create-graphs)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28590]: (CRON) error (grandchild=
#28625
> failed with exit status 1)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28589]: (CRON) error (grandchild=
#28626
> failed with exit status 1)
> Apr =A03 03:20:01 srv /USR/SBIN/CRON[28587]: (CRON) error (grandchild=
#28624
> failed with exit status 1)
> Apr =A03 03:22:10 srv kernel: [288363.070094] INFO: task jbd2/md0-8:2=
647
> blocked for more than 120 seconds.
> Apr =A03 03:22:10 srv kernel: [288363.070114] "echo 0 >
> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> Apr =A03 03:22:10 srv kernel: [288363.070132] jbd2/md0-8 =A0 =A0 =A0D
> ffff88041aa52948 =A0 =A0 0 =A02647 =A0 =A0 =A02 0x00000000
> Apr =A03 03:22:10 srv kernel: [288363.070154] =A0ffff88041aa526f0
> 0000000000000046 0000000000000000 ffff8804196769b0
> Apr =A03 03:22:10 srv kernel: [288363.070178] =A00000000000011180
> ffff88041bdc5fd8 0000000000004000 0000000000011180
> Apr =A03 03:22:10 srv kernel: [288363.070201] =A0ffff88041bdc4010
> ffff88041aa52950 ffff88041bdc5fd8 ffff88041aa52948
> Apr =A03 03:22:10 srv kernel: [288363.070224] Call Trace:
> Apr =A03 03:22:10 srv kernel: [288363.070246] =A0[]=
?
> queue_work_on+0x16/0x20
> Apr =A03 03:22:10 srv kernel: [288363.070266] =A0[]=
?
> md_write_start+0xad/0x190
> Apr =A03 03:22:10 srv kernel: [288363.070283] =A0[]=
?
> autoremove_wake_function+0x0/0x30
> Apr =A03 03:22:10 srv kernel: [288363.070299] =A0[]=
?
> make_request+0x35/0x600
> Apr =A03 03:22:10 srv kernel: [288363.070317] =A0[]=
?
> __alloc_pages_nodemask+0x10b/0x810
> Apr =A03 03:22:10 srv kernel: [288363.070335] =A0[]=
?
> T.1015+0x32/0x90
> Apr =A03 03:22:10 srv kernel: [288363.070350] =A0[]=
?
> md_make_request+0xd4/0x200
> Apr =A03 03:22:10 srv kernel: [288363.070366] =A0[]=
?
> ext4_map_blocks+0x178/0x210
> Apr =A03 03:22:10 srv kernel: [288363.070382] =A0[]=
?
> generic_make_request+0x144/0x2f0
> Apr =A03 03:22:10 srv kernel: [288363.070397] =A0[]=
?
> jbd2_journal_file_buffer+0x3d/0x70
> Apr =A03 03:22:10 srv kernel: [288363.070413] =A0[]=
?
> submit_bio+0x5c/0xd0
> Apr =A03 03:22:10 srv kernel: [288363.070430] =A0[]=
?
> submit_bh+0xe5/0x120
> Apr =A03 03:22:10 srv kernel: [288363.070445] =A0[]=
?
> jbd2_journal_commit_transaction+0x441/0x1180
> Apr =A03 03:22:10 srv kernel: [288363.070466] =A0[]=
?
> lock_timer_base+0x33/0x70
> Apr =A03 03:22:10 srv kernel: [288363.070480] =A0[]=
?
> autoremove_wake_function+0x0/0x30
> Apr =A03 03:22:10 srv kernel: [288363.070498] =A0[]=
?
> kjournald2+0xb1/0x1e0
> Apr =A03 03:22:10 srv kernel: [288363.070511] =A0[]=
?
> autoremove_wake_function+0x0/0x30
> Apr =A03 03:22:10 srv kernel: [288363.070527] =A0[]=
?
> kjournald2+0x0/0x1e0
> Apr =A03 03:22:10 srv kernel: [288363.070544] =A0[]=
?
> kjournald2+0x0/0x1e0
> Apr =A03 03:22:10 srv kernel: [288363.070557] =A0[]=
?
> kthread+0x96/0xa0
> Apr =A03 03:22:10 srv kernel: [288363.070573] =A0[]=
?
> kernel_thread_helper+0x4/0x10
> Apr =A03 03:22:10 srv kernel: [288363.070588] =A0[]=
?
> kthread+0x0/0xa0
> Apr =A03 03:22:10 srv kernel: [288363.070602] =A0[]=
?
> kernel_thread_helper+0x0/0x10
>
> So apparently sdd suffered an unknown failure (it happens) and the ar=
ray
> kicked it out (as it should). But 120 seconds later all tasks accessi=
ng that
> array trigger their 120 second hangcheck warning and are all suck in =
the D
> state.
>
> At the time the array was 12.1% of the way through a reshape. I had t=
o
> reboot the machine to get it back up and it's now continuing the resh=
ape on
> 9 drives.
>
> brad@srv:~$ cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [ra=
id4]
> md0 : active raid6 sdc[0] sdh[9] sda[8] sde[7] sdg[5] sdb[4] sdf[3] s=
dm[2]
> sdl[1]
> =A0 =A0 =A07814078464 blocks super 1.2 level 6, 512k chunk, algorithm=
2 [10/9]
> [UUUUUU_UUU]
> =A0 =A0 =A0[===3D>.................] =A0reshape =3D 16.5% (162091=
008/976759808)
> finish=3D5778.6min speed=3D2349K/sec
>
>
>
> To make matters more confusing the other arrays on the machine were i=
n the
> middle of their "Debians first Sunday of every month" "check" scrub.
>
> I have the full syslog and can probably procure any other information=
that
> might be useful. I don't think I've lost any data, the machine contin=
ued
> reshaping and we're all moving along nicely. I just wanted to report =
it and
> offer assistance in diagnosing it should that be requested.
>
> Regards,
> Brad
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid"=
in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at =A0http://vger.kernel.org/majordomo-info.html
>



--=20
Roberto Spadim
Spadim Technology / SPAEmpresarial
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" i=
n
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array? (No apparent data loss).

am 04.04.2011 07:59:03 von Brad Campbell

On 03/04/11 23:47, Roberto Spadim wrote:
> what kernel version? more informations about your linux box?

The kernel version and architecture were the first 2 lines of the E-mail
you top posted over.

What would you like to know about the box? It's a 6 core Phenom-II with
16G of ram. 2 LSI SAS 9240 controllers configured with 10 x 1TB SATA
Drives in a RAID-6(md0) & 3 x 750GB SATA drives in a RAID-5(md2).

The boot drives are a pair of 1TB SATA drives in multiple RAID-1's using
the on-board AMD chipset controller and there is a 64GB SSD on a
separate PCI-E Marvell 7042m Controller.

The array in question is :

root@srv:~# mdadm --detail /dev/md0
/dev/md0:
Version : 1.2
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Array Size : 7814078464 (7452.09 GiB 8001.62 GB)
Used Dev Size : 976759808 (931.51 GiB 1000.20 GB)
Raid Devices : 10
Total Devices : 9
Persistence : Superblock is persistent

Update Time : Mon Apr 4 13:53:59 2011
State : clean, degraded, recovering
Active Devices : 9
Working Devices : 9
Failed Devices : 0
Spare Devices : 0

Layout : left-symmetric
Chunk Size : 512K

Reshape Status : 29% complete
New Chunksize : 64K

Name : srv:server (local to host srv)
UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Events : 429198

Number Major Minor RaidDevice State
0 8 32 0 active sync /dev/sdc
1 8 176 1 active sync /dev/sdl
2 8 192 2 active sync /dev/sdm
3 8 80 3 active sync /dev/sdf
4 8 16 4 active sync /dev/sdb
5 8 96 5 active sync /dev/sdg
6 0 0 6 removed
7 8 64 7 active sync /dev/sde
8 8 0 8 active sync /dev/sda
9 8 112 9 active sync /dev/sdh
root@srv:~#

Subsequent investigation has shown sdd has a pending reallocation and I
can only assume the unidentified IO error was as a result of tripping up
on that. It still does not explain why all IO to the array froze after
the drive was kicked.

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array? (No apparent data loss).

am 04.04.2011 18:49:24 von Roberto Spadim

i don=B4t know but this happened with me on a hp server, with linux
2,6,37 i changed kernel to a older release and the problem ended,
check with neil and others md guys what=B4s the real problem
maybe realtime module and others changes inside kernel are the
problem, maybe not...
just a quick solution idea: try a older kernel

2011/4/4 Brad Campbell :
> On 03/04/11 23:47, Roberto Spadim wrote:
>>
>> what kernel version? more informations about your linux box?
>
> The kernel version and architecture were the first 2 lines of the E-m=
ail you
> top posted over.
>
> What would you like to know about the box? It's a 6 core Phenom-II wi=
th 16G
> of ram. 2 LSI SAS 9240 controllers configured with 10 x 1TB SATA Driv=
es in a
> RAID-6(md0) & 3 x 750GB SATA drives in a RAID-5(md2).
>
> The boot drives are a pair of 1TB SATA drives in multiple RAID-1's us=
ing the
> on-board AMD chipset controller and there is a 64GB SSD on a separate=
PCI-E
> Marvell 7042m Controller.
>
> The array in question is :
>
> root@srv:~# mdadm --detail /dev/md0
> /dev/md0:
> =A0 =A0 =A0 =A0Version : 1.2
> =A0Creation Time : Sat Jan =A08 11:25:17 2011
> =A0 =A0 Raid Level : raid6
> =A0 =A0 Array Size : 7814078464 (7452.09 GiB 8001.62 GB)
> =A0Used Dev Size : 976759808 (931.51 GiB 1000.20 GB)
> =A0 Raid Devices : 10
> =A0Total Devices : 9
> =A0 =A0Persistence : Superblock is persistent
>
> =A0 =A0Update Time : Mon Apr =A04 13:53:59 2011
> =A0 =A0 =A0 =A0 =A0State : clean, degraded, recovering
> =A0Active Devices : 9
> Working Devices : 9
> =A0Failed Devices : 0
> =A0Spare Devices : 0
>
> =A0 =A0 =A0 =A0 Layout : left-symmetric
> =A0 =A0 Chunk Size : 512K
>
> =A0Reshape Status : 29% complete
> =A0New Chunksize : 64K
>
> =A0 =A0 =A0 =A0 =A0 Name : srv:server =A0(local to host srv)
> =A0 =A0 =A0 =A0 =A0 UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
> =A0 =A0 =A0 =A0 Events : 429198
>
> =A0 =A0Number =A0 Major =A0 Minor =A0 RaidDevice State
> =A0 =A0 =A0 0 =A0 =A0 =A0 8 =A0 =A0 =A0 32 =A0 =A0 =A0 =A00 =A0 =A0 =A0=
active sync =A0 /dev/sdc
> =A0 =A0 =A0 1 =A0 =A0 =A0 8 =A0 =A0 =A0176 =A0 =A0 =A0 =A01 =A0 =A0 =A0=
active sync =A0 /dev/sdl
> =A0 =A0 =A0 2 =A0 =A0 =A0 8 =A0 =A0 =A0192 =A0 =A0 =A0 =A02 =A0 =A0 =A0=
active sync =A0 /dev/sdm
> =A0 =A0 =A0 3 =A0 =A0 =A0 8 =A0 =A0 =A0 80 =A0 =A0 =A0 =A03 =A0 =A0 =A0=
active sync =A0 /dev/sdf
> =A0 =A0 =A0 4 =A0 =A0 =A0 8 =A0 =A0 =A0 16 =A0 =A0 =A0 =A04 =A0 =A0 =A0=
active sync =A0 /dev/sdb
> =A0 =A0 =A0 5 =A0 =A0 =A0 8 =A0 =A0 =A0 96 =A0 =A0 =A0 =A05 =A0 =A0 =A0=
active sync =A0 /dev/sdg
> =A0 =A0 =A0 6 =A0 =A0 =A0 0 =A0 =A0 =A0 =A00 =A0 =A0 =A0 =A06 =A0 =A0=
=A0removed
> =A0 =A0 =A0 7 =A0 =A0 =A0 8 =A0 =A0 =A0 64 =A0 =A0 =A0 =A07 =A0 =A0 =A0=
active sync =A0 /dev/sde
> =A0 =A0 =A0 8 =A0 =A0 =A0 8 =A0 =A0 =A0 =A00 =A0 =A0 =A0 =A08 =A0 =A0=
=A0active sync =A0 /dev/sda
> =A0 =A0 =A0 9 =A0 =A0 =A0 8 =A0 =A0 =A0112 =A0 =A0 =A0 =A09 =A0 =A0 =A0=
active sync =A0 /dev/sdh
> root@srv:~#
>
> Subsequent investigation has shown sdd has a pending reallocation and=
I can
> only assume the unidentified IO error was as a result of tripping up =
on
> that. It still does not explain why all IO to the array froze after t=
he
> drive was kicked.
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid"=
in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at =A0http://vger.kernel.org/majordomo-info.html
>



--=20
Roberto Spadim
Spadim Technology / SPAEmpresarial
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" i=
n
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 05.04.2011 02:47:16 von Brad Campbell

On 05/04/11 00:49, Roberto Spadim wrote:
> i don=B4t know but this happened with me on a hp server, with linux
> 2,6,37 i changed kernel to a older release and the problem ended,
> check with neil and others md guys what=B4s the real problem
> maybe realtime module and others changes inside kernel are the
> problem, maybe not...
> just a quick solution idea: try a older kernel
>

Quick precis:
- Started reshape 512k to 64k chunk size.
- sdd got bad sector and was kicked.
- Array froze all IO.
- Reboot required to get system back.
- Restarted reshape with 9 drives.
- sdl suffered IO error and was kicked
- Array froze all IO.
- Reboot required to get system back.
- Array will no longer mount with 8/10 drives.
- Mdadm 3.1.5 segfaults when trying to start reshape.
Naively tried to run it under gdb to get a backtrace but was unable=20
to stop it forking
- Got array started with mdadm 3.2.1
- Attempted to re-add sdd/sdl (now marked as spares)

root@srv:~/mdadm-3.1.5# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid=
4]
md0 : active raid6 sdl[1](S) sdd[6](S) sdc[0] sdh[9] sda[8] sde[7]=20
sdg[5] sdb[4] sdf[3] sdm[2]
7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2=20
[10/8] [U_UUUU_UUU]
resync=3DDELAYED

md2 : active raid5 sdi[0] sdk[3] sdj[1]
1465146368 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3=
]=20
[UUU]

md6 : active raid1 sdo6[0] sdn6[1]
821539904 blocks [2/2] [UU]

md5 : active raid1 sdo5[0] sdn5[1]
104864192 blocks [2/2] [UU]

md4 : active raid1 sdo3[0] sdn3[1]
20980800 blocks [2/2] [UU]

md3 : active (auto-read-only) raid1 sdo2[0] sdn2[1]
8393856 blocks [2/2] [UU]

md1 : active raid1 sdo1[0] sdn1[1]
20980736 blocks [2/2] [UU]

unused devices:


[ 303.640776] md: bind
[ 303.677461] md: bind
[ 303.837358] md: bind
[ 303.846291] md: bind
[ 303.851476] md: bind
[ 303.860725] md: bind
[ 303.861055] md: bind
[ 303.861982] md: bind
[ 303.862830] md: bind
[ 303.863128] md: bind
[ 303.863306] md: kicking non-fresh sdd from array!
[ 303.863353] md: unbind
[ 303.900207] md: export_rdev(sdd)
[ 303.900260] md: kicking non-fresh sdl from array!
[ 303.900306] md: unbind
[ 303.940100] md: export_rdev(sdl)
[ 303.942181] md/raid:md0: reshape will continue
[ 303.942242] md/raid:md0: device sdc operational as raid disk 0
[ 303.942285] md/raid:md0: device sdh operational as raid disk 9
[ 303.942327] md/raid:md0: device sda operational as raid disk 8
[ 303.942368] md/raid:md0: device sde operational as raid disk 7
[ 303.942409] md/raid:md0: device sdg operational as raid disk 5
[ 303.942449] md/raid:md0: device sdb operational as raid disk 4
[ 303.942490] md/raid:md0: device sdf operational as raid disk 3
[ 303.942531] md/raid:md0: device sdm operational as raid disk 2
[ 303.943733] md/raid:md0: allocated 10572kB
[ 303.943866] md/raid:md0: raid level 6 active with 8 out of 10=20
devices, algorithm 2
[ 303.943912] RAID conf printout:
[ 303.943916] --- level:6 rd:10 wd:8
[ 303.943920] disk 0, o:1, dev:sdc
[ 303.943924] disk 2, o:1, dev:sdm
[ 303.943927] disk 3, o:1, dev:sdf
[ 303.943931] disk 4, o:1, dev:sdb
[ 303.943934] disk 5, o:1, dev:sdg
[ 303.943938] disk 7, o:1, dev:sde
[ 303.943941] disk 8, o:1, dev:sda
[ 303.943945] disk 9, o:1, dev:sdh
[ 303.944061] md0: detected capacity change from 0 to 8001616347136
[ 303.944366] md: md0 switched to read-write mode.
[ 303.944427] md: reshape of RAID array md0
[ 303.944469] md: minimum _guaranteed_ speed: 1000 KB/sec/disk.
[ 303.944511] md: using maximum available idle IO bandwidth (but not=20
more than 200000 KB/sec) for reshape.
[ 303.944573] md: using 128k window, over a total of 976759808 blocks.
[ 304.054875] md0: unknown partition table
[ 304.393245] mdadm[5940]: segfault at 7f2000 ip 00000000004480d2 sp=20
00007fffa04777b8 error 4 in mdadm[400000+64000]


root@srv:~# mdadm --detail /dev/md0
/dev/md0:
Version : 1.2
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Array Size : 7814078464 (7452.09 GiB 8001.62 GB)
Used Dev Size : 976759808 (931.51 GiB 1000.20 GB)
Raid Devices : 10
Total Devices : 10
Persistence : Superblock is persistent

Update Time : Tue Apr 5 07:54:30 2011
State : active, degraded
Active Devices : 8
Working Devices : 10
Failed Devices : 0
Spare Devices : 2

Layout : left-symmetric
Chunk Size : 512K

New Chunksize : 64K

Name : srv:server (local to host srv)
UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Events : 633835

Number Major Minor RaidDevice State
0 8 32 0 active sync /dev/sdc
1 0 0 1 removed
2 8 192 2 active sync /dev/sdm
3 8 80 3 active sync /dev/sdf
4 8 16 4 active sync /dev/sdb
5 8 96 5 active sync /dev/sdg
6 0 0 6 removed
7 8 64 7 active sync /dev/sde
8 8 0 8 active sync /dev/sda
9 8 112 9 active sync /dev/sdh

1 8 176 - spare /dev/sdl
6 8 48 - spare /dev/sdd

root@srv:~# for i in /dev/sd? ; do mdadm --examine $i ; done
/dev/sda:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 9beb9a0f:2a73328c:f0c17909:89da70fd

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : c58ed095 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 8
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdb:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 75d997f8:d9372d90:c068755b:81c8206b

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : 72321703 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 4
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdc:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 5738a232:85f23a16:0c7a9454:d770199c

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : 5c61ea2e - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 0
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdd:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 83a2c731:ba2846d0:2ce97d83:de624339

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : e1a5ebbc - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : spare
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sde:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : f1e3c1d3:ea9dc52e:a4e6b70e:e25a0321

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : 551997d7 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 7
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdf:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : c32dff71:0b8c165c:9f589b0f:bcbc82da

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : db0aa39b - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 3
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdg:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 194bc75c:97d3f507:4915b73a:51a50172

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : 344cadbe - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 5
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdh:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 1326457e:4fc0a6be:0073ccae:398d5c7f

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : 8debbb14 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 9
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdi:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : e39d73c3:75be3b52:44d195da:b240c146
Name : srv:2 (local to host srv)
Creation Time : Sat Jul 10 21:14:29 2010
Raid Level : raid5
Raid Devices : 3

Avail Dev Size : 1465147120 (698.64 GiB 750.16 GB)
Array Size : 2930292736 (1397.27 GiB 1500.31 GB)
Used Dev Size : 1465146368 (698.64 GiB 750.15 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : clean
Device UUID : b577b308:56f2e4c9:c78175f4:cf10c77f

Update Time : Tue Apr 5 07:46:18 2011
Checksum : 57ee683f - correct
Events : 455775

Layout : left-symmetric
Chunk Size : 64K

Device Role : Active device 0
Array State : AAA ('A' == active, '.' == missing)
/dev/sdj:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : e39d73c3:75be3b52:44d195da:b240c146
Name : srv:2 (local to host srv)
Creation Time : Sat Jul 10 21:14:29 2010
Raid Level : raid5
Raid Devices : 3

Avail Dev Size : 1465147120 (698.64 GiB 750.16 GB)
Array Size : 2930292736 (1397.27 GiB 1500.31 GB)
Used Dev Size : 1465146368 (698.64 GiB 750.15 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : clean
Device UUID : b127f002:a4aa8800:735ef8d7:6018564e

Update Time : Tue Apr 5 07:46:18 2011
Checksum : 3ae0b4c6 - correct
Events : 455775

Layout : left-symmetric
Chunk Size : 64K

Device Role : Active device 1
Array State : AAA ('A' == active, '.' == missing)
/dev/sdk:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : e39d73c3:75be3b52:44d195da:b240c146
Name : srv:2 (local to host srv)
Creation Time : Sat Jul 10 21:14:29 2010
Raid Level : raid5
Raid Devices : 3

Avail Dev Size : 1465147120 (698.64 GiB 750.16 GB)
Array Size : 2930292736 (1397.27 GiB 1500.31 GB)
Used Dev Size : 1465146368 (698.64 GiB 750.15 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : clean
Device UUID : 90fddf63:03d5dba4:3fcdc476:9ce3c44c

Update Time : Tue Apr 5 07:46:18 2011
Checksum : dd5eef0e - correct
Events : 455775

Layout : left-symmetric
Chunk Size : 64K

Device Role : Active device 2
Array State : AAA ('A' == active, '.' == missing)
/dev/sdl:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 769940af:66733069:37cea27d:7fb28a23

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : dc756202 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : spare
Array State : A.AAAA.AAA ('A' == active, '.' == missing)
/dev/sdm:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x4
Array UUID : d00a11d7:fe0435af:07c8d4d6:e3b8e34e
Name : srv:server (local to host srv)
Creation Time : Sat Jan 8 11:25:17 2011
Raid Level : raid6
Raid Devices : 10

Avail Dev Size : 1953523120 (931.51 GiB 1000.20 GB)
Array Size : 15628156928 (7452.09 GiB 8001.62 GB)
Used Dev Size : 1953519616 (931.51 GiB 1000.20 GB)
Data Offset : 2048 sectors
Super Offset : 8 sectors
State : active
Device UUID : 7e564e2c:7f21125b:c3b1907a:b640178f

Reshape pos'n : 3437035520 (3277.81 GiB 3519.52 GB)
New Chunksize : 64K

Update Time : Tue Apr 5 07:54:30 2011
Checksum : b3df3ee7 - correct
Events : 633835

Layout : left-symmetric
Chunk Size : 512K

Device Role : Active device 2
Array State : A.AAAA.AAA ('A' == active, '.' == missing)

root@srv:~/mdadm-3.1.5# ./mdadm --version
mdadm - v3.1.5 - 23rd March 2011

root@srv:~/mdadm-3.1.5# uname -a
Linux srv 2.6.38 #19 SMP Wed Mar 23 09:57:05 WST 2011 x86_64 GNU/Linux

Now. The array restarted with mdadm 3.2.1, but of course its now=20
reshaping 8 out of 10 disks, has no redundancy and is going at 600k/s=20
which will take over 10 days. Is there anything I can do to give it som=
e=20
redundancy while it completes or am I better to copy the data off, blow=
=20
it away and start again? All the important stuff is backed up anyway, I=
=20
just wanted to avoid restoring 8TB from backup if I could.

Regards,
Brad
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" i=
n
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 05.04.2011 08:10:43 von NeilBrown

On Tue, 05 Apr 2011 08:47:16 +0800 Brad Campbell com>
wrote:

> On 05/04/11 00:49, Roberto Spadim wrote:
> > i don=B4t know but this happened with me on a hp server, with linux
> > 2,6,37 i changed kernel to a older release and the problem ended,
> > check with neil and others md guys what=B4s the real problem
> > maybe realtime module and others changes inside kernel are the
> > problem, maybe not...
> > just a quick solution idea: try a older kernel
> >
>=20
> Quick precis:
> - Started reshape 512k to 64k chunk size.
> - sdd got bad sector and was kicked.
> - Array froze all IO.

That .... shouldn't happen. But I know why it did.

mdadm forks and runs in the back ground monitoring the reshape.
It suspends IO to a region of the array, backs up the data, then lets t=
he
reshape progress over that region, then invalidates the backup and allo=
ws IO
to resume, then moves on to the next region (it actually have two regio=
ns in
different states at the same time, but you get the idea).

If the device failed the reshape in the kernel aborted and then restart=
ed.
It is meant to do this - restore to a known state, then decide if there=
is
anything useful to do. It restarts exactly where it left off so all sh=
ould
be fine.

mdadm periodically checks the value in 'sync_completed' to see how far =
the
reshape has progressed to know if it can move on.
If it checks while the reshape is temporarily aborted it sees 'none', w=
hich
is not a number, so it aborts. That should be fixed.
It aborts with IO to a region still suspended so it is very possible fo=
r IO
to freeze if anything is destined for that region.

> - Reboot required to get system back.
> - Restarted reshape with 9 drives.
> - sdl suffered IO error and was kicked

Very sad.

> - Array froze all IO.

Same thing...

> - Reboot required to get system back.
> - Array will no longer mount with 8/10 drives.
> - Mdadm 3.1.5 segfaults when trying to start reshape.

Don't know why it would have done that... I cannot reproduce it easily.


> Naively tried to run it under gdb to get a backtrace but was unabl=
e=20
> to stop it forking

Yes, tricky .... an "strace -o /tmp/file -f mdadm ...." might have been
enough, but to late to worry about that now.

> - Got array started with mdadm 3.2.1
> - Attempted to re-add sdd/sdl (now marked as spares)

Hmm... it isn't meant to do that any more. I thought I fixed it so tha=
t it
if a device looked like part of the array it wouldn't add it as a spare=
..
Obviously that didn't work. I'd better look in to it again.


> [ 304.393245] mdadm[5940]: segfault at 7f2000 ip 00000000004480d2 sp=
=20
> 00007fffa04777b8 error 4 in mdadm[400000+64000]
>=20

If you have the exact mdadm binary that caused this segfault we should =
be
able to figure out what instruction was at 0004480d2. If you don't fe=
el up
to it, could you please email me the file privately and I'll have a loo=
k.


> root@srv:~/mdadm-3.1.5# uname -a
> Linux srv 2.6.38 #19 SMP Wed Mar 23 09:57:05 WST 2011 x86_64 GNU/Linu=
x
>=20
> Now. The array restarted with mdadm 3.2.1, but of course its now=20
> reshaping 8 out of 10 disks, has no redundancy and is going at 600k/s=
=20
> which will take over 10 days. Is there anything I can do to give it s=
ome=20
> redundancy while it completes or am I better to copy the data off, bl=
ow=20
> it away and start again? All the important stuff is backed up anyway,=
I=20
> just wanted to avoid restoring 8TB from backup if I could.

No, you cannot give it extra redundancy.
I would suggest:
copy anything that you need off, just in case - if you can.

Kill the mdadm that is running in the back ground. This will mean th=
at
if the machine crashes your array will be corrupted, but you are thin=
king
of rebuilding it any, so that isn't the end of the world.
In /sys/block/md0/md
cat suspend_hi > suspend_lo
cat component_size > sync_max

That will allow the reshape to continue without any backup. It will =
be
much faster (but less safe, as I said).

If the reshape completes without incident, it will start recovering t=
o the
two 'spares' - and then you will have a happy array again.

If something goes wrong, you will need to scrap the array, recreate i=
t, and
copy data back from where-ever you copied it to (or backups).

If anything there doesn't make sense, or doesn't seem to work - please =
ask.

Thanks for the report. I'll try to get those mdadm issues addressed -
particularly if you can get me the mdadm file which caused the segfault=


NeilBrown
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" i=
n
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 05.04.2011 11:02:43 von Brad Campbell

On 05/04/11 14:10, NeilBrown wrote:
>> - Reboot required to get system back.
>> - Restarted reshape with 9 drives.
>> - sdl suffered IO error and was kicked
>
> Very sad.

I'd say pretty damn unlucky actually.

>> - Array froze all IO.
>
> Same thing...
>
>> - Reboot required to get system back.
>> - Array will no longer mount with 8/10 drives.
>> - Mdadm 3.1.5 segfaults when trying to start reshape.
>
> Don't know why it would have done that... I cannot reproduce it easily.

No. I tried numerous incantations. The system version of mdadm is Debian
3.1.4. This segfaulted so I downloaded and compiled 3.1.5 which did the
same thing. I then composed most of this E-mail, made *really* sure my
backups were up to date and tried 3.2.1 which to my astonishment worked.
It's been ticking along _slowly_ ever since.

>> Naively tried to run it under gdb to get a backtrace but was unable
>> to stop it forking
>
> Yes, tricky .... an "strace -o /tmp/file -f mdadm ...." might have been
> enough, but to late to worry about that now.

I wondered about using strace but for some reason got it into my head
that a gdb backtrace would be more useful. Then of course I got it
started with 3.2.1 and have not tried again.

>> - Got array started with mdadm 3.2.1
>> - Attempted to re-add sdd/sdl (now marked as spares)
>
> Hmm... it isn't meant to do that any more. I thought I fixed it so
that it
> if a device looked like part of the array it wouldn't add it as a
spare...
> Obviously that didn't work. I'd better look in to it again.

Now the chain of events that led up to this was along these lines.
- Rebooted machine.
- Tried to --assemble with 3.1.4
- mdadm told me it did not really want to continue with 8/10 devices and
I should use --force if I really wanted it to try.
- I used --force
- I did a mdadm --add /dev/md0 /dev/sdd and the same for sdl
- I checked and they were listed as spares.

So this was all done with Debian's mdadm 3.1.4, *not* 3.1.5

>
> No, you cannot give it extra redundancy.
> I would suggest:
> copy anything that you need off, just in case - if you can.
>
> Kill the mdadm that is running in the back ground. This will mean
that
> if the machine crashes your array will be corrupted, but you are
thinking
> of rebuilding it any, so that isn't the end of the world.
> In /sys/block/md0/md
> cat suspend_hi> suspend_lo
> cat component_size> sync_max
>
> That will allow the reshape to continue without any backup. It
will be
> much faster (but less safe, as I said).

Well, I have nothing to lose, but I've just picked up some extra drives
so I'll make second backups and then give this a whirl.

> If something goes wrong, you will need to scrap the array,
recreate it, and
> copy data back from where-ever you copied it to (or backups).

I did go into this with the niggling feeling that something bad might
happen, so I made sure all my backups were up to date before I started.
No biggie if it does die.

The very odd thing is I did a complete array check, plus SMART long
tests on all drives literally hours before I started the reshape. Goes
to show how ropey these large drives can be in big(iash) arrays.

> If anything there doesn't make sense, or doesn't seem to work -
please ask.
>
> Thanks for the report. I'll try to get those mdadm issues addressed -
> particularly if you can get me the mdadm file which caused the segfault.
>

Well, luckily I preserved the entire build tree then. I was planning on
running nm over the binary and have a two thumbs type of look into it
with gdb, but seeing as you probably have a much better idea what you
are looking for I'll just send you the binary!

Thanks for the help Neil. Much appreciated.

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 05.04.2011 13:31:11 von NeilBrown

On Tue, 05 Apr 2011 17:02:43 +0800 Brad Campbell
wrote:

> Well, luckily I preserved the entire build tree then. I was planning on
> running nm over the binary and have a two thumbs type of look into it
> with gdb, but seeing as you probably have a much better idea what you
> are looking for I'll just send you the binary!

Thanks. It took me a little while, but I've found the problem.

The code was failing at
wd0 = sources[z][d];
in qsyndrome in restripe.c.
It is looking up 'd' in 'sources[z]' and having problems.
The error address (from dmesg) is 0x7f2000 so it isn't a
NULL pointer, but rather it is falling off the end of an allocation.

When doing qsyndrome calculations we often need a block full of zeros, so
restripe.c allocates one and stores it in in a global pointer.

You were restriping from 512K to 64K chunk size.

The first thing restripe.c was called on to do was to restore data from the
backup file into the array. This uses the new chunk size - 64K. So the
'zero' buffer was allocated at 64K and cleared.


The next thing it does is read the next section of the array and write it to
the backup. As the array was missing 2 devices it needed to do a qsyndrome
calculation to get the missing data block(s). This was a calculation done on
old-style chunks so it needed a 512K zero block.
However as a zero block had already been allocated it didn't bother to
allocate another one. It just used what it had, which was too small.
So it fell off the end and got the result we saw.

I don't know why this works in 3.2.1 where it didn't work in 3.1.4.

However when it successfully recovers from the backup it should update the
metadata so that it knows it has successfully recovered and doesn't need to
recover any more. So maybe the time it worked, it found there wasn't any
recovery needed and so didn't allocate a 'zero' buffer until it was working
with the old, bigger, chunk size.

Anyway, this is easy to fix which I will do.

It only affects restarting a reshape of a double-degraded RAID6 which reduced
the chunksize.

Thanks,
NeilBrown
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 05.04.2011 13:47:59 von Brad Campbell

On 05/04/11 19:31, NeilBrown wrote:

> It only affects restarting a reshape of a double-degraded RAID6 which reduced
> the chunksize.

Talk about the planets falling into alignment! I was obviously holding
my mouth wrong when I pushed the enter key.

Thanks for the explanation. Much appreciated (again!)


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 08.04.2011 03:19:01 von Brad Campbell

On 05/04/11 14:10, NeilBrown wrote:

> I would suggest:
> copy anything that you need off, just in case - if you can.
>
> Kill the mdadm that is running in the back ground. This will mean that
> if the machine crashes your array will be corrupted, but you are thinking
> of rebuilding it any, so that isn't the end of the world.
> In /sys/block/md0/md
> cat suspend_hi> suspend_lo
> cat component_size> sync_max
>

root@srv:/sys/block/md0/md# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
sdg[5] sdb[4] sdf[3] sdm[2]
7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2
[10/8] [U_UUUU_UUU]
[=================>...] reshape = 88.2% (861696000/976759808)
finish=3713.3min speed=516K/sec

md2 : active raid5 sdi[0] sdk[3] sdj[1]
1465146368 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3]
[UUU]

md6 : active raid1 sdp6[0] sdo6[1]
821539904 blocks [2/2] [UU]

md5 : active raid1 sdp5[0] sdo5[1]
104864192 blocks [2/2] [UU]

md4 : active raid1 sdp3[0] sdo3[1]
20980800 blocks [2/2] [UU]

md3 : active raid1 sdp2[0] sdo2[1]
8393856 blocks [2/2] [UU]

md1 : active raid1 sdp1[0] sdo1[1]
20980736 blocks [2/2] [UU]

unused devices:
root@srv:/sys/block/md0/md# cat component_size > sync_max
cat: write error: Device or resource busy

root@srv:/sys/block/md0/md# cat suspend_hi suspend_lo
13788774400
13788774400

root@srv:/sys/block/md0/md# grep . sync_*
sync_action:reshape
sync_completed:1723392000 / 1953519616
sync_force_parallel:0
sync_max:1723392000
sync_min:0
sync_speed:281
sync_speed_max:200000 (system)
sync_speed_min:200000 (local)

So I killed mdadm, then did the cat suspend_hi > suspend_lo.. but as you
can see it won't let me change sync_max. The array above reports
516K/sec, but that was just on its way down to 0 on a time based
average. It was not moving at all.

I then tried stopping the array, restarting it with mdadm 3.1.4 which
immediately segfaulted and left the array in state resync=DELAYED.

I issued the above commands again, which succeeded this time but while
the array looked good, it was not resyncing :
root@srv:/sys/block/md0/md# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
sdg[5] sdb[4] sdf[3] sdm[2]
7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2
[10/8] [U_UUUU_UUU]
[=================>...] reshape = 88.2% (861698048/976759808)
finish=30203712.0min speed=0K/sec

md2 : active raid5 sdi[0] sdk[3] sdj[1]
1465146368 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3]
[UUU]

md6 : active raid1 sdp6[0] sdo6[1]
821539904 blocks [2/2] [UU]

md5 : active raid1 sdp5[0] sdo5[1]
104864192 blocks [2/2] [UU]

md4 : active raid1 sdp3[0] sdo3[1]
20980800 blocks [2/2] [UU]

md3 : active raid1 sdp2[0] sdo2[1]
8393856 blocks [2/2] [UU]

md1 : active raid1 sdp1[0] sdo1[1]
20980736 blocks [2/2] [UU]

unused devices:

root@srv:/sys/block/md0/md# grep . sync*
sync_action:reshape
sync_completed:1723396096 / 1953519616
sync_force_parallel:0
sync_max:976759808
sync_min:0
sync_speed:0
sync_speed_max:200000 (system)
sync_speed_min:200000 (local)

I stopped the array and restarted it with mdadm 3.2.1 and it continued
along its merry way.

Not an issue, and I don't much care if it blew something up, but I
thought it worthy of a follow up.

If there is anything you need tested while it's in this state I've got ~
1000 minutes of resync time left and I'm happy to damage it if requested.

Regards,
Brad
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 08.04.2011 11:52:38 von NeilBrown

On Fri, 08 Apr 2011 09:19:01 +0800 Brad Campbell
wrote:

> On 05/04/11 14:10, NeilBrown wrote:
>
> > I would suggest:
> > copy anything that you need off, just in case - if you can.
> >
> > Kill the mdadm that is running in the back ground. This will mean that
> > if the machine crashes your array will be corrupted, but you are thinking
> > of rebuilding it any, so that isn't the end of the world.
> > In /sys/block/md0/md
> > cat suspend_hi> suspend_lo
> > cat component_size> sync_max
> >
>
> root@srv:/sys/block/md0/md# cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
> sdg[5] sdb[4] sdf[3] sdm[2]
> 7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2
> [10/8] [U_UUUU_UUU]
> [=================>...] reshape = 88.2% (861696000/976759808)
> finish=3713.3min speed=516K/sec
>
> md2 : active raid5 sdi[0] sdk[3] sdj[1]
> 1465146368 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3]
> [UUU]
>
> md6 : active raid1 sdp6[0] sdo6[1]
> 821539904 blocks [2/2] [UU]
>
> md5 : active raid1 sdp5[0] sdo5[1]
> 104864192 blocks [2/2] [UU]
>
> md4 : active raid1 sdp3[0] sdo3[1]
> 20980800 blocks [2/2] [UU]
>
> md3 : active raid1 sdp2[0] sdo2[1]
> 8393856 blocks [2/2] [UU]
>
> md1 : active raid1 sdp1[0] sdo1[1]
> 20980736 blocks [2/2] [UU]
>
> unused devices:
> root@srv:/sys/block/md0/md# cat component_size > sync_max
> cat: write error: Device or resource busy

Sorry, I should have checked the source code.


echo max > sync_max

is what you want.
Or just a much bigger number.

>
> root@srv:/sys/block/md0/md# cat suspend_hi suspend_lo
> 13788774400
> 13788774400

They are the same so that is good - nothing will be suspended.

>
> root@srv:/sys/block/md0/md# grep . sync_*
> sync_action:reshape
> sync_completed:1723392000 / 1953519616
> sync_force_parallel:0
> sync_max:1723392000
> sync_min:0
> sync_speed:281
> sync_speed_max:200000 (system)
> sync_speed_min:200000 (local)
>
> So I killed mdadm, then did the cat suspend_hi > suspend_lo.. but as you
> can see it won't let me change sync_max. The array above reports
> 516K/sec, but that was just on its way down to 0 on a time based
> average. It was not moving at all.
>
> I then tried stopping the array, restarting it with mdadm 3.1.4 which
> immediately segfaulted and left the array in state resync=DELAYED.
>
> I issued the above commands again, which succeeded this time but while
> the array looked good, it was not resyncing :
> root@srv:/sys/block/md0/md# cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
> sdg[5] sdb[4] sdf[3] sdm[2]
> 7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2
> [10/8] [U_UUUU_UUU]
> [=================>...] reshape = 88.2% (861698048/976759808)
> finish=30203712.0min speed=0K/sec
>
> md2 : active raid5 sdi[0] sdk[3] sdj[1]
> 1465146368 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3]
> [UUU]
>
> md6 : active raid1 sdp6[0] sdo6[1]
> 821539904 blocks [2/2] [UU]
>
> md5 : active raid1 sdp5[0] sdo5[1]
> 104864192 blocks [2/2] [UU]
>
> md4 : active raid1 sdp3[0] sdo3[1]
> 20980800 blocks [2/2] [UU]
>
> md3 : active raid1 sdp2[0] sdo2[1]
> 8393856 blocks [2/2] [UU]
>
> md1 : active raid1 sdp1[0] sdo1[1]
> 20980736 blocks [2/2] [UU]
>
> unused devices:
>
> root@srv:/sys/block/md0/md# grep . sync*
> sync_action:reshape
> sync_completed:1723396096 / 1953519616
> sync_force_parallel:0
> sync_max:976759808
> sync_min:0
> sync_speed:0
> sync_speed_max:200000 (system)
> sync_speed_min:200000 (local)
>
> I stopped the array and restarted it with mdadm 3.2.1 and it continued
> along its merry way.
>
> Not an issue, and I don't much care if it blew something up, but I
> thought it worthy of a follow up.
>
> If there is anything you need tested while it's in this state I've got ~
> 1000 minutes of resync time left and I'm happy to damage it if requested.

No thank - I think I know what happened. Main problem is that there is
confusion between 'k' and 'sectors' and there are random other values that
sometimes work (like 'max') and I never remember which is which. sysfs in md
is a bit of a mess.... one day I hope to completely replace it (with back
compatibility of course...)

Thanks for the feedback.

NeilBrown


>
> Regards,
> Brad
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: What the heck happened to my array?

am 08.04.2011 17:27:55 von Roberto Spadim

hi neil, with time i could help changin sysfs information (add unit
information at output, and remove it (unit text) from input)
what's the current kernel version of develop?

2011/4/8 NeilBrown :
> On Fri, 08 Apr 2011 09:19:01 +0800 Brad Campbell le.com>
> wrote:
>
>> On 05/04/11 14:10, NeilBrown wrote:
>>
>> > I would suggest:
>> > =A0 =A0copy anything that you need off, just in case - if you can.
>> >
>> > =A0 =A0Kill the mdadm that is running in the back ground. =A0This =
will mean that
>> > =A0 =A0if the machine crashes your array will be corrupted, but yo=
u are thinking
>> > =A0 =A0of rebuilding it any, so that isn't the end of the world.
>> > =A0 =A0In /sys/block/md0/md
>> > =A0 =A0 =A0 cat suspend_hi> =A0suspend_lo
>> > =A0 =A0 =A0 cat component_size> =A0sync_max
>> >
>>
>> root@srv:/sys/block/md0/md# cat /proc/mdstat
>> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [r=
aid4]
>> md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
>> sdg[5] sdb[4] sdf[3] sdm[2]
>> =A0 =A0 =A0 =A07814078464 blocks super 1.2 level 6, 512k chunk, algo=
rithm 2
>> [10/8] [U_UUUU_UUU]
>> =A0 =A0 =A0 =A0[=================3D>=
..] =A0reshape =3D 88.2% (861696000/976759808)
>> finish=3D3713.3min speed=3D516K/sec
>>
>> md2 : active raid5 sdi[0] sdk[3] sdj[1]
>> =A0 =A0 =A0 =A01465146368 blocks super 1.2 level 5, 64k chunk, algor=
ithm 2 [3/3]
>> [UUU]
>>
>> md6 : active raid1 sdp6[0] sdo6[1]
>> =A0 =A0 =A0 =A0821539904 blocks [2/2] [UU]
>>
>> md5 : active raid1 sdp5[0] sdo5[1]
>> =A0 =A0 =A0 =A0104864192 blocks [2/2] [UU]
>>
>> md4 : active raid1 sdp3[0] sdo3[1]
>> =A0 =A0 =A0 =A020980800 blocks [2/2] [UU]
>>
>> md3 : active raid1 sdp2[0] sdo2[1]
>> =A0 =A0 =A0 =A08393856 blocks [2/2] [UU]
>>
>> md1 : active raid1 sdp1[0] sdo1[1]
>> =A0 =A0 =A0 =A020980736 blocks [2/2] [UU]
>>
>> unused devices:
>> root@srv:/sys/block/md0/md# cat component_size > sync_max
>> cat: write error: Device or resource busy
>
> Sorry, I should have checked the source code.
>
>
> =A0 echo max > sync_max
>
> is what you want.
> Or just a much bigger number.
>
>>
>> root@srv:/sys/block/md0/md# cat suspend_hi suspend_lo
>> 13788774400
>> 13788774400
>
> They are the same so that is good - nothing will be suspended.
>
>>
>> root@srv:/sys/block/md0/md# grep . sync_*
>> sync_action:reshape
>> sync_completed:1723392000 / 1953519616
>> sync_force_parallel:0
>> sync_max:1723392000
>> sync_min:0
>> sync_speed:281
>> sync_speed_max:200000 (system)
>> sync_speed_min:200000 (local)
>>
>> So I killed mdadm, then did the cat suspend_hi > suspend_lo.. but as=
you
>> can see it won't let me change sync_max. The array above reports
>> 516K/sec, but that was just on its way down to 0 on a time based
>> average. It was not moving at all.
>>
>> I then tried stopping the array, restarting it with mdadm 3.1.4 whic=
h
>> immediately segfaulted and left the array in state resync=3DDELAYED.
>>
>> I issued the above commands again, which succeeded this time but whi=
le
>> the array looked good, it was not resyncing :
>> root@srv:/sys/block/md0/md# cat /proc/mdstat
>> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [r=
aid4]
>> md0 : active raid6 sdc[0] sdd[6](S) sdl[1](S) sdh[9] sda[8] sde[7]
>> sdg[5] sdb[4] sdf[3] sdm[2]
>> =A0 =A0 =A0 =A07814078464 blocks super 1.2 level 6, 512k chunk, algo=
rithm 2
>> [10/8] [U_UUUU_UUU]
>> =A0 =A0 =A0 =A0[=================3D>=
..] =A0reshape =3D 88.2% (861698048/976759808)
>> finish=3D30203712.0min speed=3D0K/sec
>>
>> md2 : active raid5 sdi[0] sdk[3] sdj[1]
>> =A0 =A0 =A0 =A01465146368 blocks super 1.2 level 5, 64k chunk, algor=
ithm 2 [3/3]
>> [UUU]
>>
>> md6 : active raid1 sdp6[0] sdo6[1]
>> =A0 =A0 =A0 =A0821539904 blocks [2/2] [UU]
>>
>> md5 : active raid1 sdp5[0] sdo5[1]
>> =A0 =A0 =A0 =A0104864192 blocks [2/2] [UU]
>>
>> md4 : active raid1 sdp3[0] sdo3[1]
>> =A0 =A0 =A0 =A020980800 blocks [2/2] [UU]
>>
>> md3 : active raid1 sdp2[0] sdo2[1]
>> =A0 =A0 =A0 =A08393856 blocks [2/2] [UU]
>>
>> md1 : active raid1 sdp1[0] sdo1[1]
>> =A0 =A0 =A0 =A020980736 blocks [2/2] [UU]
>>
>> unused devices:
>>
>> root@srv:/sys/block/md0/md# grep . sync*
>> sync_action:reshape
>> sync_completed:1723396096 / 1953519616
>> sync_force_parallel:0
>> sync_max:976759808
>> sync_min:0
>> sync_speed:0
>> sync_speed_max:200000 (system)
>> sync_speed_min:200000 (local)
>>
>> I stopped the array and restarted it with mdadm 3.2.1 and it continu=
ed
>> along its merry way.
>>
>> Not an issue, and I don't much care if it blew something up, but I
>> thought it worthy of a follow up.
>>
>> If there is anything you need tested while it's in this state I've g=
ot ~
>> 1000 minutes of resync time left and I'm happy to damage it if reque=
sted.
>
> No thank - I think I know what happened. =A0Main problem is that ther=
e is
> confusion between 'k' and 'sectors' and there are random other values=
that
> sometimes work (like 'max') and I never remember which is which. =A0s=
ysfs in md
> is a bit of a mess.... one day I hope to completely replace it (with =
back
> compatibility of course...)
>
> Thanks for the feedback.
>
> NeilBrown
>
>
>>
>> Regards,
>> Brad
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-raid=
" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at =A0http://vger.kernel.org/majordomo-info.html
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid"=
in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at =A0http://vger.kernel.org/majordomo-info.html
>



--=20
Roberto Spadim
Spadim Technology / SPAEmpresarial
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" i=
n
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html