Bookmarks

Yahoo Gmail Google Facebook Delicious Twitter Reddit Stumpleupon Myspace Digg

Search queries

sqldatasource dal, wwwxxxenden, convert raid5 to raid 10 mdadm, apache force chunked, nrao wwwxxx, xxxxxdup, procmail change subject header, wwwXxx not20, Wwwxxx.doks sas, linux raid resync after reboot

Links

XODOX
Impressum

#1: always dirty RAID5 arrays

Posted on 2004-11-09 22:03:24 by Scott Ransom

Hi All,

On one of our servers we have 4 6-disk RAID5 arrays running. Each of the
arrays was created using the following command:

mdadm --create /dev/md3 --level=5 --verbose --force --chunk=128 \
--raid-devices=6 /dev/sd[ijklmn]1

After building and a resync (and a reboot and a resync...), the array looks
like this:

--------------------------------------------
spigot2 sransom 84: mdadm --detail /dev/md3
/dev/md3:
Version : 00.90.00
Creation Time : Tue Sep 21 15:54:31 2004
Raid Level : raid5
Array Size : 1220979200 (1164.42 GiB 1250.28 GB)
Device Size : 244195840 (232.88 GiB 250.06 GB)
Raid Devices : 6
Total Devices : 6
Preferred Minor : 3
Persistence : Superblock is persistent

Update Time : Sat Nov 6 10:58:12 2004
State : dirty
Active Devices : 6
Working Devices : 6
Failed Devices : 0
Spare Devices : 0

Layout : left-symmetric
Chunk Size : 128K

UUID : 264b9069:96d7a1a6:d3b17be5:23fa47ce
Events : 0.30

Number Major Minor RaidDevice State
0 8 129 0 active sync /dev/sdi1
1 8 145 1 active sync /dev/sdj1
2 8 161 2 active sync /dev/sdk1
3 8 177 3 active sync /dev/sdl1
4 8 193 4 active sync /dev/sdm1
5 8 209 5 active sync /dev/sdn1
------------------------------------------------------------ -----


Notice that the State is "dirty". If we reboot, the arrays are in a dirty
status and always need to resync. Any idea why would this be? (I'm
running XFS on the RAIDs and am using a slightly modified RH9.0 kernel:
2.4.24aa1-xfs)

Thanks for the help,

Scott

--
Scott M. Ransom Address: NRAO
Phone: (434) 296-0320 520 Edgemont Rd.
email: sransom@nrao.edu Charlottesville, VA 22903 USA
GPG Fingerprint: 06A9 9553 78BE 16DB 407B FFCA 9BFA B6FF FFD3 2989
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Report this message

#2: RE: always dirty RAID5 arrays

Posted on 2004-11-09 23:27:10 by bugzilla

Dirty does not indicate a re-sync is needed, just that changes have
occurred. If the array were shutdown incorrectly (like a crash) then on
re-start a re-build of parity would be done.

For more details:
man md

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Scott Ransom
Sent: Tuesday, November 09, 2004 4:03 PM
To: linux-raid@vger.kernel.org
Cc: sransom@nrao.edu
Subject: always dirty RAID5 arrays

Hi All,

On one of our servers we have 4 6-disk RAID5 arrays running. Each of the
arrays was created using the following command:

mdadm --create /dev/md3 --level=5 --verbose --force --chunk=128 \
--raid-devices=6 /dev/sd[ijklmn]1

After building and a resync (and a reboot and a resync...), the array looks
like this:

--------------------------------------------
spigot2 sransom 84: mdadm --detail /dev/md3
/dev/md3:
Version : 00.90.00
Creation Time : Tue Sep 21 15:54:31 2004
Raid Level : raid5
Array Size : 1220979200 (1164.42 GiB 1250.28 GB)
Device Size : 244195840 (232.88 GiB 250.06 GB)
Raid Devices : 6
Total Devices : 6
Preferred Minor : 3
Persistence : Superblock is persistent

Update Time : Sat Nov 6 10:58:12 2004
State : dirty
Active Devices : 6
Working Devices : 6
Failed Devices : 0
Spare Devices : 0

Layout : left-symmetric
Chunk Size : 128K

UUID : 264b9069:96d7a1a6:d3b17be5:23fa47ce
Events : 0.30

Number Major Minor RaidDevice State
0 8 129 0 active sync /dev/sdi1
1 8 145 1 active sync /dev/sdj1
2 8 161 2 active sync /dev/sdk1
3 8 177 3 active sync /dev/sdl1
4 8 193 4 active sync /dev/sdm1
5 8 209 5 active sync /dev/sdn1
------------------------------------------------------------ -----


Notice that the State is "dirty". If we reboot, the arrays are in a dirty
status and always need to resync. Any idea why would this be? (I'm
running XFS on the RAIDs and am using a slightly modified RH9.0 kernel:
2.4.24aa1-xfs)

Thanks for the help,

Scott

--
Scott M. Ransom Address: NRAO
Phone: (434) 296-0320 520 Edgemont Rd.
email: sransom@nrao.edu Charlottesville, VA 22903 USA
GPG Fingerprint: 06A9 9553 78BE 16DB 407B FFCA 9BFA B6FF FFD3 2989
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html

Report this message