Antw: Problems creating MD-RAID1: "device .. not suitable forany style of raid array" / &q
am 11.05.2011 16:38:24 von Ulrich WindlHi!
I could reduce the problem to "--bitmap internal": It works if I leave it out. Unfortunately the RAID will fully resync every assembly when there was an unclean shutdown.
The thing that kept one device busy was MD-RAID itself: "mdadm -- stop" released the device.
So to summarize:
1) mdadm fails to set up an internal bitmap
2) Even when mdadm fails to do 1), it starts the array
Regards,
Ulrich
>>> Ulrich Windl schrieb am 06.05.2011 um 14:15 in Nachricht <4DC3E67F.8BA : 161 :
60728>:
> Hello!
>
> I'm having strange trouble with SLES11 SP1 amd MD RAID1 (mdadm - v3.0.3
> (mdadm-3.0.3-0.22.4), 2.6.32.36-0.5-xen):
>
> I was able to create one RAID1 array, but not a second one. I have no idea
> what's wrong, but my guesses are:
>
> 1) An error when using "--bitmap internal" for a 30GB disk
> 2) Unsure whether the disks needs an msdos signature or a RAID partition
> 3) It seems a failed attempt to create the array keeps one device busy (a
> reboot(!) resolves that problem for one attempt)
>
> Some output:
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: device /dev/xvdd not suitable for any style of array
> (Reboot)
> # mdadm -C -l1 -n2 /dev/md0 /dev/xvdd /dev/xvde
> mdadm: /dev/xvdd appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:00:12 2011
> Continue creating array? y
> mdadm: array /dev/md0 started.
>
> # mdadm --grow --bitmap internal /dev/md0
> mdadm: failed to set internal bitmap.
> # mdadm --stop /dev/md0
> mdadm: stopped /dev/md0
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: /dev/xvdd appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:38:20 2011
> mdadm: /dev/xvde appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:38:20 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdd
> mdadm: ADD_NEW_DISK for /dev/xvde failed: Device or resource busy
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: device /dev/xvdd not suitable for any style of array
> # xm reboot rksapv01
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: /dev/xvdd appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 13:26:55 2011
> mdadm: /dev/xvde appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:38:20 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdd
> mdadm: ADD_NEW_DISK for /dev/xvde failed: Device or resource busy
>
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: /dev/xvdd appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 13:37:28 2011
> mdadm: /dev/xvde appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:38:20 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdd
> mdadm: ADD_NEW_DISK for /dev/xvde failed: Device or resource busy
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: device /dev/xvdd not suitable for any style of array
>
> (At this point I filed a service request for SLES with no result until now)
>
> Trying some other disks (of varying size):
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: /dev/xvdd appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 13:37:28 2011
> mdadm: /dev/xvde appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 12:38:20 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdd
> mdadm: ADD_NEW_DISK for /dev/xvde failed: Device or resource busy
> # mdadm -C -l1 -n2 --bitmap internal /dev/md0 /dev/xvdd /dev/xvde
> mdadm: device /dev/xvdd not suitable for any style of array
>
> (another Reboot)
> # mdadm -C -l1 -n2 --bitmap internal /dev/md1 /dev/xvdf /dev/xvdg
> mdadm: /dev/xvdf appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 16:36:59 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdf
> mdadm: ADD_NEW_DISK for /dev/xvdg failed: Device or resource busy
> # mdadm -C -l1 -n2 --bitmap internal /dev/md2 /dev/xvdh /dev/xvdi
> mdadm: /dev/xvdh appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 16:37:26 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdh
> mdadm: ADD_NEW_DISK for /dev/xvdi failed: Device or resource busy
> # mdadm -C -l1 -n2 --bitmap internal /dev/md3 /dev/xvdj /dev/xvdk
> mdadm: /dev/xvdj appears to be part of a raid array:
> level=raid1 devices=2 ctime=Wed May 4 16:37:38 2011
> Continue creating array? y
> mdadm: failed to write superblock to /dev/xvdj
> mdadm: ADD_NEW_DISK for /dev/xvdk failed: Device or resource busy
>
> Corresponding Syslog messages:
> May 4 17:18:54 rksapv01 kernel: [ 231.942241] md: bind
> May 4 17:18:54 rksapv01 kernel: [ 231.942265] md: could not bd_claim xvdg.
> May 4 17:18:54 rksapv01 kernel: [ 231.942269] md: md_import_device
> returned
> -16
> May 4 17:19:13 rksapv01 kernel: [ 250.118561] md: bind
> May 4 17:19:13 rksapv01 kernel: [ 250.118586] md: could not bd_claim xvdi.
> May 4 17:19:13 rksapv01 kernel: [ 250.118590] md: md_import_device
> returned
> -16
> May 4 17:19:27 rksapv01 kernel: [ 264.505337] md: bind
> May 4 17:19:27 rksapv01 kernel: [ 264.505365] md: could not bd_claim xvdk.
> May 4 17:19:27 rksapv01 kernel: [ 264.505368] md: md_import_device
> returned
> -16
>
> Do you understand that I'm quite frustrated? Maybe I should mention that the
> disks are from a FC-SAN with a 4-way multipath
> (multipath-tools-0.4.8-40.25.1). Also "lsof" finds no process that has the
> device open.
>
> Finally: I had a similar problem with SLES10 SP3 which made me quit using
> MD-RAID about two years ago...
>
> Regards,
> Ulrich
> P.S: I'm not subscribed to the list, so please CC: -- Thank you
>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html