[CentOS] Re: Help repairing a broken Software Raid5
Scott Silva
ssilva at sgvwater.com
Fri Jul 6 17:59:29 UTC 2007
Andrew @ ATM Logic spake the following on 7/6/2007 4:39 AM:
>> Hi,
>> Can you explain what have you tried till now? All I can say
>> "man mdamd" is sufficient.
>>
>> --
>> Regards,
>> Sudev Barar
>> _______________________________________________
>> CentOS mailing list
>> CentOS at centos.org
>> http://lists.centos.org/mailman/listinfo/centos
>>
>
>
>
> These are a few of the commands I have ran so far...
>
> cat /proc/mdstat
> Personalities : [raid0] [raid1] [raid5] [raid6]
> Md1 : active raid1 hdd1[2] hdc1[1] hda1[0]
> 104320 blocks [3/3] [UUU]
>
>
> (Just so you know... That's missing the "needed" MD2)
>
> lvm pvscan
> - No matching physical volumes found
> lvm lvscan
> - No volume groups found
> lvm vgscan
> - Reading all physical volumes. This may take a while....
> No volume groups found
>
> lvm vgchange -ay
> No volume groups found
>
> lvm vgdisplay
> No volume groups found
>
>
>
> fdisk -l /dev/hda
> Disk /dev/hda: 160.0 GB, 160041885696 bytes
> 255 heads, 63 sectors/track, 19457 cylinders Units = cylinders of 16065 *
> 512 = 8225280 bytes
>
> Device Boot Start End Blocks Id System
> /dev/hda1 * 1 13 104391 fd Linux raid
> autodetect
> /dev/hda2 14 19457 156183930 fd Linux raid
> autodetect
>
> fdisk -l /dev/hdc
> Disk /dev/hdc: 160.0 GB, 160041885696 bytes
> 255 heads, 63 sectors/track, 19457 cylinders Units = cylinders of 16065 *
> 512 = 8225280 bytes
>
> Device Boot Start End Blocks Id System
> /dev/hdc1 * 1 13 104391 fd Linux raid
> autodetect
> /dev/hdc2 14 19457 156183930 fd Linux raid
> autodetect
>
> fdisk -l /dev/hdd
> Disk /dev/hdd: 160.0 GB, 160041885696 bytes
> 255 heads, 63 sectors/track, 19457 cylinders Units = cylinders of 16065 *
> 512 = 8225280 bytes
>
> Device Boot Start End Blocks Id System
> /dev/hdd1 * 1 13 104391 fd Linux raid
> autodetect
> /dev/hdd2 14 19457 156183930 fd Linux raid
> autodetect
>
>
> raidstart /dev/md1
> - No errors, just returns command prompt
>
> raidstart /dev/md2
> - No errors, just returns command prompt
>
>
>
>
> Sooooooooo, Then I tried to start fixing things.....
>
> mdadm --assemble -m 2 /dev/md2 /dev/hda2 /dev/hdc2 /dev/hdd2
>
> I get
> Mdadm: Bad super-minor number: /dev/md2
>
> mdadm --assemble --run -m 2 /dev/md2 /dev/hda2 /dev/hdc2 /dev/hdd2
>
>
> Mdadm: failed to RUN_ARRAY /dev/md2: Invalid argument
>
> Then... Looking at
>
> cat /proc/mdstat
>
> Personalities : [raid0] [raid1] [raid5] [raid6]
> Md1 : active raid1 hdd1[2] hdc1[1] hda1[0]
> 104320 blocks [3/3] [UUU]
>
> Md2 : inactive hdc2[1] hdd2[2]
> 312367616
> Unused devices: <none>
>
>
> Then... Trying to get a little more pushy...
>
> mdadm --stop /dev/md2
> mdadm --verbose --assemble --run -m 2 /dev/md2 /dev/hda2 /dev/hdc2
> /dev/hdd2
>
> Mdadm: looking for devices for /dev/md2
> Mdadm: /dev/hda2 is identified as a member of /dev/md2, slot 0
> Mdadm: /dev/hdc2 is identified as a member of /dev/md2, slot 1
> Mdadm: /dev/hdd2 is identified as a member of /dev/md2, slot 2
> Mdadm: added /dev/hda2 to /dev/md2 as 0
> Mdadm: added /dev/hdd2 to /dev/md2 as 2
> Mdadm: added /dev/hdc2 to /dev/md2 as 1
> Mdadm: failed to RUN_ARRAY /dev/md2: Invalid argument
>
>
> Ok... So... That is where I quit... Any Idea what kind of hope I should
> be holding out for?
I recently had luck with the following on an array that wouldn't start after
the system was unplugged.
mdadm --assemble --run --force --update=summaries /dev/"raid array"
/dev/"drive1" /dev/"drive2" /dev/"drive3"
remember to fix the quoted parts with your actual entries.
--
MailScanner is like deodorant...
You hope everybody uses it, and
you notice quickly if they don't!!!!
More information about the CentOS
mailing list