[CentOS] odd mdadm behavior

Tue Dec 20 01:30:43 UTC 2011
Ljubomir Ljubojevic <office at plnet.rs>

Vreme: 12/19/2011 10:58 PM, Paul Heinlein piše:
> On Mon, 19 Dec 2011, aurfalien at gmail.com wrote:
>>> I'm interested to know if you used mdadm to fail and remove the bad
>>> disk from the array when it first started acting up.
>> No, I should have but left it alone.
>> I know, my bad.
> I was merely interested.
> Recently I had a RAID-1 device get marked as bad, but I couldn't see
> any SMART errors. So I failed, removed, and then re-added the device.
> It worked for about a week, then it failed again, but time the SMART
> errors were obvious.
> I'd ordered a new drive at the first failure, so I was ready when it
> failed the second time.
> I guess the point is that I've seen "bad" drives go "good" again, at
> least for short periods of time.

Les, I will reply to your mail here.

This would better explain my troubles. I've had to manually re-add them 
few times, but I've had little experience and I thought these things 
happen, never gave it much thought.

I am still not 100% on what actually happened, but my guess would be 
that since boot partition was active only on one drive, and that one was 
creating problems, I went with line of easier resistance and just patch 
it up. I am going to use C6 ability to boot from full raid partition, 
and maybe even add IDE DOM module for boot partition.

I am now good and watching like a hawk.


Ljubomir Ljubojevic
(Love is in the Air)
PL Computers
Serbia, Europe

Google is the Mother, Google is the Father, and traceroute is your
trusty Spiderman...
StarOS, Mikrotik and CentOS/RHEL/Linux consultant