Vreme: 12/19/2011 10:58 PM, Paul Heinlein piše:
On Mon, 19 Dec 2011, aurfalien@gmail.com wrote:
I'm interested to know if you used mdadm to fail and remove the bad disk from the array when it first started acting up.
No, I should have but left it alone.
I know, my bad.
I was merely interested.
Recently I had a RAID-1 device get marked as bad, but I couldn't see any SMART errors. So I failed, removed, and then re-added the device. It worked for about a week, then it failed again, but time the SMART errors were obvious.
I'd ordered a new drive at the first failure, so I was ready when it failed the second time.
I guess the point is that I've seen "bad" drives go "good" again, at least for short periods of time.
Les, I will reply to your mail here.
This would better explain my troubles. I've had to manually re-add them few times, but I've had little experience and I thought these things happen, never gave it much thought.
I am still not 100% on what actually happened, but my guess would be that since boot partition was active only on one drive, and that one was creating problems, I went with line of easier resistance and just patch it up. I am going to use C6 ability to boot from full raid partition, and maybe even add IDE DOM module for boot partition.
I am now good and watching like a hawk.