Koozali.org formerly Contribs.org

SME Software Raid problem

SME Software Raid problem
« on: January 04, 2008, 08:57:15 PM »
Im a bit in distress on what is wrong here.

From raid monitor I get alarm:

ALARM! RAID configuration problem

Current configuration is:

Personalities : [raid1]
read_ahead 1024 sectors
md2 : active raid1 hda3[0] hdc3[1]
      262016 blocks [2/2] [UU]
     
md1 : active raid1 hdc2[1]
      119695872 blocks [2/1] [_U]
     
md0 : active raid1 hda1[0] hdc1[1]
      102208 blocks [2/2] [UU]
     
unused devices: <none>

Last known good configuration was:

Personalities : [raid1]
read_ahead 1024 sectors
md2 : active raid1 hda3[0] hdc3[1]
      262016 blocks [2/2] [UU]
     
md1 : active raid1 hda2[0] hdc2[1]
      119695872 blocks [2/2] [UU]
     
md0 : active raid1 hda1[0] hdc1[1]
      102208 blocks [2/2] [UU]
     
unused devices: <none>


It seems to be a problem with hda,  but only on md1? and this confuses me: [2/1] [_U]

Is hda broken and should I just replace it or is it just md1 thats broken?  :shock:

from dmesg log:

md: autorun ...
md: considering hda1 ...
md:  adding hda1 ...
md:  adding hdc1 ...
md: created md0
md: bind<hdc1,1>
md: bind<hda1,2>
md: running: <hda1><hdc1>
md: hda1's event counter: 0000002b
md: hdc1's event counter: 0000002b
md: RAID level 1 does not need chunksize! Continuing anyway.
md0: max total readahead window set to 124k
md0: 1 data-disks, max readahead per data-disk: 124k
raid1: device hda1 operational as mirror 0
raid1: device hdc1 operational as mirror 1
raid1: raid set md0 active with 2 out of 2 mirrors
md: updating md0 RAID superblock on device
md: hda1 [events: 0000002c]<6>(write) hda1's sb offset: 102208
md: hdc1 [events: 0000002c]<6>(write) hdc1's sb offset: 102208
md: considering hdc2 ...
md:  adding hdc2 ...
md:  adding hda2 ...
md: created md1
md: bind<hda2,1>
md: bind<hdc2,2>
md: running: <hdc2><hda2>
md: hdc2's event counter: 0000002c
md: hda2's event counter: 0000002a
md: superblock update time inconsistency -- using the most recent one
md: freshest: hdc2
md: kicking non-fresh hda2 from array!
md: unbind<hda2,1>
md: export_rdev(hda2)
md1: removing former faulty hda2!
md: md1: raid array is not clean -- starting background reconstruction
md: RAID level 1 does not need chunksize! Continuing anyway.
md1: max total readahead window set to 124k
md1: 1 data-disks, max readahead per data-disk: 124k
raid1: device hdc2 operational as mirror 1
raid1: md1, not all disks are operational -- trying to recover array
raid1: raid set md1 active with 1 out of 2 mirrors
md: updating md1 RAID superblock on device
md: hdc2 [events: 0000002d]<6>(write) hdc2's sb offset: 119695872
md: recovery thread got woken up ...
md1: no spare disk to reconstruct array! -- continuing in degraded mode
md: recovery thread finished ...
md: considering hda3 ...
md:  adding hda3 ...
md:  adding hdc3 ...
md: created md2
md: bind<hdc3,1>
md: bind<hda3,2>
md: running: <hda3><hdc3>
md: hda3's event counter: 0000002b
md: hdc3's event counter: 0000002b
md: RAID level 1 does not need chunksize! Continuing anyway.
md2: max total readahead window set to 124k
md2: 1 data-disks, max readahead per data-disk: 124k
raid1: device hda3 operational as mirror 0
raid1: device hdc3 operational as mirror 1
raid1: raid set md2 active with 2 out of 2 mirrors
md: updating md2 RAID superblock on device
md: hda3 [events: 0000002c]<6>(write) hda3's sb offset: 262016
md: hdc3 [events: 0000002c]<6>(write) hdc3's sb offset: 262016
md: ... autorun DONE.


Re: SME Software Raid problem
« Reply #1 on: January 05, 2008, 03:28:15 PM »
I would try rebuilding the array again and see if the problem persists or goes away. Things happen. I had one go over a year ago and after rebuild I had no issues; perhaps a temporary power issue caused a bad rewrite. A few years ago a different array failed and it was defintely an HDD issue.

check out the following in the wiki:
http://wiki.contribs.org/Raid

It will also show your condition.

Christian
SME since 2003

Re: SME Software Raid problem
« Reply #2 on: January 05, 2008, 11:07:24 PM »
I would try rebuilding the array again and see if the problem persists or goes away. Things happen. I had one go over a year ago and after rebuild I had no issues; perhaps a temporary power issue caused a bad rewrite. A few years ago a different array failed and it was defintely an HDD issue.

check out the following in the wiki:
http://wiki.contribs.org/Raid

It will also show your condition.

Christian


Thanks! I tried adding the partition again from hda and it is rebuiding again with success!  :-) I believe that just saved me a 120km drive and a customer a new hd.  :eek: