Im a bit in distress on what is wrong here.
From raid monitor I get alarm:
ALARM! RAID configuration problem
Current configuration is:
Personalities : [raid1]
read_ahead 1024 sectors
md2 : active raid1 hda3[0] hdc3[1]
262016 blocks [2/2] [UU]
md1 : active raid1 hdc2[1]
119695872 blocks [2/1] [_U]
md0 : active raid1 hda1[0] hdc1[1]
102208 blocks [2/2] [UU]
unused devices: <none>
Last known good configuration was:
Personalities : [raid1]
read_ahead 1024 sectors
md2 : active raid1 hda3[0] hdc3[1]
262016 blocks [2/2] [UU]
md1 : active raid1 hda2[0] hdc2[1]
119695872 blocks [2/2] [UU]
md0 : active raid1 hda1[0] hdc1[1]
102208 blocks [2/2] [UU]
unused devices: <none>
It seems to be a problem with hda, but only on md1? and this confuses me: [2/1] [_U]
Is hda broken and should I just replace it or is it just md1 thats broken?
from dmesg log:
md: autorun ...
md: considering hda1 ...
md: adding hda1 ...
md: adding hdc1 ...
md: created md0
md: bind<hdc1,1>
md: bind<hda1,2>
md: running: <hda1><hdc1>
md: hda1's event counter: 0000002b
md: hdc1's event counter: 0000002b
md: RAID level 1 does not need chunksize! Continuing anyway.
md0: max total readahead window set to 124k
md0: 1 data-disks, max readahead per data-disk: 124k
raid1: device hda1 operational as mirror 0
raid1: device hdc1 operational as mirror 1
raid1: raid set md0 active with 2 out of 2 mirrors
md: updating md0 RAID superblock on device
md: hda1 [events: 0000002c]<6>(write) hda1's sb offset: 102208
md: hdc1 [events: 0000002c]<6>(write) hdc1's sb offset: 102208
md: considering hdc2 ...
md: adding hdc2 ...
md: adding hda2 ...
md: created md1
md: bind<hda2,1>
md: bind<hdc2,2>
md: running: <hdc2><hda2>
md: hdc2's event counter: 0000002c
md: hda2's event counter: 0000002a
md: superblock update time inconsistency -- using the most recent one
md: freshest: hdc2
md: kicking non-fresh hda2 from array!
md: unbind<hda2,1>
md: export_rdev(hda2)
md1: removing former faulty hda2!
md: md1: raid array is not clean -- starting background reconstruction
md: RAID level 1 does not need chunksize! Continuing anyway.
md1: max total readahead window set to 124k
md1: 1 data-disks, max readahead per data-disk: 124k
raid1: device hdc2 operational as mirror 1
raid1: md1, not all disks are operational -- trying to recover array
raid1: raid set md1 active with 1 out of 2 mirrors
md: updating md1 RAID superblock on device
md: hdc2 [events: 0000002d]<6>(write) hdc2's sb offset: 119695872
md: recovery thread got woken up ...
md1: no spare disk to reconstruct array! -- continuing in degraded mode
md: recovery thread finished ...
md: considering hda3 ...
md: adding hda3 ...
md: adding hdc3 ...
md: created md2
md: bind<hdc3,1>
md: bind<hda3,2>
md: running: <hda3><hdc3>
md: hda3's event counter: 0000002b
md: hdc3's event counter: 0000002b
md: RAID level 1 does not need chunksize! Continuing anyway.
md2: max total readahead window set to 124k
md2: 1 data-disks, max readahead per data-disk: 124k
raid1: device hda3 operational as mirror 0
raid1: device hdc3 operational as mirror 1
raid1: raid set md2 active with 2 out of 2 mirrors
md: updating md2 RAID superblock on device
md: hda3 [events: 0000002c]<6>(write) hda3's sb offset: 262016
md: hdc3 [events: 0000002c]<6>(write) hdc3's sb offset: 262016
md: ... autorun DONE.