RAID5 (mdadm) array hosed after grow operation
I think growing my RAID array after replacing all the
drives with bigger ones has somehow hosed the array.
The system is Etch with a stock 2.6.18 kernel and
mdadm v. 2.5.6, running on an Athlon 1700 box.
The array is 6 disk (5 active, one spare) RAID 5
that has been humming along quite nicely for
a few months now. However, I decided to replace
all the drives with larger ones.
The RAID reassembled fine at each boot as the drives
were replaced one by one. After the last drive was
partitioned and added to the array, I issued the
"mdadm -G /dev/md/0 -z max"
to grow the array to the maximum space available
on the smallest drive. That appeared to work just
fine at the time, but booting today the array
refused to assemble with the following error:
md: hdg1 has invalid sb, not importing!
md: md_import_device returned -22
I tried to force assembly but only two of the remaining
4 active drives appeared to be fault free. dmesg gives
md: kicking non-fresh hde1 from array!
md: kicking non-fresh hdi1 from array!
I also noticed that "mdadm -X <drive>" shows
the pre-grow device size for 2 of the devices
and some discrepancies between event and event cleared
One last thing I found curious---from dmesg:
EXT3-fs error (device hdg1): ext3_check_descriptors: Block
bitmap for group 0 not in group (block 2040936682)!
EXT3-fs: group descriptors corrupted!
There is not ext3 directly on hdg1. LVM sits between the
and the filesystem, so the above message seems suspect.
I hope someone will be able to help me with this. I feel
like the info above is pertinent, but I don't know where to
go from here.
http://www.fastmail.fm - Does exactly what it says on the tin