[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Re: raid problem



Le 30/11/2013 06:39, Stan Hoeppner a écrit :
> On 11/29/2013 4:43 PM, François Patte wrote:
>> Bonsoir,
>>
>> I have a problem with 2 raid arrays: I have 2 disks (sdc and sdd) in
>> raid1 arrays.
>>
>> One disk (sdc) failed and I replaced it by a new one. Copying the
>> partition table from sdd disk using sfdisk:
>>
>> sfdisk -d /dev/sdd | sfdisk /dev/sdc
>>
>> then I "added" the 2 partitions (sdc1 and sdc3) to the arrays md0 and md1:
>>
>> mdadm --add /dev/md0 /dev/sdc1
>>
>> mdadm --add /dev/md1 /dev/sdc3
>>
>> There were no problem with the md0 array:
>>
>>
>> cat /proc/mdstat gives:
>>
>> md0 : active raid1 sdc1[1] sdd1[0]
>>       1052160 blocks [2/2] [UU]
>>
>>
>> But for the md1 array, I get:
>>
>> md1 : active raid1 sdc3[2](S) sdd3[0]
>>       483138688 blocks [2/1] [U_]
>>
>>
>> And mdadm --detail /dev/md1 returns:
>>
>> /dev/md1:
>>         Version : 0.90
>>   Creation Time : Sat Mar  7 11:48:30 2009
>>      Raid Level : raid1
>>      Array Size : 483138688 (460.76 GiB 494.73 GB)
>>   Used Dev Size : 483138688 (460.76 GiB 494.73 GB)
>>    Raid Devices : 2
>>   Total Devices : 2
>> Preferred Minor : 1
>>     Persistence : Superblock is persistent
>>
>>     Update Time : Fri Nov 29 21:23:25 2013
>>           State : clean, degraded
>>  Active Devices : 1
>> Working Devices : 2
>>  Failed Devices : 0
>>   Spare Devices : 1
>>
>>            UUID : 2e8294de:9b0d8d96:680a5413:2aac5c13
>>          Events : 0.72076
>>
>>     Number   Major   Minor   RaidDevice State
>>        0       8       51        0      active sync   /dev/sdd3
>>        2       0        0        2      removed
>>
>>        2       8       35        -      spare   /dev/sdc3
>>
>> While mdadm --examine /dev/sdc3 returns:
>>
>> /dev/sdc3:
>>           Magic : a92b4efc
>>         Version : 0.90.00
>>            UUID : 2e8294de:9b0d8d96:680a5413:2aac5c13
>>   Creation Time : Sat Mar  7 11:48:30 2009
>>      Raid Level : raid1
>>   Used Dev Size : 483138688 (460.76 GiB 494.73 GB)
>>
>>
>>      Array Size : 483138688 (460.76 GiB 494.73 GB)
>>    Raid Devices : 2
>>   Total Devices : 2
>> Preferred Minor : 1
>>
>>     Update Time : Fri Nov 29 23:03:41 2013
>>           State : clean
>>  Active Devices : 1
>> Working Devices : 2
>>  Failed Devices : 1
>>   Spare Devices : 1
>>        Checksum : be8bd27f - correct
>>          Events : 72078
>>
>>
>>       Number   Major   Minor   RaidDevice State
>> this     2       8       35        2      spare   /dev/sdc3
>>
>>    0     0       8       51        0      active sync   /dev/sdd3
>>    1     1       0        0        1      faulty removed
>>    2     2       8       35        2      spare   /dev/sdc3
>>
>>
>> What is the problem? And how can I recover a correct md1 array?
> 
> IIRC Linux md rebuilds multiple degraded arrays sequentially, not in
> parallel.  This is due to system performance impact and other reasons.
> When the rebuild of md0 is finished, the rebuild of md1/sdc3 should
> start automatically.  If this did not occur please let us know and we'll
> go from there.

I thought it was clear enough that the result of commands mdadm
--details or mdadm --examine were what they return *after* the rebuild
of array md1.

On reboot, I am warned that md1 is started with one disk out of two and
one spare and recovery starts immediately:

md1 : active raid1 sdd3[0] sdc3[2]
      483138688 blocks [2/1] [U_]
      [=>...................]  recovery =  7.5% (36521408/483138688)
finish=89.1min speed=83445K/sec

After that, the situation is what is quoted in my previous message....

Regards.


-- 
François Patte
UFR de mathématiques et informatique
Laboratoire CNRS MAP5, UMR 8145
Université Paris Descartes
45, rue des Saints Pères
F-75270 Paris Cedex 06
Tél. +33 (0)1 8394 5849
http://www.math-info.univ-paris5.fr/~patte

Attachment: signature.asc
Description: OpenPGP digital signature


Reply to: