[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Re: LILO with root-raid nightmare - please help.



Alvin Oga wrote:

what exaactly did you type BEFORE you removed the bad disk ??

raidhotadd, raidhotremove, etc, etc.. is required ( aka good idea )

I use mdadm to manage my array. The command was

mdadm --set-faulty /dev/md1 /dev/hda1 --remove /dev/md1 /dev/hda1

(repeated for the other partitions substituting md2,hda2, md3,hda3 etc; only some of the partitions had been marked as failed in the old drive I tried re-adding them but they failed again after a few days so I swapped the drive)

some bios cannot boot off of additional ide cards ( pci cards )
and mixing with onboard ide might be confusing the system

My system wouldn't boot at all with the new drive in place until I went into the bios and explicitly told it to boot off the drive on the ide card. The drives attached to the card were listed in the bios setup screen as boot options so I'm pretty sure my bios is compatible with booting off raid. Since Windows will only boot off the drive with bios device code 0x80 I believe that the bios may be switching the codes around and this is what is confusing lilo. I don't know why this is a problem though since I thought that since version 22.5 lilo didn't care about device codes.

LILO version 22.6.1

too old ...

you should be using lilo-22.7.1 for raid stuff if you want a working
bootable raid system

It's the version that comes with sarge. I set up the array with the LILO that came with woody, surely upgrading isn't necessary? I'll look for a backport.

lba32
boot=/dev/md1
raid-extra-boot=/dev/hda,/dev/hde

why do you use "extra-boot" ?? - try commenting that line out

AFAIK that line is essential. It tells lilo to write the mbr to both members of the raid array, so that if one drive dies the system remains bootable. If I'm wrong about this please let me know. I set up the array using the instructions at http://alioth.debian.org/projects/rootraiddoc/

this implies /dev/md1 is your /boot ??
and that your rootfs ( /bin, /etc, /sbin .. ) is in /dev/md3 ??

That is correct, however this was my first Linux system and I went a bit overboard on the partitioning: tmp, usr, var, and home each have their own partition.

for more debugging ...
	what is the contents of /etc/fstab
# /etc/fstab: static file system information.
#
# <file system> <mount point> <type> <options> <dump> <pass>
/dev/md3        /               ext3    errors=remount-ro       0       1
/dev/md2        none            swap    sw                      0       0
proc            /proc           proc    defaults                0       0
/dev/fd0        /floppy         auto    user,noauto             0       0
/dev/cdrom      /cdrom          iso9660 ro,user,noauto          0       0
/dev/md5        /tmp    ext3    defaults                        0       2
/dev/md6        /usr    ext3    defaults                        0       2
/dev/md7        /var    ext3    defaults                        0       2
/dev/md8        /home   ext3    defaults                        0       2
/dev/md1        /boot   ext3    defaults                        0       2

	what is the contents of /etc/mdadm.conf
:~# cat /etc/mdadm/mdadm.conf
DEVICE /dev/hda* /dev/hde*
ARRAY /dev/md1 level=raid1 disks=2 UUID=f126c8af:1fd56337:a653c52b:3e50c388
  devices=/dev/hda1,/dev/hde1
ARRAY /dev/md2 level=raid1 disks=2 UUID=6a91045f:babc1717:3a48db49:e10249c1
  devices=/dev/hda2,/dev/hde2
ARRAY /dev/md3 level=raid1 disks=2 UUID=b32dcb66:a802ec78:795cdb4e:3ed450ae
  devices=/dev/hda3,/dev/hde3
ARRAY /dev/md5 level=raid1 disks=2 UUID=173c680c:200676a5:29113ce8:5e0c54e9
  devices=/dev/hda5,/dev/hde5
ARRAY /dev/md6 level=raid1 disks=2 UUID=73a930b9:674865c7:1c5030f1:55ae62a9
  devices=/dev/hda6,/dev/hde6
ARRAY /dev/md7 level=raid1 disks=2 UUID=641c416c:2ffd8890:3cce2a6f:811f694b
  devices=/dev/hda7,/dev/hde7
ARRAY /dev/md8 level=raid1 disks=2 UUID=850a410c:1fa77dd9:5234f468:43814e67
  devices=/dev/hda8,/dev/hde8

	what is the output of "cat /proc/mdstat"
:~# cat /proc/mdstat
Personalities : [raid1] [raid5]
read_ahead 1024 sectors
md1 : active raid1 hde1[1] hda1[0]
     16000 blocks [2/2] [UU]

md2 : active raid1 hde2[1] hda2[0]
     490304 blocks [2/2] [UU]

md3 : active raid1 hde3[1] hda3[0]
     144576 blocks [2/2] [UU]

md5 : active raid1 hde5[1] hda5[0]
     96640 blocks [2/2] [UU]

md6 : active raid1 hde6[1] hda6[0]
     4000128 blocks [2/2] [UU]

md7 : active raid1 hde7[1] hda7[0]
     4000128 blocks [2/2] [UU]

md8 : active raid1 hde8[1] hda8[0]
     31273088 blocks [2/2] [UU]


Thanks for any help/advice you can give.





Reply to: