Unable to access intel fake RAID 1 array in Fedora 14 after reboot

Posted by Sim on Server Fault See other posts from Server Fault or by Sim
Published on 2011-01-06T23:41:38Z Indexed on 2011/01/06 23:55 UTC
Read the original article Hit count: 386

Filed under:
|
|
|
|

Hello everyone, 1st I am relatively new to linux (but not to *nix). I have 4 disks assembled in the following intel ahci bios fake raid arrays: 2x320GB RAID1 - used for operating systems md126 2x1TB RAID1 - used for data md125

I have used the raid of size 320GB to install my operating system and the second raid I didn't even select during the installation of Fedora 14. After successful partitioning and installation of Fedora, I tried to make the second array available, it was possible to make it visible in linux with mdadm --assembe --scan , after that I created one maximum size partition and 1 maximum size ext4 filesystem in it. Mounted, and used it. After restart - a few I/O errors during boot regarding md125 + inability to mount the filesystem on it and dropped into repair shell. I commented the filesystem in fstab and it booted. To my surprise, the array was marked as "auto read only":

[root@localhost ~]# cat /proc/mdstat 
Personalities : [raid1] 
md125 : active (auto-read-only) raid1 sdc[1] sdd[0]
      976759808 blocks super external:/md127/0 [2/2] [UU]

md127 : inactive sdc[1](S) sdd[0](S)
      4514 blocks super external:imsm

md126 : active raid1 sda[1] sdb[0]
      312566784 blocks super external:/md1/0 [2/2] [UU]

md1 : inactive sdb[1](S) sda[0](S)
      4514 blocks super external:imsm

unused devices: <none>
[root@localhost ~]#

and the partition in it was not available as device special file in /dev:

[root@localhost ~]# ls -l /dev/md125*
brw-rw---- 1 root disk 9, 125 Jan  6 15:50 /dev/md125
[root@localhost ~]#

But the partition is there according to fdisk:

[root@localhost ~]# fdisk -l /dev/md125

Disk /dev/md125: 1000.2 GB, 1000202043392 bytes
19 heads, 10 sectors/track, 10281682 cylinders, total 1953519616 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x1b238ea9

      Device Boot      Start         End      Blocks   Id  System
/dev/md125p1            2048  1953519615   976758784   83  Linux
[root@localhost ~]# 

I tried to "activate" the array in different ways (I'm not experienced with mdadm and the man page is gigantic so I was only browsing it looking for my answer) but it was impossible - the array would still stay in "auto read only" and the device special file for the partition it will not be in /dev. It was only after I recreated the partition via fdisk that it reappeared in /dev... until next reboot.

So, my question is - How do I make the array automatically available after reboot?

Here is some additional information:

1st I am able to see the UUID of the array in blkid:

[root@localhost ~]# blkid 
/dev/sdc: UUID="b9a1149f-ae11-4fc8-a600-0d77354dc42a" SEC_TYPE="ext2" TYPE="ext3" 
/dev/sdd: UUID="b9a1149f-ae11-4fc8-a600-0d77354dc42a" SEC_TYPE="ext2" TYPE="ext3" 
/dev/md126p1: UUID="60C8D9A7C8D97C2A" TYPE="ntfs" 
/dev/md126p2: UUID="3d1b38a3-b469-4b7c-b016-8abfb26a5d7d" TYPE="ext4" 
/dev/md126p3: UUID="1Msqqr-AAF8-k0wi-VYnq-uWJU-y0OD-uIFBHL" TYPE="LVM2_member" 
/dev/mapper/vg00-rootlv: LABEL="_Fedora-14-x86_6" UUID="34cc1cf5-6845-4489-8303-7a90c7663f0a" TYPE="ext4" 
/dev/mapper/vg00-swaplv: UUID="4644d857-e13b-456c-ac03-6f26299c1046" TYPE="swap" 
/dev/mapper/vg00-homelv: UUID="82bd58b2-edab-4b4b-aec4-b79595ecd0e3" TYPE="ext4" 
/dev/mapper/vg00-varlv: UUID="1b001444-5fdd-41b6-a59a-9712ec6def33" TYPE="ext4" 
/dev/mapper/vg00-tmplv: UUID="bf7d2459-2b35-4a1c-9b81-d4c4f24a9842" TYPE="ext4" 
/dev/md125: UUID="b9a1149f-ae11-4fc8-a600-0d77354dc42a" SEC_TYPE="ext2" TYPE="ext3"
/dev/sda: TYPE="isw_raid_member" 
/dev/md125p1: UUID="420adfdd-6c4e-4552-93f0-2608938a4059" TYPE="ext4" 
[root@localhost ~]# 

Here is how /etc/mdadm.conf looks like:

[root@localhost ~]# cat /etc/mdadm.conf
# mdadm.conf written out by anaconda
MAILADDR root
AUTO +imsm +1.x -all
ARRAY /dev/md1 UUID=89f60dee:e46a251f:7475814b:d4cc19a9
ARRAY /dev/md126 UUID=a8775c90:cee66376:5310fc13:63bcba5b
ARRAY /dev/md125 UUID=b9a1149f:ae114fc8:a6000d77:354dc42a
[root@localhost ~]# 

here is how /proc/mdstat looks like after I recreate the partition in the array so that it becomes available:

[root@localhost ~]# cat /proc/mdstat 
Personalities : [raid1] 
md125 : active raid1 sdc[1] sdd[0]
      976759808 blocks super external:/md127/0 [2/2] [UU]

md127 : inactive sdc[1](S) sdd[0](S)
      4514 blocks super external:imsm

md126 : active raid1 sda[1] sdb[0]
      312566784 blocks super external:/md1/0 [2/2] [UU]

md1 : inactive sdb[1](S) sda[0](S)
      4514 blocks super external:imsm

unused devices: <none>
[root@localhost ~]# 


Detailed output regarding the array in subject:

[root@localhost ~]# mdadm --detail /dev/md125
/dev/md125:
      Container : /dev/md127, member 0
     Raid Level : raid1
     Array Size : 976759808 (931.51 GiB 1000.20 GB)
  Used Dev Size : 976759940 (931.51 GiB 1000.20 GB)
   Raid Devices : 2
  Total Devices : 2

    Update Time : Fri Jan  7 00:38:00 2011
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0


           UUID : 30ebc3c2:b6a64751:4758d05c:fa8ff782
    Number   Major   Minor   RaidDevice State
       1       8       32        0      active sync   /dev/sdc
       0       8       48        1      active sync   /dev/sdd
[root@localhost ~]# 

and /etc/fstab, with /data commented (the filesystem that is on this array):

#
# /etc/fstab
# Created by anaconda on Thu Jan  6 03:32:40 2011
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
/dev/mapper/vg00-rootlv /                       ext4    defaults        1 1
UUID=3d1b38a3-b469-4b7c-b016-8abfb26a5d7d /boot                   ext4    defaults        1 2
#UUID=420adfdd-6c4e-4552-93f0-2608938a4059 /data ext4 defaults 0 1
/dev/mapper/vg00-homelv /home                   ext4    defaults        1 2
/dev/mapper/vg00-tmplv  /tmp                    ext4    defaults        1 2
/dev/mapper/vg00-varlv  /var                    ext4    defaults        1 2
/dev/mapper/vg00-swaplv swap                    swap    defaults        0 0
tmpfs                   /dev/shm                tmpfs   defaults        0 0
devpts                  /dev/pts                devpts  gid=5,mode=620  0 0
sysfs                   /sys                    sysfs   defaults        0 0
proc                    /proc                   proc    defaults        0 0
[root@localhost ~]# 

Thanks in advance to everyone that even read this whole issue :-)

© Server Fault or respective owner

Related posts about linux

Related posts about hardware-raid