Results 1 to 39 of 39

Thread: RAID Superblock disappeared from one of my disks after crash. Fix?

  1. #1
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733

    RAID Superblock disappeared from one of my disks after crash. Fix?

    My fileserver crashed a couple times lately, and I just noticed that the RAID array didn't come back up the last time. I stored the configuration in /etc/mdadm.conf:
    Code:
    # cat /etc/mdadm.conf
    DEVICE  /dev/sda1 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/hda1
    ARRAY /dev/md0 level=raid5 num-devices=5 UUID=20694a10:5013084e:6dde8b75:f422ec4e
    So I figured I could simply run:
    Code:
    # mdadm --assemble --scan -v
    But only two of the five disks were added to the array. Not enough to start it. So I tried:
    Code:
     # mdadm --assemble --scan -fv
    mdadm: looking for devices for /dev/md0
    mdadm: /dev/sda1 is identified as a member of /dev/md0, slot 0.
    mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 1.
    mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 2.
    mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 3.
    mdadm: /dev/hda1 is identified as a member of /dev/md0, slot 4.
    mdadm: forcing event count in /dev/sda1(0) from 312084 upto 312087
    mdadm: forcing event count in /dev/sdb1(1) from 312084 upto 312087
    mdadm: RAID superblock has disappeared from /dev/sda1
    # cat /proc/mdstat
    Personalities : [raid5] [raid4]
    unused devices: <none>
    # mount /dev/md0 /exports
    mount: /dev/md0: can't read superblock
    # mdadm --assemble --scan -v
    mdadm: looking for devices for /dev/md0
    mdadm: no recogniseable superblock on /dev/sda1
    mdadm: /dev/sda1 has wrong uuid.
    mdadm: no recogniseable superblock on /dev/sdb1
    mdadm: /dev/sdb1 has wrong uuid.
    mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 2.
    mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 3.
    mdadm: /dev/hda1 is identified as a member of /dev/md0, slot 4.
    mdadm: no uptodate device for slot 0 of /dev/md0
    mdadm: no uptodate device for slot 1 of /dev/md0
    mdadm: added /dev/sdd1 to /dev/md0 as 3
    mdadm: added /dev/hda1 to /dev/md0 as 4
    mdadm: added /dev/sdc1 to /dev/md0 as 2
    mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.
    Now, I suppose I have to restore the superblocks on sda1 & sdb1. Does anyone have any advice? I'm thinking I have to recreate the array identically to restore the superblocks. I'm 90% sure I can do that from memory, but does anyone know if there is any way to confirm the exact configuration of the array just in case my memory is wrong? Since most of the superblocks are present, I think this should be possible.

  2. #2
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    Another question. If I recreate the array should I use the "--assume-clean" option? That's what I'm leaning towards right now.

    If I don't will the data be wiped? I'm worried the sync will wipe everything. OTOH, if I do, do I risk corruption because of the crashes? The filesystem was a journalling one (XFS).

  3. #3
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    I found:
    Code:
    mdadm --misc --examine
    I'm going to fiddle with that and see if I can learn anything.

  4. #4
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    Oh, and I think
    Code:
    mdadm --assemble --update=uuid
    Might be handy. It should update the uuid on the superblocks of the RAID disks.

    Also:
    Code:
    mdadm --assemble --update=resync
    Is something I'm considering trying after that and before I use the array --if the first command can fix the superblocks.

  5. #5
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    Here's the contents of the superblocks on the good disks:
    Code:
    # mdadm --misc --examine /dev/sdc1
    /dev/sdc1:
              Magic : a92b4efc
            Version : 00.90.03
               UUID : 20694a10:5013084e:6dde8b75:f422ec4e
      Creation Time : Fri Jul  7 14:56:21 2006
         Raid Level : raid5
        Device Size : 245111552 (233.76 GiB 250.99 GB)
         Array Size : 980446208 (935.03 GiB 1003.98 GB)
       Raid Devices : 5
      Total Devices : 5
    Preferred Minor : 0
    
        Update Time : Wed Sep  6 18:03:51 2006
              State : clean
     Active Devices : 2
    Working Devices : 2
     Failed Devices : 5
      Spare Devices : 0
           Checksum : 13fe93de - correct
             Events : 0.312087
    
             Layout : left-symmetric
         Chunk Size : 128K
    
          Number   Major   Minor   RaidDevice State
    this     2       8       33        2      active sync   /dev/sdc1
    
       0     0       0        0        0      removed
       1     1       0        0        1      faulty removed
       2     2       8       33        2      active sync   /dev/sdc1
       3     3       0        0        3      faulty removed
       4     4       3        1        4      active sync   /dev/hda1
    It appears that some disks have been noted as removed/faulty because of their bad superblocks when I tried to reassemble the array.

  6. #6
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    Oops. The other two disks:
    Code:
    # mdadm --misc --examine /dev/sdd1
    /dev/sdd1:
              Magic : a92b4efc
            Version : 00.90.03
               UUID : 20694a10:5013084e:6dde8b75:f422ec4e
      Creation Time : Fri Jul  7 14:56:21 2006
         Raid Level : raid5
        Device Size : 245111552 (233.76 GiB 250.99 GB)
         Array Size : 980446208 (935.03 GiB 1003.98 GB)
       Raid Devices : 5
      Total Devices : 5
    Preferred Minor : 0
    
        Update Time : Wed Sep  6 10:03:55 2006
              State : clean
     Active Devices : 5
    Working Devices : 5
     Failed Devices : 0
      Spare Devices : 0
           Checksum : 13fe2340 - correct
             Events : 0.312084
    
             Layout : left-symmetric
         Chunk Size : 128K
    
          Number   Major   Minor   RaidDevice State
    this     3       8       49        3      active sync   /dev/sdd1
    
       0     0       8        1        0      active sync   /dev/sda1
       1     1       8       17        1      active sync   /dev/sdb1
       2     2       8       33        2      active sync   /dev/sdc1
       3     3       8       49        3      active sync   /dev/sdd1
       4     4       3        1        4      active sync   /dev/hda1
    # mdadm --misc --examine /dev/hda1
    /dev/hda1:
              Magic : a92b4efc
            Version : 00.90.03
               UUID : 20694a10:5013084e:6dde8b75:f422ec4e
      Creation Time : Fri Jul  7 14:56:21 2006
         Raid Level : raid5
        Device Size : 245111552 (233.76 GiB 250.99 GB)
         Array Size : 980446208 (935.03 GiB 1003.98 GB)
       Raid Devices : 5
      Total Devices : 5
    Preferred Minor : 0
    
        Update Time : Wed Sep  6 18:03:51 2006
              State : clean
     Active Devices : 2
    Working Devices : 2
     Failed Devices : 5
      Spare Devices : 0
           Checksum : 13fe93bd - correct
             Events : 0.312087
    
             Layout : left-symmetric
         Chunk Size : 128K
    
          Number   Major   Minor   RaidDevice State
    this     4       3        1        4      active sync   /dev/hda1
    
       0     0       0        0        0      removed
       1     1       0        0        1      faulty removed
       2     2       8       33        2      active sync   /dev/sdc1
       3     3       0        0        3      faulty removed
       4     4       3        1        4      active sync   /dev/hda1

  7. #7
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    Attempting to re-add the devices doesn't work:
    Code:
    # mdadm --manage /dev/md0 --re-add /dev/sda1
    mdadm: cannot get array info for /dev/md0
    # mdadm --manage /dev/md0 --re-add /dev/sdb1
    mdadm: cannot get array info for /dev/md0
    Annoying. Sorry for the spam-style thread. It helps me think through this. I suppose it may be useful to someone else with a similar problem one day.

  8. #8
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    One thing I don't understand is why mdadm doesn't just resync the superblocks when I specify "--force" with "--scan". It has the configuration information in mdadm.conf. That's why I put it there so if things went wrong I would be able to restore the array easily.

    So much for that >. Bah!

  9. #9
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    SUCCESS!!

  10. #10
    Storage Is My Life LiamC's Avatar
    Join Date
    Feb 2002
    Location
    Canberra
    Posts
    2,007
    How?
    It's good to be king

  11. #11
    Storage is cool
    Join Date
    Aug 2004
    Location
    Ottawa, ON
    Posts
    733
    1. I examined the good superblocks to make sure I knew how to recreate the array exactly (i.e. chunk size, etc).
    2. I zeroed all the superblocks on all the disks.
    3. I recreated the array with the "--assume-clean" option.
    4. I marked it possibly dirty with:
    Code:
    mdadm --assemble /dev/md0 --update=resync
    5. I let it try to resync. This only took about 30-60 minutes, probably because all the data was good and it was just reading from each disk and not writing.
    6. I mounted it & everything so far appears to be good.

    Alternately, I think I could have simply recreated it without "--assume-clean" and the initial creation sync wouldn't have destroyed the data, but that wasn't a risk I was willing to take.

    Once I got over my fear of accidentally destroying the data, it didn't really take me long.

  12. #12
    Fatwah on Western Digital Fixture Mercutio's Avatar
    Join Date
    Jan 2002
    Location
    I am omnipresent
    Posts
    18,635
    This is very useful information Gilbo. Thank you for sharing it. Messing around with data recovery on a RAID array is pretty much a cold-sweat nightmare for any techie.

  13. #13
    What is this storage?
    Join Date
    Jul 2010
    Posts
    3

    Hey Gilbo - Please help

    Your scenario seems most similar and promising to the one I am facing and I was hoping you could help me out. I started a new thread at: http://www.storageforum.net/forum/sh...653#post134653 and would very much appreciate your input.

  14. #14
    What is this storage? cyberwiz's Avatar
    Join Date
    Nov 2010
    Posts
    1
    This thread still saves lives

  15. #15
    What is this storage?
    Join Date
    Nov 2010
    Posts
    2
    Hi,
    Guys can you help me, my raid failed trying to recover and details as follows:
    #mdadm --assemble -v /dev/sd[b-i]1 --run
    mdadm: looking for devices for /dev/sdb1
    mdadm: /dev/sdc1 is identified as a member of /dev/sdb1, slot 1.
    mdadm: /dev/sdd1 is identified as a member of /dev/sdb1, slot 7.
    mdadm: /dev/sde1 is identified as a member of /dev/sdb1, slot 6.
    mdadm: /dev/sdf1 is identified as a member of /dev/sdb1, slot 2.
    mdadm: /dev/sdg1 is identified as a member of /dev/sdb1, slot 3.
    mdadm: /dev/sdh1 is identified as a member of /dev/sdb1, slot 4.
    mdadm: /dev/sdi1 is identified as a member of /dev/sdb1, slot 5.
    mdadm: no uptodate device for slot 0 of /dev/sdb1
    mdadm: added /dev/sdf1 to /dev/sdb1 as 2
    mdadm: added /dev/sdg1 to /dev/sdb1 as 3
    mdadm: added /dev/sdh1 to /dev/sdb1 as 4
    mdadm: added /dev/sdi1 to /dev/sdb1 as 5
    mdadm: added /dev/sde1 to /dev/sdb1 as 6
    mdadm: added /dev/sdd1 to /dev/sdb1 as 7
    mdadm: added /dev/sdc1 to /dev/sdb1 as 1
    mdadm: failed to RUN_ARRAY /dev/sdb1: Input/output error
    mdadm: Not enough devices to start the array.

    http://www.storageforum.net/forum/sh...d.php?p=139021

    # mdadm --misc --examine
    mdadm: No devices to examine

    # mdadm --assemble --update=uuid /dev/md0
    mdadm: failed to add 8:17 to /dev/md0: Device or resource busy
    mdadm: /dev/md0 assembled from 0 drives - not enough to start the array.

    # cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
    [raid4] [raid10]
    md127 : inactive sdc1[1] sdd1[7] sde1[6]
    2930279808 blocks

    md0 : inactive sdi1[5](S) sdh1[4](S) sdg1[3](S) sdf1[2](S)
    3907039744 blocks

    unused devices: <none>


    ------------
    dcerouter_1000085:~# cat /etc/mdadm.conf
    cat: /etc/mdadm.conf: No such file or directory
    dcerouter_1000085:~# mdadm --assemble --scan -v
    mdadm: looking for devices for /dev/md0
    mdadm: no RAID superblock on /dev/sdi
    mdadm: /dev/sdi has wrong uuid.
    mdadm: no RAID superblock on /dev/sdh
    mdadm: /dev/sdh has wrong uuid.
    mdadm: no RAID superblock on /dev/sdg
    mdadm: /dev/sdg has wrong uuid.
    mdadm: no RAID superblock on /dev/sdf
    mdadm: /dev/sdf has wrong uuid.
    mdadm: cannot open device /dev/sde1: Device or resource busy
    mdadm: /dev/sde1 has wrong uuid.
    mdadm: cannot open device /dev/sde: Device or resource busy
    mdadm: /dev/sde has wrong uuid.
    mdadm: cannot open device /dev/sdd1: Device or resource busy
    mdadm: /dev/sdd1 has wrong uuid.
    mdadm: cannot open device /dev/sdd: Device or resource busy
    mdadm: /dev/sdd has wrong uuid.
    mdadm: cannot open device /dev/sdc1: Device or resource busy
    mdadm: /dev/sdc1 has wrong uuid.
    mdadm: cannot open device /dev/sdc: Device or resource busy
    mdadm: /dev/sdc has wrong uuid.
    mdadm: no RAID superblock on /dev/sdb
    mdadm: /dev/sdb has wrong uuid.
    mdadm: cannot open device /dev/sda5: Device or resource busy
    mdadm: /dev/sda5 has wrong uuid.
    mdadm: no RAID superblock on /dev/sda2
    mdadm: /dev/sda2 has wrong uuid.
    mdadm: cannot open device /dev/sda1: Device or resource busy
    mdadm: /dev/sda1 has wrong uuid.
    mdadm: cannot open device /dev/sda: Device or resource busy
    mdadm: /dev/sda has wrong uuid.
    mdadm: /dev/sdi1 is identified as a member of /dev/md0, slot 5.
    mdadm: /dev/sdh1 is identified as a member of /dev/md0, slot 4.
    mdadm: /dev/sdg1 is identified as a member of /dev/md0, slot 3.
    mdadm: /dev/sdf1 is identified as a member of /dev/md0, slot 2.
    mdadm: 8:17 is identified as a member of /dev/md0, slot 0.
    mdadm: no uptodate device for slot 1 of /dev/md0
    mdadm: added /dev/sdf1 to /dev/md0 as 2
    mdadm: added /dev/sdg1 to /dev/md0 as 3
    mdadm: added /dev/sdh1 to /dev/md0 as 4
    mdadm: added /dev/sdi1 to /dev/md0 as 5
    mdadm: no uptodate device for slot 6 of /dev/md0
    mdadm: no uptodate device for slot 7 of /dev/md0
    mdadm: failed to add 8:17 to /dev/md0: Device or resource busy
    mdadm: /dev/md0 assembled from 0 drives - not enough to start the array.
    dcerouter_1000085:~# mdadm --assemble --scan -fv
    mdadm: looking for devices for /dev/md0
    mdadm: no RAID superblock on /dev/sdi
    mdadm: /dev/sdi has wrong uuid.
    mdadm: no RAID superblock on /dev/sdh
    mdadm: /dev/sdh has wrong uuid.
    mdadm: no RAID superblock on /dev/sdg
    mdadm: /dev/sdg has wrong uuid.
    mdadm: no RAID superblock on /dev/sdf
    mdadm: /dev/sdf has wrong uuid.
    mdadm: cannot open device /dev/sde1: Device or resource busy
    mdadm: /dev/sde1 has wrong uuid.
    mdadm: cannot open device /dev/sde: Device or resource busy
    mdadm: /dev/sde has wrong uuid.
    mdadm: cannot open device /dev/sdd1: Device or resource busy
    mdadm: /dev/sdd1 has wrong uuid.
    mdadm: cannot open device /dev/sdd: Device or resource busy
    mdadm: /dev/sdd has wrong uuid.
    mdadm: cannot open device /dev/sdc1: Device or resource busy
    mdadm: /dev/sdc1 has wrong uuid.
    mdadm: cannot open device /dev/sdc: Device or resource busy
    mdadm: /dev/sdc has wrong uuid.
    mdadm: no RAID superblock on /dev/sdb
    mdadm: /dev/sdb has wrong uuid.
    mdadm: cannot open device /dev/sda5: Device or resource busy
    mdadm: /dev/sda5 has wrong uuid.
    mdadm: no RAID superblock on /dev/sda2
    mdadm: /dev/sda2 has wrong uuid.
    mdadm: cannot open device /dev/sda1: Device or resource busy
    mdadm: /dev/sda1 has wrong uuid.
    mdadm: cannot open device /dev/sda: Device or resource busy
    mdadm: /dev/sda has wrong uuid.
    mdadm: /dev/sdi1 is identified as a member of /dev/md0, slot 5.
    mdadm: /dev/sdh1 is identified as a member of /dev/md0, slot 4.
    mdadm: /dev/sdg1 is identified as a member of /dev/md0, slot 3.
    mdadm: /dev/sdf1 is identified as a member of /dev/md0, slot 2.
    mdadm: 8:17 is identified as a member of /dev/md0, slot 0.
    mdadm: forcing event count in /dev/sdf1(2) from 696186 upto 696190
    Segmentation fault (core dumped)
    dcerouter_1000085:~# cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
    md127 : inactive sdc1[1] sdd1[7] sde1[6]
    2930279808 blocks

    unused devices: <none>
    dcerouter_1000085:~# history |grep stop
    456 mdadm --stop /dev/md0
    458 mdadm --stop /dev/md0
    466 mdadm --stop /dev/md0
    489 mdadm --stop /dev/md0
    491 mdadm --stop /dev/md0
    495 mdadm --stop /dev/md0
    501 mdadm --stop /dev/md0
    515 history |grep stop
    516 mdadm --stop /dev/md0
    518 mdadm --stop /dev/md0
    520 mdadm --stop /dev/md0
    521 mdadm --stop /dev/md0
    522 mdadm --stop /dev/md0
    531 mdadm --stop /dev/md0
    539 history |grep stop
    dcerouter_1000085:~#
    dcerouter_1000085:~# mdadm --stop /dev/md127
    mdadm: stopped /dev/md127
    # cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
    unused devices: <none>
    dcerouter_1000085:~#



    #mdadm --misc --examine /dev/md0
    mdadm: No md superblock detected on /dev/md0.


    Thanks,
    Srini

  16. #16
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    Thanks for the helpful post. I believe I am in the same issue with my new XFS mdadm raid6.
    The /dev devices have changed in my system and It looks like the superblocks are incorrect on the disks. My hunch is that because I had my old array and new array hooked up in the same box when I built this array, and now the old array is removed this might have caused an issue. It also crashed while resyncing (sata cables got yanked out).

    note: I have not let it successfully resync so this may be an issue.

    [root@machine ~]# mdadm --assemble --scan -fv
    mdadm: looking for devices for /dev/md1
    mdadm: /dev/sdf has wrong uuid.
    mdadm: /dev/sde has wrong uuid.
    mdadm: /dev/sdd has wrong uuid.
    mdadm: /dev/sdb has wrong uuid.
    mdadm: /dev/sdc has wrong uuid.

    I then started the array by using the update=uuid option.

    [root@poopsock ~]# mdadm --assemble --update=uuid /dev/md1
    mdadm: /dev/md1 has been started with 5 drives.

    It is currently rebuilding and estimated to finish in 16hrs.

    I had a few questions

    1) Should I let resync finish before mounting/using the array? I've been copying data to it but it will randomly fail all the drives after some read/write activity.

    2) I read that you used mdadm --zero-superblock $devices - will this destroy data? I see that listed as one of your steps.

    I will wait for it to resync before remounting it and see.

    Thanks.

  17. #17
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    Ok, as an update, I think I'm having multiple issues.

    The resync finished but something else made all the drives fail out of the array. Also, I seem to have a mystery "md127" device appear occasionally and usurp my /dev/sdc.

    I also feel like NCQ being enabled might be an issue as I'm using low-power seagate and samsung drives.

    Finally, I think it could be possible that the drives aren't responding to default timeout set when under load, explaining the stack traces (dealing with timeouts) and subsequent tossing of drives out of the array.

    Here is what I've done so far.

    (disable NCQ for each drive)
    echo "1" > /sys/block/sdb/device/queue_depth

    (increase drive timeout of each device)
    echo "60" > /sys/block/sdb/device/timeout

    The ghost train /dev/md127 array would sometimes snatch /dev/sdc from me. I also zeroed the superblock of /dev/sdc and re-added it to my array.

    Right now it's recovering, I'll post my results.

    Personalities : [raid6] [raid5] [raid4]
    md1 : active raid6 sdc[5] sdf[0] sdb[4] sdd[2] sde[1]
    5860540224 blocks super 1.2 level 6, 64k chunk, algorithm 2 [5/4] [UUU_U]
    [>....................] recovery = 1.4% (28604096/195351340 finish=1442.1min speed=22245K/sec

  18. #18
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    I think I might have fixed this issue by just swapping the cards. I'm currently recovering the raid (purposely failed a drive to fully test the superblock thing from earlier).

    This was a Marvel chipset that will randomly offline disks, looking at forum posts and behavior.

    avoid marvel 88SE9123 like the plague. I just ended up using the Highpoint rocketraid 622A 2 port SATA card with my rig.

    This is exactly my issue:
    http://vip.asus.com/forum/view.aspx?...Language=en-us

    (other articles)
    https://koitsu.wordpress.com/2009/07...6g-controller/
    http://www.extremetech.com/article2/...2350340,00.asp
    http://opensolaris.org/jive/thread.j...ssageID=490180

    in RHEL6 it requires compilation of a kernel module, I found a good post here:
    https://help.ubuntu.com/community/RocketRaid

  19. #19
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    I think the issue I was having is from using a buggy marvel chipset

    avoid marvel 88SE9123 if you're doing a 6-8 drive setup using an external SATA bay (sansdigital).

    This was exactly my issue:
    http://vip.asus.com/forum/view.aspx?...Language=en-us

    (other articles about the chipset)
    https://koitsu.wordpress.com/2009/07...6g-controller/
    http://www.extremetech.com/article2/...2350340,00.asp
    http://opensolaris.org/jive/thread.j...ssageID=490180

    Right now I've swapped out the marvel chipset with a rocketraid 622A that came with the SANS digital unit and the array is rebuilding (i manually failed a drive to troubleshoot the superblock thing from earlier).

    Crossing my fingers.. but I'm hoping the buggy Marvel chipset was the culprit here.

  20. #20
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    Apologies for duplication of info above.

    Unfortunately I spoke too soon, after about 20minutes of running the rr62x module I started seeing hpt resets and FIS calls were unable to go through. The same behaviour occurred - drives were booted out of the controller one at a time and were no longer visible to the system.

    I found a post here that summarizes the same issue I had pretty much.
    http://ubuntuforums.org/showthread.php?t=1592227

    I've since ordered a Sil based chipset SATA controller because other folks have had good experience with it + external SATA bay and I should receive it today,
    (ADSA3GPX1-2E Raid 5 2CH Esata II Pcie). It will be slower internally as it's only SATA II/3gpbs however most of my writes are limited to gigabit line speed anyways.

    For clarification here is the hardware I had issues with. Everything was done in RHEL6 using software raid (mdadm) and the XFS filesystem.

    (storage)
    Sans Digital TowerRAID TR8M-B 8 Bay SATA (comes with rocketraid622a below)
    4 x 2TB Samsung spinpoint F4 5400rpm low-power SATA
    1 x 2TB Seagate 5400 rpm low-power SATA (all drives in 5drive mdadm RAID6)

    (SATA controllers)
    rocket622a raidcard (rr62x kernel module) - hpt resets, drives dissapear
    rocketraid622 controller card (marvel 88SE9123) - drives dissapear.

    When the new sil 332 card arrives today I'll give that a shot and post my findings.

    Here are the logs below from the last card I tried.
    Note: both the Marvell 88SE9123 (622) and rocketrai622a (rr62x) had the same I/O errors below right before the drives were kicked out of the system. Only the rr62x module was verbose enough to log hpt resets and FIS traversal errors.

    == rr62x ==
    rr62x:hpt_reset(8/0/1)
    rr62x:[0 0 ] failed to disable comm status change bits
    rr62x:[0 0 ] start port.
    rr62x:[0 0 ] start port hard reset (probe 1).
    rr62x:[0 1 ] failed to disable comm status change bits
    rr62x:[0 1 ] start port.
    rr62x:[0 1 ] start port hard reset (probe 1).
    rr62x:[0 0 ] start port soft reset (probe 1).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port soft reset (probe 1).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port hard reset (probe 2).
    Clocksource tsc unstable (delta = 202255234 ns)
    Switching to clocksource jiffies
    rr62x:[0 1 ] start port hard reset (probe 2).
    rr62x:[0 1 ] start port soft reset (probe 2).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 2).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 3).
    rr62x:[0 0 ] start port hard reset (probe 3).
    rr62x:[0 1 ] start port soft reset (probe 3).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 3).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 4).
    rr62x:[0 0 ] start port hard reset (probe 4).
    rr62x:[0 1 ] start port soft reset (probe 4).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 4).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 5).
    rr62x:[0 0 ] start port hard reset (probe 5).
    rr62x:[0 1 ] start port soft reset (probe 5).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 5).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 6).
    rr62x:[0 0 ] start port hard reset (probe 6).
    rr62x:[0 1 ] start port soft reset (probe 6).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 6).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 7).
    rr62x:[0 0 ] start port hard reset (probe 7).
    rr62x:[0 1 ] start port soft reset (probe 7).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 0 ] start port soft reset (probe 7).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe .
    rr62x:[0 0 ] start port hard reset (probe .
    rr62x:[0 1 ] start port soft reset (probe .
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 9).
    rr62x:[0 0 ] start port soft reset (probe .
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 0 ] start port hard reset (probe 9).
    rr62x:[0 1 ] start port soft reset (probe 9).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 10).
    rr62x:[0 0 ] start port soft reset (probe 9).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 0 ] start port hard reset (probe 10).
    rr62x:[0 1 ] start port soft reset (probe a).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 11).
    rr62x:[0 0 ] start port soft reset (probe a).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 0 ] start port hard reset (probe 11).
    rr62x:[0 1 ] start port soft reset (probe b).
    rr62x:[0 1 f] failed to send 1st FIS
    rr62x:[0 1 ] start port hard reset (probe 12).
    rr62x:[0 1 3] device disconnected on port.
    rr62x:[0 0 ] start port soft reset (probe b).
    rr62x:[0 0 f] failed to send 1st FIS
    rr62x:[0 0 ] start port hard reset (probe 12).
    rr62x:[0 0 0] device disconnected on port.
    rr62x:[0 0 1] device disconnected on port.
    rr62x:[0 0 2] device disconnected on port.
    rr62x:[0 0 3] device disconnected on port.
    rr62x:hpt_reset(8/0/
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d1 08 00 00 78 00
    end_request: I/O error, dev sdc, sector 885117192
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d8 80 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119104
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d9 00 00 00 50 00
    end_request: I/O error, dev sdc, sector 885119232
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d9 50 00 00 20 00
    end_request: I/O error, dev sdc, sector 885119312
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d9 70 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119344
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d9 f0 00 00 10 00
    end_request: I/O error, dev sdc, sector 885119472
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 da 00 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119488
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 da 80 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119616
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d3 80 00 00 80 00
    end_request: I/O error, dev sdc, sector 885117824
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 db 00 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119744
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 db 80 00 00 80 00
    end_request: I/O error, dev sdc, sector 885119872
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 dc 00 00 00 80 00
    end_request: I/O error, dev sdc, sector 885120000
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 dc 80 00 00 80 00
    end_request: I/O error, dev sdc, sector 885120128
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 dd 00 00 00 08 00
    end_request: I/O error, dev sdc, sector 885120256
    scsi 8:0:1:0: [sdc] Unhandled error code
    scsi 8:0:1:0: [sdc] Result: hostbyte=DID_BAD_TARGET driverbyte=DRIVER_OK
    scsi 8:0:1:0: [sdc] CDB: Read(10): 28 00 34 c1 d6 00 00 00 80 00
    end_request: I/O error, dev sdc, sector 885118464
    scsi 8:0:8:0: rejecting I/O to offline device
    scsi 8:0:8:0: [sdf] Unhandled error code
    scsi 8:0:8:0: [sdf] Result: hostbyte=DID_NO_CONNECT driverbyte=DRIVER_OK
    scsi 8:0:8:0: [sdf] CDB: Write(10): 2a 00 34 c1 d8 00 00 00 80 00
    end_request: I/O error, dev sdf, sector 885118976
    raid5: Disk failure on sdf, disabling device.
    raid5: Operation continuing on 4 devices.
    raid5: Disk failure on sdc, disabling device.
    raid5: Operation continuing on 3 devices.
    end_request: I/O error, dev sde, sector 8
    md: super_written gets error=-5, uptodate=0
    raid5: Disk failure on sde, disabling device.
    raid5: Operation continuing on 2 devices.
    end_request: I/O error, dev sdb, sector 8
    md: super_written gets error=-5, uptodate=0
    raid5: Disk failure on sdb, disabling device.
    raid5: Operation continuing on 1 devices.
    end_request: I/O error, dev sdd, sector 8
    md: super_written gets error=-5, uptodate=0
    raid5: Disk failure on sdd, disabling device.
    raid5: Operation continuing on 0 devices.
    md: md1: recovery done.
    RAID5 conf printout:

  21. #21
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    I've got the ADSA3GPX1-2E Raid 5 2CH Esata II Pcie in place.

    My drive order changed (sda is no longer my OS drive, it's now sdf)
    My raid drives are now sda/sdb/sdc/sdd/sde and I'm recovering the raid.

    I tried setting my mobo SATA chipset to AHCI but it would only boot if the external 8-bay was not connected.

    I'll post my findings but it will be a bit. I'd like to be optimistic but this is the 3rd SATA controller I've tried, though a different chipset (SiL based vs. Marvell based).

    [root@poopsock ~]# cat /proc/mdstat
    Personalities : [raid6] [raid5] [raid4]
    md1 : active raid6 sda[5] sde[0] sdb[4] sdc[2] sdd[1]
    5860540224 blocks super 1.2 level 6, 64k chunk, algorithm 2 [5/4] [UUU_U]
    [>....................] recovery = 3.2% (64301120/195351340 finish=1061.4min speed=29663K/sec

    (yes, my hostname is poopsock)
    http://www.urbandictionary.com/define.php?term=poopsock

  22. #22
    What is this storage?
    Join Date
    Jun 2011
    Posts
    1
    I CANNOT THANK YOU ENOUGH!

    I have subscribed to this forum just to say THANK YOU after searching google!

    I was expanding my raid 6 array when one drive failed during reshaping.

    Thought the array was gone for good 4tb worth of precious data (photos videos music ALL PERSONAL).

    Tried recreating and all sorts of stuff.

    This command saved my life:
    mdadm --assemble --scan -fv

    THANK YOU SO MUCH FROM AUSTRALIA!

  23. #23
    Storage? I am Storage! Bozo's Avatar
    Join Date
    Feb 2002
    Location
    Twilight Zone
    Posts
    4,218
    You're welcome! Stop in anytime.
    Bozo

  24. #24
    What is this storage?
    Join Date
    Oct 2011
    Posts
    4
    So I've got a 6 disk RAID 5 array and recently lost two disks. I replaced them but can't get the array back online. I believe part of the problem is the order of the disks has changed. What I'm getting from this post is that I can delete the superblock on ALL the disks and then recreate the raid array with the same info (number of disks, chunk size, etc.) without necessarily losing all the data?

  25. #25
    Fixture ddrueding's Avatar
    Join Date
    Feb 2002
    Location
    Monterey, CA
    Age
    34
    Posts
    17,161
    RAID5 only protects against a single drive failure. If you weren't able to replace that drive before another drive failed, you might be in big trouble.
    Work1: i7-990X@4.6Ghz, 24GB RAM, RevoDrive 3 X2 480GB, 2x GTX690, G1.Assassin, Noctua NH-C14, AX1200i
    Work2:
    i7-4770, 32GB RAM, 840 Pro 512GB, 3x 3TB, 2x Titan, GA-G1.SNIPER 5, H100i, AX1200i
    Home1: i7-3970X@4.7GHz, 32GB, Vertex4 512GB, GTX690, ASUS P9X79 WS, Corsair AX1200i, Watercooled
    Home2: Intel i3-2120T, 16GB RAM, OCZ NOC-MSATA 120GB, PicoPSU 120W
    , Watercooled

  26. #26
    I can't believe I'm a Fixture LunarMist's Avatar
    Join Date
    Jan 2003
    Location
    USA
    Posts
    10,595
    Yeah. It's time for a rebuild and restoration of data from backup.
    --Lunar

  27. #27
    What is this storage?
    Join Date
    Oct 2011
    Posts
    4
    Quote Originally Posted by ddrueding View Post
    RAID5 only protects against a single drive failure. If you weren't able to replace that drive before another drive failed, you might be in big trouble.
    Yea, I realize that. I was just holding onto some sliver of hope that I might be able to recover some of the data, no matter how small. I guess since I have nothing to lose I'll nuke the superblocks of the existing drives and see what happens.....

  28. #28
    Fixture ddrueding's Avatar
    Join Date
    Feb 2002
    Location
    Monterey, CA
    Age
    34
    Posts
    17,161
    How did your drives fail? If they can still be detected, you can try some RAID recovery programs and see what it can find.

    If the drives are completely inaccessible, time to throw in the towel. As every file is striped across every drive, you have lost some part of any file big enough to fill a stripe.
    Work1: i7-990X@4.6Ghz, 24GB RAM, RevoDrive 3 X2 480GB, 2x GTX690, G1.Assassin, Noctua NH-C14, AX1200i
    Work2:
    i7-4770, 32GB RAM, 840 Pro 512GB, 3x 3TB, 2x Titan, GA-G1.SNIPER 5, H100i, AX1200i
    Home1: i7-3970X@4.7GHz, 32GB, Vertex4 512GB, GTX690, ASUS P9X79 WS, Corsair AX1200i, Watercooled
    Home2: Intel i3-2120T, 16GB RAM, OCZ NOC-MSATA 120GB, PicoPSU 120W
    , Watercooled

  29. #29
    What is this storage?
    Join Date
    Oct 2011
    Posts
    4
    I'll have to hook them back up and see if I can get them to respond. When I noticed, they were clicking loudly and the machine would essentially lock when detecting them (upon reboot).
    Perhaps I'll try freezing them and then doing a dd. I'm not really sure how much data is on each drive...

  30. #30
    Fixture ddrueding's Avatar
    Join Date
    Feb 2002
    Location
    Monterey, CA
    Age
    34
    Posts
    17,161
    I've used RAID Reconstructor before with positive results, but it sounds like those drives might be very dead.
    Work1: i7-990X@4.6Ghz, 24GB RAM, RevoDrive 3 X2 480GB, 2x GTX690, G1.Assassin, Noctua NH-C14, AX1200i
    Work2:
    i7-4770, 32GB RAM, 840 Pro 512GB, 3x 3TB, 2x Titan, GA-G1.SNIPER 5, H100i, AX1200i
    Home1: i7-3970X@4.7GHz, 32GB, Vertex4 512GB, GTX690, ASUS P9X79 WS, Corsair AX1200i, Watercooled
    Home2: Intel i3-2120T, 16GB RAM, OCZ NOC-MSATA 120GB, PicoPSU 120W
    , Watercooled

  31. #31
    Serial computer killer Hairy Aussie CougTek's Avatar
    Join Date
    Jan 2002
    Location
    Québec, Québec
    Posts
    7,873
    Quote Originally Posted by ogp View Post
    When I noticed, they were clicking loudly and the machine would essentially lock when detecting them (upon reboot).
    Dead drives exhibiting the click-of-death syndrome are unrecoverable by common means. You'll need to deal with professional recovery services. That means a company that disassembles your hard drive in a clean room. Here, they charges between one and two thousand per drive and the result isn't garanteed. If it's valuable company data, it might be worth it, but if it's just the family photos of your last trip to Disneyland and the ugly face of aunt Betty, then you might want to reconsider.

  32. #32
    What is this storage?
    Join Date
    Oct 2011
    Posts
    4
    Yea, nothing important or work related. Just lots of large media files... I'm not holding my breath for any sort of restoration.

  33. #33
    What is this storage?
    Join Date
    Mar 2011
    Location
    Raleigh, NC
    Posts
    8
    Hey Folks,

    As an update, I had no problems with the Addonics ADSA3GPX1-2E (SiL Based).
    It cost me around $30 from Amazon.

    http://www.amazon.com/ADSA3GPX1-2E-R...dp/B000UUN70A/

    The SIL 3132 chipset based cards are the only ones that worked for me.
    My setup is now working 100% with no issues, the SATA I/O disconnect issues are gone.

    A friend of mine purchased a similiar tower made by Rosewill which included a SiL 3132 based card and his works as intended, it seems the SANS digital ship with highpoint (Marvell based) cards.

    My setup ::

    8x 2TB Samsung Spinpoint F4
    Linux mdadm RAID6
    RHEL6.2

  34. #34
    What is this storage?
    Join Date
    Feb 2012
    Posts
    1
    Quote Originally Posted by Gilbo View Post
    1. I examined the good superblocks to make sure I knew how to recreate the array exactly (i.e. chunk size, etc).
    2. I zeroed all the superblocks on all the disks.
    3. I recreated the array with the "--assume-clean" option.
    4. I marked it possibly dirty with:
    Code:
    mdadm --assemble /dev/md0 --update=resync
    5. I let it try to resync. This only took about 30-60 minutes, probably because all the data was good and it was just reading from each disk and not writing.
    6. I mounted it & everything so far appears to be good.

    Alternately, I think I could have simply recreated it without "--assume-clean" and the initial creation sync wouldn't have destroyed the data, but that wasn't a risk I was willing to take.

    Once I got over my fear of accidentally destroying the data, it didn't really take me long.
    Still saving lives!!!

    Had an old TeraStation Buffalo drive die and using google and this post was able to finally get it up and running!

    Thanks again!!!

  35. #35
    What is this storage?
    Join Date
    Feb 2012
    Posts
    2
    I am getting "failed to send 1st FIS" with the Sans Digital TowerRAID TR8M-B 8 Bay SATA and rocketraid 622a and rr62x userland raid module. I am wondering if I replaced the rocketraid with another controller if my errors would go away. Any suggestions are welcome.

  36. #36
    Fixture ddrueding's Avatar
    Join Date
    Feb 2002
    Location
    Monterey, CA
    Age
    34
    Posts
    17,161
    I have that hardware configuration, but in Win7, and without error. Nice kit, but one of the Linux guys around here will need to help you. Shouldn't be long
    Work1: i7-990X@4.6Ghz, 24GB RAM, RevoDrive 3 X2 480GB, 2x GTX690, G1.Assassin, Noctua NH-C14, AX1200i
    Work2:
    i7-4770, 32GB RAM, 840 Pro 512GB, 3x 3TB, 2x Titan, GA-G1.SNIPER 5, H100i, AX1200i
    Home1: i7-3970X@4.7GHz, 32GB, Vertex4 512GB, GTX690, ASUS P9X79 WS, Corsair AX1200i, Watercooled
    Home2: Intel i3-2120T, 16GB RAM, OCZ NOC-MSATA 120GB, PicoPSU 120W
    , Watercooled

  37. #37
    What is this storage?
    Join Date
    Feb 2012
    Posts
    2
    Thanks ddrueding. I have it working in Win7 too. As you say, it works well. My issues are with the linux driver/card, fedora 14 to be more specific. For RAID redundency, it's just not stable enough with my configuration. I'm willing to buy a GOOD card for good stabilty and speed.

  38. #38
    I had a failed drive sda and I replaced it and when I wanted to re-add it to md device I encountered following issue.


    Before adding the disk I just did fdisk this was the output



    root@host ~]# fdisk -l

    Disk /dev/sda: 1000.2 GB, 1000204886016 bytes
    255 heads, 63 sectors/track, 121601 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes

    Device Boot Start End Blocks Id System
    /dev/sda1 * 1 65 522081 fd Linux raid autodetect
    /dev/sda2 66 121601 976237920 fd Linux raid autodetect

    Disk /dev/sdb: 1000.2 GB, 1000204886016 bytes
    255 heads, 63 sectors/track, 121601 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes

    Device Boot Start End Blocks Id System
    /dev/sdb1 * 1 65 522081 fd Linux raid autodetect
    /dev/sdb2 66 121601 976237920 fd Linux raid autodetect

    Disk /dev/md1: 999.6 GB, 999667531776 bytes
    2 heads, 4 sectors/track, 244059456 cylinders
    Units = cylinders of 8 * 512 = 4096 bytes

    Disk /dev/md1 doesn't contain a valid partition table

    Disk /dev/md0: 534 MB, 534511616 bytes
    2 heads, 4 sectors/track, 130496 cylinders
    Units = cylinders of 8 * 512 = 4096 bytes

    Disk /dev/md0 doesn't contain a valid partition table



    ==============================

    But when I tried to add the sda1 into my md0 raid it went perfect but when I tried to add sda2 into md1 it failed telling that no such device found. And when I did fdisk -l again I saw



    [root@host ~]# fdisk -l

    Disk /dev/sdb: 1000.2 GB, 1000204886016 bytes
    255 heads, 63 sectors/track, 121601 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes

    Device Boot Start End Blocks Id System
    /dev/sdb1 * 1 65 522081 fd Linux raid autodetect
    /dev/sdb2 66 121601 976237920 fd Linux raid autodetect

    Disk /dev/md1: 999.6 GB, 999667531776 bytes
    2 heads, 4 sectors/track, 244059456 cylinders
    Units = cylinders of 8 * 512 = 4096 bytes

    Disk /dev/md1 doesn't contain a valid partition table

    Disk /dev/md0: 534 MB, 534511616 bytes
    2 heads, 4 sectors/track, 130496 cylinders
    Units = cylinders of 8 * 512 = 4096 bytes

    Disk /dev/md0 doesn't contain a valid partition table

    Disk /dev/sdc: 1000.2 GB, 1000204886016 bytes
    255 heads, 63 sectors/track, 121601 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes

    Device Boot Start End Blocks Id System
    /dev/sdc1 * 1 65 522081 fd Linux raid autodetect
    /dev/sdc2 66 121601 976237920 fd Linux raid autodetect
    You have new mail in /var/spool/mail/root

    ==============



    Suprisingly linux detected the new drive suddenly as sdc1. And now if I want to delete the sda1 from md0 so that I could add sdc1 its not allowing me saying sda1 no such device. Please help...

    Dmesh at below pastebin

    http://fpaste.org/qwdh/

  39. #39
    What is this storage?
    Join Date
    Aug 2012
    Posts
    1
    I have a raid 5 array that is in a bit of a messed up state. Fully detailed here, http://www.linuxquestions.org/questi...22#post4763622.

    I am trying to follow the procedure Gilbo outlines above to rebuild the array, since the superblock is missing from one of the partitions, and when I attempt to build the array after zeroing out the superblocks, I get this,

    [root@moneypit RAID_RECOVERY]# mdadm --build /dev/md5 --chunk=128 --level=raid5 --raid-devices=3 --assume-clean /dev/sdc1 /dev/sdd1 /dev/sdf1
    mdadm: Raid level raid5 not permitted with --build.

    I am not sure exactly what the command is meant when he says, " I recreated the array with the "--assume-clean" option." I assumed it was the --build option, but clearly it is not.

    Anybody have any suggestions?

Thread Information

Users Browsing this Thread

There are currently 1 users browsing this thread. (0 members and 1 guests)

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •