MDADM can't automount existing RAID1 arraymdadm; previously working; after “failure”, cannot join array due to disk sizemdadm - RAID5 array size vs. actual disk size mismatchAn existing mdadm RAID5 is not mounting, Either a problem drive or SuperblockHow to replace faulty device on RAID5Replacing a disk to RAID5 failedHow do I (re)build/create/assemble an IMSM RAID-0 array from disk images instead of disk drives using mdadm?MDADM lists same device twiceError setting up Raid 1 with mdadm at RaspberryPi 2Help with mdadm will not reassembleRebuilding mdadm RAID 5 array with multiple failed drives

Why is a very small peak with larger m/z not considered to be the molecular ion?

Shifting between bemols and diesis in the key signature

Gaining more land

I reported the illegal activity of my boss to his boss. My boss found out. Now I am being punished. What should I do?

Trig Subsitution When There's No Square Root

Which situations would cause a company to ground or recall a aircraft series?

Can the alpha, lambda values of a glmnet object output determine whether ridge or Lasso?

What would be the most expensive material to an intergalactic society?

Is it a Cyclops number? "Nobody" knows!

Power Strip for Europe

How to design an organic heat-shield?

Why does Solve lock up when trying to solve the quadratic equation with large integers?

Is this Paypal Github SDK reference really a dangerous site?

Is it possible to find 2014 distinct positive integers whose sum is divisible by each of them?

Why do phishing e-mails use faked e-mail addresses instead of the real one?

Can't make sense of a paragraph from Lovecraft

From an axiomatic set theoric approach why can we take uncountable unions?

How does Ehrenfest's theorem apply to the quantum harmonic oscillator?

MySQL importing CSV files really slow

Can I negotiate a patent idea for a raise, under French law?

Would an aboleth's Phantasmal Force lair action be affected by Counterspell, Dispel Magic, and/or Slow?

What is the population of Romulus in the TNG era?

Are small insurances worth it?

What is better: yes / no radio, or simple checkbox?



MDADM can't automount existing RAID1 array


mdadm; previously working; after “failure”, cannot join array due to disk sizemdadm - RAID5 array size vs. actual disk size mismatchAn existing mdadm RAID5 is not mounting, Either a problem drive or SuperblockHow to replace faulty device on RAID5Replacing a disk to RAID5 failedHow do I (re)build/create/assemble an IMSM RAID-0 array from disk images instead of disk drives using mdadm?MDADM lists same device twiceError setting up Raid 1 with mdadm at RaspberryPi 2Help with mdadm will not reassembleRebuilding mdadm RAID 5 array with multiple failed drives













0















I recently reformatted my system and cannot get my RAID 1 array to load on boot. Would appreciate some help figuring out why this is. Using ubuntu 14 with MDADM. Its a very simple RAID1 array consisting of 2 drives sda and sdb. I tried to edit fstab to mount the RAID but it fails at boot and asks me to press a key to skip mounting.



Here's my MDADM.conf file



# mdadm.conf
#
# Please refer to mdadm.conf(5) for information about this file.
#

# by default (built-in), scan all partitions (/proc/partitions) and all
# containers for MD superblocks. alternatively, specify devices to scan, using
# wildcards if desired.
#DEVICE partitions containers

# auto-create devices with Debian standard permissions
CREATE owner=root group=disk mode=0660 auto=yes

# automatically tag new arrays as belonging to the local system
HOMEHOST <system>

# instruct the monitoring daemon where to send mail alerts
MAILADDR root

# definitions of existing MD arrays
ARRAY metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
ARRAY /dev/md/Storage(RAID) container=4e85cd11:34b3cd40:f263b2be:616ef7fb member=0 UUID=c26e7f0d:a956fc70:db00c387:58e7a99a

# This file was auto-generated on Sun, 10 Mar 2019 16:31:50 -0500
# by mkconf $Id$


Output from mdadm --Examine /dev/sda:



/dev/sda:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.3.00
Orig Family : 96f17bdd
Family : 96f17bdd
Generation : 000bd2b7
Attributes : All supported
UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
Checksum : 7eee6b4f correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1

Disk00 Serial : ZAD04VWC
State : active
Id : 00000004
Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

[Storage(RAID)]:
UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
RAID Level : 1
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 0
Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
Sector Offset : 0
Num Stripes : 45785312
Chunk Size : 64 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean

Disk01 Serial : ZAD05L4F
State : active
Id : 00000006
Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


Output from mdadm --Examine /dev/sdb:



/dev/sdb:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.3.00
Orig Family : 96f17bdd
Family : 96f17bdd
Generation : 000bd2b7
Attributes : All supported
UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
Checksum : 7eee6b4f correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1

Disk01 Serial : ZAD05L4F
State : active
Id : 00000006
Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

[Storage(RAID)]:
UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
RAID Level : 1
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 1
Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
Sector Offset : 0
Num Stripes : 45785312
Chunk Size : 64 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : clean

Disk00 Serial : ZAD04VWC
State : active
Id : 00000004
Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


Output from lsblk -o NAME,SIZE,FSTYPE,TYPE,MOUNTPOINT gives the following:



NAME SIZE FSTYPE TYPE MOUNTPOINT
sda 5.5T disk
└─md126 5.5T raid1
├─md126p1 128M md
└─md126p2 5.5T md
sdb 5.5T disk
└─md126 5.5T raid1
├─md126p1 128M md
└─md126p2 5.5T md
sr0 1024M rom
nvme0n1 477G disk
├─nvme0n1p1 499M part
├─nvme0n1p2 99M part /boot/efi
├─nvme0n1p3 16M part
├─nvme0n1p4 237.4G part
├─nvme0n1p5 470M part
├─nvme0n1p6 232.9G part /
└─nvme0n1p9 1.9G part [SWAP]


If I run sudo mdadm --verbose --assemble --scan after boot, here is the output



mdadm: looking for devices for further assembly
mdadm: cannot open device /dev/sr0: No medium found
mdadm: no RAID superblock on /dev/nvme0n1p9
mdadm: no RAID superblock on /dev/nvme0n1p6
mdadm: no RAID superblock on /dev/nvme0n1p5
mdadm: no RAID superblock on /dev/nvme0n1p4
mdadm: no RAID superblock on /dev/nvme0n1p3
mdadm: no RAID superblock on /dev/nvme0n1p2
mdadm: no RAID superblock on /dev/nvme0n1p1
mdadm: no RAID superblock on /dev/nvme0n1
mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot -1.
mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot -1.
mdadm: added /dev/sda to /dev/md/imsm0 as -1
mdadm: added /dev/sdb to /dev/md/imsm0 as -1
mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
mdadm: looking for devices for further assembly
mdadm: looking for devices for further assembly
mdadm: /dev/sdb is busy - skipping
mdadm: /dev/sda is busy - skipping
mdadm: cannot open device /dev/sr0: No medium found
mdadm: no recogniseable superblock on /dev/nvme0n1p9
mdadm: no recogniseable superblock on /dev/nvme0n1p6
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
mdadm: no recogniseable superblock on /dev/nvme0n1p3
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
mdadm: looking in container /dev/md127
mdadm: found match on member /md127/0 in /dev/md127
mdadm: Started /dev/md/Storage(RAID)_0 with 2 devices
mdadm: looking for devices for further assembly
mdadm: /dev/sdb is busy - skipping
mdadm: /dev/sda is busy - skipping
mdadm: looking in container /dev/md127
mdadm: member /md127/0 in /dev/md127 is already assembled
mdadm: looking for devices for further assembly
mdadm: Cannot assemble mbr metadata on /dev/md/Storage_RAID__0p2
mdadm: no recogniseable superblock on /dev/md/Storage_RAID__0p1
mdadm: Cannot assemble mbr metadata on /dev/md126
mdadm: /dev/sdb is busy - skipping
mdadm: /dev/sda is busy - skipping
mdadm: cannot open device /dev/sr0: No medium found
mdadm: no recogniseable superblock on /dev/nvme0n1p9
mdadm: no recogniseable superblock on /dev/nvme0n1p6
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
mdadm: no recogniseable superblock on /dev/nvme0n1p3
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
mdadm: looking in container /dev/md127
mdadm: member /md127/0 in /dev/md127 is already assembled


I've made sure initramfs is updated when making changes to mdadm.conf. And once I run the --assemble --scan command, it does mount the arrays. If I then do cat /proc/mdstat I get:



md127 : inactive sdb[1](S) sda[0](S)
5032 blocks super external:imsm

md126 : active (auto-read-only) raid1 sda[1] sdb[0]
5860519936 blocks super external:/md127/0 [2/2] [UU]


But if I try to do mdadm --detail --scan to add to the mdadm.conf file I get this:



ARRAY /dev/md/imsm0 metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
mdadm: cannot open /dev/md/Storage(RAID)_0: No such file or directory


I need this array to mount at boot but have no idea what is stopping that from happening.










share|improve this question







New contributor




ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.
























    0















    I recently reformatted my system and cannot get my RAID 1 array to load on boot. Would appreciate some help figuring out why this is. Using ubuntu 14 with MDADM. Its a very simple RAID1 array consisting of 2 drives sda and sdb. I tried to edit fstab to mount the RAID but it fails at boot and asks me to press a key to skip mounting.



    Here's my MDADM.conf file



    # mdadm.conf
    #
    # Please refer to mdadm.conf(5) for information about this file.
    #

    # by default (built-in), scan all partitions (/proc/partitions) and all
    # containers for MD superblocks. alternatively, specify devices to scan, using
    # wildcards if desired.
    #DEVICE partitions containers

    # auto-create devices with Debian standard permissions
    CREATE owner=root group=disk mode=0660 auto=yes

    # automatically tag new arrays as belonging to the local system
    HOMEHOST <system>

    # instruct the monitoring daemon where to send mail alerts
    MAILADDR root

    # definitions of existing MD arrays
    ARRAY metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
    ARRAY /dev/md/Storage(RAID) container=4e85cd11:34b3cd40:f263b2be:616ef7fb member=0 UUID=c26e7f0d:a956fc70:db00c387:58e7a99a

    # This file was auto-generated on Sun, 10 Mar 2019 16:31:50 -0500
    # by mkconf $Id$


    Output from mdadm --Examine /dev/sda:



    /dev/sda:
    Magic : Intel Raid ISM Cfg Sig.
    Version : 1.3.00
    Orig Family : 96f17bdd
    Family : 96f17bdd
    Generation : 000bd2b7
    Attributes : All supported
    UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
    Checksum : 7eee6b4f correct
    MPB Sectors : 1
    Disks : 2
    RAID Devices : 1

    Disk00 Serial : ZAD04VWC
    State : active
    Id : 00000004
    Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

    [Storage(RAID)]:
    UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
    RAID Level : 1
    Members : 2
    Slots : [UU]
    Failed disk : none
    This Slot : 0
    Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
    Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
    Sector Offset : 0
    Num Stripes : 45785312
    Chunk Size : 64 KiB
    Reserved : 0
    Migrate State : idle
    Map State : normal
    Dirty State : clean

    Disk01 Serial : ZAD05L4F
    State : active
    Id : 00000006
    Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


    Output from mdadm --Examine /dev/sdb:



    /dev/sdb:
    Magic : Intel Raid ISM Cfg Sig.
    Version : 1.3.00
    Orig Family : 96f17bdd
    Family : 96f17bdd
    Generation : 000bd2b7
    Attributes : All supported
    UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
    Checksum : 7eee6b4f correct
    MPB Sectors : 1
    Disks : 2
    RAID Devices : 1

    Disk01 Serial : ZAD05L4F
    State : active
    Id : 00000006
    Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

    [Storage(RAID)]:
    UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
    RAID Level : 1
    Members : 2
    Slots : [UU]
    Failed disk : none
    This Slot : 1
    Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
    Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
    Sector Offset : 0
    Num Stripes : 45785312
    Chunk Size : 64 KiB
    Reserved : 0
    Migrate State : idle
    Map State : normal
    Dirty State : clean

    Disk00 Serial : ZAD04VWC
    State : active
    Id : 00000004
    Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


    Output from lsblk -o NAME,SIZE,FSTYPE,TYPE,MOUNTPOINT gives the following:



    NAME SIZE FSTYPE TYPE MOUNTPOINT
    sda 5.5T disk
    └─md126 5.5T raid1
    ├─md126p1 128M md
    └─md126p2 5.5T md
    sdb 5.5T disk
    └─md126 5.5T raid1
    ├─md126p1 128M md
    └─md126p2 5.5T md
    sr0 1024M rom
    nvme0n1 477G disk
    ├─nvme0n1p1 499M part
    ├─nvme0n1p2 99M part /boot/efi
    ├─nvme0n1p3 16M part
    ├─nvme0n1p4 237.4G part
    ├─nvme0n1p5 470M part
    ├─nvme0n1p6 232.9G part /
    └─nvme0n1p9 1.9G part [SWAP]


    If I run sudo mdadm --verbose --assemble --scan after boot, here is the output



    mdadm: looking for devices for further assembly
    mdadm: cannot open device /dev/sr0: No medium found
    mdadm: no RAID superblock on /dev/nvme0n1p9
    mdadm: no RAID superblock on /dev/nvme0n1p6
    mdadm: no RAID superblock on /dev/nvme0n1p5
    mdadm: no RAID superblock on /dev/nvme0n1p4
    mdadm: no RAID superblock on /dev/nvme0n1p3
    mdadm: no RAID superblock on /dev/nvme0n1p2
    mdadm: no RAID superblock on /dev/nvme0n1p1
    mdadm: no RAID superblock on /dev/nvme0n1
    mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot -1.
    mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot -1.
    mdadm: added /dev/sda to /dev/md/imsm0 as -1
    mdadm: added /dev/sdb to /dev/md/imsm0 as -1
    mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
    mdadm: looking for devices for further assembly
    mdadm: looking for devices for further assembly
    mdadm: /dev/sdb is busy - skipping
    mdadm: /dev/sda is busy - skipping
    mdadm: cannot open device /dev/sr0: No medium found
    mdadm: no recogniseable superblock on /dev/nvme0n1p9
    mdadm: no recogniseable superblock on /dev/nvme0n1p6
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
    mdadm: no recogniseable superblock on /dev/nvme0n1p3
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
    mdadm: looking in container /dev/md127
    mdadm: found match on member /md127/0 in /dev/md127
    mdadm: Started /dev/md/Storage(RAID)_0 with 2 devices
    mdadm: looking for devices for further assembly
    mdadm: /dev/sdb is busy - skipping
    mdadm: /dev/sda is busy - skipping
    mdadm: looking in container /dev/md127
    mdadm: member /md127/0 in /dev/md127 is already assembled
    mdadm: looking for devices for further assembly
    mdadm: Cannot assemble mbr metadata on /dev/md/Storage_RAID__0p2
    mdadm: no recogniseable superblock on /dev/md/Storage_RAID__0p1
    mdadm: Cannot assemble mbr metadata on /dev/md126
    mdadm: /dev/sdb is busy - skipping
    mdadm: /dev/sda is busy - skipping
    mdadm: cannot open device /dev/sr0: No medium found
    mdadm: no recogniseable superblock on /dev/nvme0n1p9
    mdadm: no recogniseable superblock on /dev/nvme0n1p6
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
    mdadm: no recogniseable superblock on /dev/nvme0n1p3
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
    mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
    mdadm: looking in container /dev/md127
    mdadm: member /md127/0 in /dev/md127 is already assembled


    I've made sure initramfs is updated when making changes to mdadm.conf. And once I run the --assemble --scan command, it does mount the arrays. If I then do cat /proc/mdstat I get:



    md127 : inactive sdb[1](S) sda[0](S)
    5032 blocks super external:imsm

    md126 : active (auto-read-only) raid1 sda[1] sdb[0]
    5860519936 blocks super external:/md127/0 [2/2] [UU]


    But if I try to do mdadm --detail --scan to add to the mdadm.conf file I get this:



    ARRAY /dev/md/imsm0 metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
    mdadm: cannot open /dev/md/Storage(RAID)_0: No such file or directory


    I need this array to mount at boot but have no idea what is stopping that from happening.










    share|improve this question







    New contributor




    ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
    Check out our Code of Conduct.






















      0












      0








      0








      I recently reformatted my system and cannot get my RAID 1 array to load on boot. Would appreciate some help figuring out why this is. Using ubuntu 14 with MDADM. Its a very simple RAID1 array consisting of 2 drives sda and sdb. I tried to edit fstab to mount the RAID but it fails at boot and asks me to press a key to skip mounting.



      Here's my MDADM.conf file



      # mdadm.conf
      #
      # Please refer to mdadm.conf(5) for information about this file.
      #

      # by default (built-in), scan all partitions (/proc/partitions) and all
      # containers for MD superblocks. alternatively, specify devices to scan, using
      # wildcards if desired.
      #DEVICE partitions containers

      # auto-create devices with Debian standard permissions
      CREATE owner=root group=disk mode=0660 auto=yes

      # automatically tag new arrays as belonging to the local system
      HOMEHOST <system>

      # instruct the monitoring daemon where to send mail alerts
      MAILADDR root

      # definitions of existing MD arrays
      ARRAY metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
      ARRAY /dev/md/Storage(RAID) container=4e85cd11:34b3cd40:f263b2be:616ef7fb member=0 UUID=c26e7f0d:a956fc70:db00c387:58e7a99a

      # This file was auto-generated on Sun, 10 Mar 2019 16:31:50 -0500
      # by mkconf $Id$


      Output from mdadm --Examine /dev/sda:



      /dev/sda:
      Magic : Intel Raid ISM Cfg Sig.
      Version : 1.3.00
      Orig Family : 96f17bdd
      Family : 96f17bdd
      Generation : 000bd2b7
      Attributes : All supported
      UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
      Checksum : 7eee6b4f correct
      MPB Sectors : 1
      Disks : 2
      RAID Devices : 1

      Disk00 Serial : ZAD04VWC
      State : active
      Id : 00000004
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

      [Storage(RAID)]:
      UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
      RAID Level : 1
      Members : 2
      Slots : [UU]
      Failed disk : none
      This Slot : 0
      Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
      Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
      Sector Offset : 0
      Num Stripes : 45785312
      Chunk Size : 64 KiB
      Reserved : 0
      Migrate State : idle
      Map State : normal
      Dirty State : clean

      Disk01 Serial : ZAD05L4F
      State : active
      Id : 00000006
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


      Output from mdadm --Examine /dev/sdb:



      /dev/sdb:
      Magic : Intel Raid ISM Cfg Sig.
      Version : 1.3.00
      Orig Family : 96f17bdd
      Family : 96f17bdd
      Generation : 000bd2b7
      Attributes : All supported
      UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
      Checksum : 7eee6b4f correct
      MPB Sectors : 1
      Disks : 2
      RAID Devices : 1

      Disk01 Serial : ZAD05L4F
      State : active
      Id : 00000006
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

      [Storage(RAID)]:
      UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
      RAID Level : 1
      Members : 2
      Slots : [UU]
      Failed disk : none
      This Slot : 1
      Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
      Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
      Sector Offset : 0
      Num Stripes : 45785312
      Chunk Size : 64 KiB
      Reserved : 0
      Migrate State : idle
      Map State : normal
      Dirty State : clean

      Disk00 Serial : ZAD04VWC
      State : active
      Id : 00000004
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


      Output from lsblk -o NAME,SIZE,FSTYPE,TYPE,MOUNTPOINT gives the following:



      NAME SIZE FSTYPE TYPE MOUNTPOINT
      sda 5.5T disk
      └─md126 5.5T raid1
      ├─md126p1 128M md
      └─md126p2 5.5T md
      sdb 5.5T disk
      └─md126 5.5T raid1
      ├─md126p1 128M md
      └─md126p2 5.5T md
      sr0 1024M rom
      nvme0n1 477G disk
      ├─nvme0n1p1 499M part
      ├─nvme0n1p2 99M part /boot/efi
      ├─nvme0n1p3 16M part
      ├─nvme0n1p4 237.4G part
      ├─nvme0n1p5 470M part
      ├─nvme0n1p6 232.9G part /
      └─nvme0n1p9 1.9G part [SWAP]


      If I run sudo mdadm --verbose --assemble --scan after boot, here is the output



      mdadm: looking for devices for further assembly
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no RAID superblock on /dev/nvme0n1p9
      mdadm: no RAID superblock on /dev/nvme0n1p6
      mdadm: no RAID superblock on /dev/nvme0n1p5
      mdadm: no RAID superblock on /dev/nvme0n1p4
      mdadm: no RAID superblock on /dev/nvme0n1p3
      mdadm: no RAID superblock on /dev/nvme0n1p2
      mdadm: no RAID superblock on /dev/nvme0n1p1
      mdadm: no RAID superblock on /dev/nvme0n1
      mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot -1.
      mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot -1.
      mdadm: added /dev/sda to /dev/md/imsm0 as -1
      mdadm: added /dev/sdb to /dev/md/imsm0 as -1
      mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
      mdadm: looking for devices for further assembly
      mdadm: looking for devices for further assembly
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no recogniseable superblock on /dev/nvme0n1p9
      mdadm: no recogniseable superblock on /dev/nvme0n1p6
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
      mdadm: no recogniseable superblock on /dev/nvme0n1p3
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
      mdadm: looking in container /dev/md127
      mdadm: found match on member /md127/0 in /dev/md127
      mdadm: Started /dev/md/Storage(RAID)_0 with 2 devices
      mdadm: looking for devices for further assembly
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: looking in container /dev/md127
      mdadm: member /md127/0 in /dev/md127 is already assembled
      mdadm: looking for devices for further assembly
      mdadm: Cannot assemble mbr metadata on /dev/md/Storage_RAID__0p2
      mdadm: no recogniseable superblock on /dev/md/Storage_RAID__0p1
      mdadm: Cannot assemble mbr metadata on /dev/md126
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no recogniseable superblock on /dev/nvme0n1p9
      mdadm: no recogniseable superblock on /dev/nvme0n1p6
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
      mdadm: no recogniseable superblock on /dev/nvme0n1p3
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
      mdadm: looking in container /dev/md127
      mdadm: member /md127/0 in /dev/md127 is already assembled


      I've made sure initramfs is updated when making changes to mdadm.conf. And once I run the --assemble --scan command, it does mount the arrays. If I then do cat /proc/mdstat I get:



      md127 : inactive sdb[1](S) sda[0](S)
      5032 blocks super external:imsm

      md126 : active (auto-read-only) raid1 sda[1] sdb[0]
      5860519936 blocks super external:/md127/0 [2/2] [UU]


      But if I try to do mdadm --detail --scan to add to the mdadm.conf file I get this:



      ARRAY /dev/md/imsm0 metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
      mdadm: cannot open /dev/md/Storage(RAID)_0: No such file or directory


      I need this array to mount at boot but have no idea what is stopping that from happening.










      share|improve this question







      New contributor




      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.












      I recently reformatted my system and cannot get my RAID 1 array to load on boot. Would appreciate some help figuring out why this is. Using ubuntu 14 with MDADM. Its a very simple RAID1 array consisting of 2 drives sda and sdb. I tried to edit fstab to mount the RAID but it fails at boot and asks me to press a key to skip mounting.



      Here's my MDADM.conf file



      # mdadm.conf
      #
      # Please refer to mdadm.conf(5) for information about this file.
      #

      # by default (built-in), scan all partitions (/proc/partitions) and all
      # containers for MD superblocks. alternatively, specify devices to scan, using
      # wildcards if desired.
      #DEVICE partitions containers

      # auto-create devices with Debian standard permissions
      CREATE owner=root group=disk mode=0660 auto=yes

      # automatically tag new arrays as belonging to the local system
      HOMEHOST <system>

      # instruct the monitoring daemon where to send mail alerts
      MAILADDR root

      # definitions of existing MD arrays
      ARRAY metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
      ARRAY /dev/md/Storage(RAID) container=4e85cd11:34b3cd40:f263b2be:616ef7fb member=0 UUID=c26e7f0d:a956fc70:db00c387:58e7a99a

      # This file was auto-generated on Sun, 10 Mar 2019 16:31:50 -0500
      # by mkconf $Id$


      Output from mdadm --Examine /dev/sda:



      /dev/sda:
      Magic : Intel Raid ISM Cfg Sig.
      Version : 1.3.00
      Orig Family : 96f17bdd
      Family : 96f17bdd
      Generation : 000bd2b7
      Attributes : All supported
      UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
      Checksum : 7eee6b4f correct
      MPB Sectors : 1
      Disks : 2
      RAID Devices : 1

      Disk00 Serial : ZAD04VWC
      State : active
      Id : 00000004
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

      [Storage(RAID)]:
      UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
      RAID Level : 1
      Members : 2
      Slots : [UU]
      Failed disk : none
      This Slot : 0
      Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
      Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
      Sector Offset : 0
      Num Stripes : 45785312
      Chunk Size : 64 KiB
      Reserved : 0
      Migrate State : idle
      Map State : normal
      Dirty State : clean

      Disk01 Serial : ZAD05L4F
      State : active
      Id : 00000006
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


      Output from mdadm --Examine /dev/sdb:



      /dev/sdb:
      Magic : Intel Raid ISM Cfg Sig.
      Version : 1.3.00
      Orig Family : 96f17bdd
      Family : 96f17bdd
      Generation : 000bd2b7
      Attributes : All supported
      UUID : 4e85cd11:34b3cd40:f263b2be:616ef7fb
      Checksum : 7eee6b4f correct
      MPB Sectors : 1
      Disks : 2
      RAID Devices : 1

      Disk01 Serial : ZAD05L4F
      State : active
      Id : 00000006
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)

      [Storage(RAID)]:
      UUID : c26e7f0d:a956fc70:db00c387:58e7a99a
      RAID Level : 1
      Members : 2
      Slots : [UU]
      Failed disk : none
      This Slot : 1
      Array Size : 11721039872 (5589.03 GiB 6001.17 GB)
      Per Dev Size : 11721040136 (5589.03 GiB 6001.17 GB)
      Sector Offset : 0
      Num Stripes : 45785312
      Chunk Size : 64 KiB
      Reserved : 0
      Migrate State : idle
      Map State : normal
      Dirty State : clean

      Disk00 Serial : ZAD04VWC
      State : active
      Id : 00000004
      Usable Size : 11721040136 (5589.03 GiB 6001.17 GB)


      Output from lsblk -o NAME,SIZE,FSTYPE,TYPE,MOUNTPOINT gives the following:



      NAME SIZE FSTYPE TYPE MOUNTPOINT
      sda 5.5T disk
      └─md126 5.5T raid1
      ├─md126p1 128M md
      └─md126p2 5.5T md
      sdb 5.5T disk
      └─md126 5.5T raid1
      ├─md126p1 128M md
      └─md126p2 5.5T md
      sr0 1024M rom
      nvme0n1 477G disk
      ├─nvme0n1p1 499M part
      ├─nvme0n1p2 99M part /boot/efi
      ├─nvme0n1p3 16M part
      ├─nvme0n1p4 237.4G part
      ├─nvme0n1p5 470M part
      ├─nvme0n1p6 232.9G part /
      └─nvme0n1p9 1.9G part [SWAP]


      If I run sudo mdadm --verbose --assemble --scan after boot, here is the output



      mdadm: looking for devices for further assembly
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no RAID superblock on /dev/nvme0n1p9
      mdadm: no RAID superblock on /dev/nvme0n1p6
      mdadm: no RAID superblock on /dev/nvme0n1p5
      mdadm: no RAID superblock on /dev/nvme0n1p4
      mdadm: no RAID superblock on /dev/nvme0n1p3
      mdadm: no RAID superblock on /dev/nvme0n1p2
      mdadm: no RAID superblock on /dev/nvme0n1p1
      mdadm: no RAID superblock on /dev/nvme0n1
      mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot -1.
      mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot -1.
      mdadm: added /dev/sda to /dev/md/imsm0 as -1
      mdadm: added /dev/sdb to /dev/md/imsm0 as -1
      mdadm: Container /dev/md/imsm0 has been assembled with 2 drives
      mdadm: looking for devices for further assembly
      mdadm: looking for devices for further assembly
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no recogniseable superblock on /dev/nvme0n1p9
      mdadm: no recogniseable superblock on /dev/nvme0n1p6
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
      mdadm: no recogniseable superblock on /dev/nvme0n1p3
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
      mdadm: looking in container /dev/md127
      mdadm: found match on member /md127/0 in /dev/md127
      mdadm: Started /dev/md/Storage(RAID)_0 with 2 devices
      mdadm: looking for devices for further assembly
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: looking in container /dev/md127
      mdadm: member /md127/0 in /dev/md127 is already assembled
      mdadm: looking for devices for further assembly
      mdadm: Cannot assemble mbr metadata on /dev/md/Storage_RAID__0p2
      mdadm: no recogniseable superblock on /dev/md/Storage_RAID__0p1
      mdadm: Cannot assemble mbr metadata on /dev/md126
      mdadm: /dev/sdb is busy - skipping
      mdadm: /dev/sda is busy - skipping
      mdadm: cannot open device /dev/sr0: No medium found
      mdadm: no recogniseable superblock on /dev/nvme0n1p9
      mdadm: no recogniseable superblock on /dev/nvme0n1p6
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p5
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p4
      mdadm: no recogniseable superblock on /dev/nvme0n1p3
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p2
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1p1
      mdadm: Cannot assemble mbr metadata on /dev/nvme0n1
      mdadm: looking in container /dev/md127
      mdadm: member /md127/0 in /dev/md127 is already assembled


      I've made sure initramfs is updated when making changes to mdadm.conf. And once I run the --assemble --scan command, it does mount the arrays. If I then do cat /proc/mdstat I get:



      md127 : inactive sdb[1](S) sda[0](S)
      5032 blocks super external:imsm

      md126 : active (auto-read-only) raid1 sda[1] sdb[0]
      5860519936 blocks super external:/md127/0 [2/2] [UU]


      But if I try to do mdadm --detail --scan to add to the mdadm.conf file I get this:



      ARRAY /dev/md/imsm0 metadata=imsm UUID=4e85cd11:34b3cd40:f263b2be:616ef7fb
      mdadm: cannot open /dev/md/Storage(RAID)_0: No such file or directory


      I need this array to mount at boot but have no idea what is stopping that from happening.







      ubuntu raid mdadm






      share|improve this question







      New contributor




      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.











      share|improve this question







      New contributor




      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      share|improve this question




      share|improve this question






      New contributor




      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.









      asked 1 hour ago









      ccac3ccac3

      1




      1




      New contributor




      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.





      New contributor





      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.






      ccac3 is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
      Check out our Code of Conduct.




















          0






          active

          oldest

          votes











          Your Answer








          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "106"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );






          ccac3 is a new contributor. Be nice, and check out our Code of Conduct.









          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2funix.stackexchange.com%2fquestions%2f505540%2fmdadm-cant-automount-existing-raid1-array%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          0






          active

          oldest

          votes








          0






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes








          ccac3 is a new contributor. Be nice, and check out our Code of Conduct.









          draft saved

          draft discarded


















          ccac3 is a new contributor. Be nice, and check out our Code of Conduct.












          ccac3 is a new contributor. Be nice, and check out our Code of Conduct.











          ccac3 is a new contributor. Be nice, and check out our Code of Conduct.














          Thanks for contributing an answer to Unix & Linux Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid


          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.

          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2funix.stackexchange.com%2fquestions%2f505540%2fmdadm-cant-automount-existing-raid1-array%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          -mdadm, raid, ubuntu

          Popular posts from this blog

          Frič See also Navigation menuinternal link

          Identify plant with long narrow paired leaves and reddish stems Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?What is this plant with long sharp leaves? Is it a weed?What is this 3ft high, stalky plant, with mid sized narrow leaves?What is this young shrub with opposite ovate, crenate leaves and reddish stems?What is this plant with large broad serrated leaves?Identify this upright branching weed with long leaves and reddish stemsPlease help me identify this bulbous plant with long, broad leaves and white flowersWhat is this small annual with narrow gray/green leaves and rust colored daisy-type flowers?What is this chilli plant?Does anyone know what type of chilli plant this is?Help identify this plant

          fontconfig warning: “/etc/fonts/fonts.conf”, line 100: unknown “element blank” The 2019 Stack Overflow Developer Survey Results Are In“tar: unrecognized option --warning” during 'apt-get install'How to fix Fontconfig errorHow do I figure out which font file is chosen for a system generic font alias?Why are some apt-get-installed fonts being ignored by fc-list, xfontsel, etc?Reload settings in /etc/fonts/conf.dTaking 30 seconds longer to boot after upgrade from jessie to stretchHow to match multiple font names with a single <match> element?Adding a custom font to fontconfigRemoving fonts from fontconfig <match> resultsBroken fonts after upgrading Firefox ESR to latest Firefox