Sponsored Content
Operating Systems Linux Red Hat missing raid array after reboot Post 302528830 by sriniv666 on Wednesday 8th of June 2011 04:49:48 AM
Old 06-08-2011
missing raid array after reboot

Dear all ,

i ve configured raid 0 in redhat machine(VM ware), by following steps:

Code:
#mdadm -C /dev/md0 -l 0 -n 2 /dev/sdb1 /dev/sdc1 
#mkfs.ext3 /dev/md0 
#mdadm --detail --scan --config=mdadm.conf >/etc/mdadm.conf

then
mounted the/dev/md0 device and also added a entry in fstab.

how the isssue is when the system rebooted the raid array configuration is missing, there is no device as /dev/md0..
what's the solution for this??? any help !!!!

Last edited by pludi; 06-08-2011 at 07:06 AM.. Reason: content editing...
 

9 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

Percent complete error while scanning RAID array during 5.0.6 load

Percent complete SCO 5.0.6 / No longer an issue (0 Replies)
Discussion started by: Henrys
0 Replies

2. AIX

RAID level of array = false?

I created a RAID 5 array and when I list out the attributes of the "hdisk" it reports back raid_level = 5 but the RAID level of the array = false. What does this actually indicate about my array? I've never paid much attention to this until now since I have a disk reporting failure I want to make... (0 Replies)
Discussion started by: scottsl
0 Replies

3. UNIX for Advanced & Expert Users

Create RAID - Smart Array Tool - ML370

Hi guys, i must install an old old old ml370 server... I must create a RAID 5 with my 4 SCSI disk. I need a SmartStart disk for create it or a Floppy Disk called "Array configuration Tool". I don't find it on the hp website...:mad::mad::mad: Anyone have it?? Thanks in advance. Zio (0 Replies)
Discussion started by: Zio Bill
0 Replies

4. Solaris

EFI Disk labels on 3510 raid array

Hi Peeps, Can anyone help me an EFI lablel on a 3510 raid array that I cannot get rid of, format -e and label just asks you if you want to label it. Want an SMI label writing to it. Anyone got any ideas on how to remove the EFI label? Thanks in advance Martin (2 Replies)
Discussion started by: callmebob
2 Replies

5. Emergency UNIX and Linux Support

Loading a RAID array after OS crash

One of my very old drive farm servers had an OS fault and can't boot now but I'd like to restore some files from it. I tried booting Ubuntu from a CD, but it couldn't see the drives -- possibly because they're RAIDed together. Is there a good way to get at my files? The data in question is a... (2 Replies)
Discussion started by: CRGreathouse
2 Replies

6. Solaris

How to find missing disks on Sun x4150 without reboot?

Hi, Here is the issue: There are 4 disks on this Sun x4150 system under Solaris 10, but only 1 disk can be seen by the OS. I've tried commands disks and devfsadm but not working. It's an important production server, so 'reboot -r' is not a choice. # format < /dev/null Searching for... (6 Replies)
Discussion started by: aixlover
6 Replies

7. Fedora

Missing entries in log files just before/after reboot

Hello world, One of the servers, a Fedora one,rebooted today (Luckily, a testbox). I tried to get the reason the server rebooted. After going through the messages, I think that the log entries just before and after reboot are missing. Please below: (****** is the server name, for privacy... (0 Replies)
Discussion started by: satish51392111
0 Replies

8. Solaris

Solaris 10 Installation - Disks missing, and Raid

Hey everyone. First, let me start by saying I'm primarily focused on linux boxes, and just happened to get pulled into building two T5220's. I'm not super educated on sun boxes. Both T5220's have 8 146GB 15k SAS drives. Inside the service processor, I can run SHOW /SYS/HDD{0-7} and they all come... (2 Replies)
Discussion started by: msarro
2 Replies

9. UNIX for Advanced & Expert Users

Revive RAID 0 Array From Buffalo Duo NAS

Thank you in advanced, I had a Buffalo DUO crap out on me that was setup as RAID 0. I dont belive it was the drives but rather the controller in the DUO unit. I bought another external HDD enclosure and was able to fireup the two older DUO drives in it and I think I resembled the RAID... (12 Replies)
Discussion started by: metallica1973
12 Replies
MOUNT.CONF(8)						    BSD System Manager's Manual 					     MOUNT.CONF(8)

NAME
mount.conf -- root file system mount configuration file SYNOPSIS
/.mount.conf DESCRIPTION
During the bootup process, the FreeBSD kernel will try to mount the root file system using the logic in the vfs_mountroot() function in src/sys/kern/vfs_mountroot.c. The root mount logic can be described as follows: 1. The kernel will synthesize in memory a config file with default directives for mounting the root file system. The logic for this is in vfs_mountroot_conf0(). 2. The kernel will first mount devfs(8) as the root file system. 3. Next, the kernel will parse the in-memory config file created in step 1 and try to mount the actual root file system. See FILE FORMAT for the format of the config file. 4. When the actual root file system is mounted, devfs will be re-mounted on the /dev directory. 5. If a /.mount.conf file does not exist in the root file system which was just mounted, the root mount logic stops here. 6. If a /.mount.conf file exists in the root file system which was just mounted, this file will be parsed, and the kernel will use this new config file to try to re-mount the root file system. See FILE FORMAT for the format of the config file. 7. If the new root file system has a /.mount directory, the old root file system will be re-mounted on /.mount. 8. The root mount logic will go back to step 4. The root mount logic is recursive, and step 8 will be repeated as long as each new root file system which is mounted has a /.mount.conf file. FILE FORMAT
The kernel parses each line in .mount.conf and then tries to perform the action specified on that line as soon as it is parsed. # A line beginning with a # is a comment and is ignored. {FS}:{MOUNTPOINT} {OPTIONS} The kernel will try to mount this in an operation equivalent to: mount -t {FS} -o {OPTIONS} {MOUNTPOINT} / If this is successfully mounted, further lines in .mount.conf are ignored. If all lines in .mount.conf have been processed and no root file system has been successfully mounted, then the action specified by .onfail is performed. .ask When the kernel processes this line, a mountroot> command-line prompt is displayed. At this prompt, the operator can enter the the root mount. .md file Create a memory backed md(4) virtual disk, using file as the backing store. .onfail [panic|reboot|retry|continue] If after parsing all the lines in .mount.conf the kernel is unable to mount a root file system, the .onfail directive tells the kernel what action to perform. .timeout N Before trying to mount a root file system, if the root mount device does not exist, wait at most N seconds for the device to appear before trying to mount it. If .timeout is not specified, the default timeout is 3 seconds. EXAMPLES
The following example .mount.conf will direct the kernel to try mounting the root file system first as an ISO CD9660 file system on /dev/cd0, then if that does not work, as an ISO CD9660 file system on /dev/acd0, and then if that does not work, as a UFS file system on /dev/ada0s1a. If that does not work, a mountroot> command-line prompt will be displayed where the operator can manually enter the root file system to mount. Finally if that does not work, the kernel will panic. .onfail panic .timeout 3 cd9660:/dev/cd0 ro .timeout 0 cd9660:/dev/acd0 ro .timeout 3 ufs:/dev/ada0s1a .ask The following example .mount.conf will direct the kernel to create a md(4) memory disk attached to the file /data/OS-1.0.iso and then mount the ISO CD9660 file system on the md device which was just created. The last line is a comment which is ignored. .timeout 3 .md /data/OS-1.0.iso cd9600:/dev/md# ro # Can also use cd9660:/dev/md0 ro The following example .mount.conf will direct the kernel to create a md(4) memory disk attached to the file /data/base.ufs.uzip and then mount the UFS file system on the md uzip device which was just created by the geom_uzip(4) driver. .md /data/base.ufs.uzip ufs:/dev/md#.uzip ro # Can also use ufs:/dev/md0.uzip ro The following example .mount.conf will direct the kernel to do a unionfs mount on a directory /jail/freebsd-8-stable which has a chroot(2) environment. .timeout 3 unionfs:/jail/freebsd-8-stable NOTES
For each root file system which is mounted, a /dev directory must exist so that the root mount logic can properly re-mount devfs(8). If this directory does not exist, the system may hang during the bootup process. SEE ALSO
nmount(2), md(4), boot.config(5), fstab(5), boot(8), loader(8), mount(8) HISTORY
The mount.conf file first appeared in FreeBSD 9.0. AUTHORS
The root mount logic in the FreeBSD kernel which parses /.mount.conf was written by Marcel Moolenaar <marcel@FreeBSD.org>. This man page was written by Craig Rodrigues <rodrigc@FreeBSD.org>. BSD
July 7, 2013 BSD
All times are GMT -4. The time now is 05:36 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy