If i don't have raid disks can i shut down dmraid device-mapper?


 
Thread Tools Search this Thread
Operating Systems Linux If i don't have raid disks can i shut down dmraid device-mapper?
# 8  
Old 04-13-2011
Getting RAID to properly autodetect isn't trivial, which is probably why they decided to get that working for them if anyone needed it.
Login or Register to Ask a Question

Previous Thread | Next Thread

10 More Discussions You Might Find Interesting

1. Ubuntu

Md0 raid don't see my folders

I suddenly don't see my folders into /mnt/md0. What can be reason? mdadm --detail /dev/md* /dev/md0: Version : 1.2 Creation Time : Fri Jan 18 09:54:27 2019 Raid Level : raid1 Array Size : 1953383488 (1862.89 GiB 2000.26 GB) Used Dev Size : 1953383488 (1862.89 GiB... (8 Replies)
Discussion started by: tomislav91
8 Replies

2. UNIX for Advanced & Expert Users

Command to see the logical volume path, device mapper path and its corresponding dm device path

Currently I am using this laborious command lvdisplay | awk '/LV Path/ {p=$3} /LV Name/ {n=$3} /VG Name/ {v=$3} /Block device/ {d=$3; sub(".*:", "/dev/dm-", d); printf "%s\t%s\t%s\n", p, "/dev/mapper/"v"-"n, d}' Would like to know if there is any shorter method to get this mapping of... (2 Replies)
Discussion started by: royalibrahim
2 Replies

3. Solaris

Hardware RAID using three disks

Dear All , Pl find the below command , # raidctl -l Controller: 1 Volume:c1t0d0 Disk: 0.0.0 Disk: 0.1.0 Disk: 0.3.0 # raidctl -l c1t0d0 Volume Size Stripe Status Cache RAID Sub Size ... (10 Replies)
Discussion started by: jegaraman
10 Replies

4. Red Hat

Device-mapper behaviour booting on init=bin/bash

Good morning Recently we needed to change the password from a redhat 6.5 system that no one knew the root password. Starting the system with the init=/bin/bash method took us to the following scenario: system_vg active with only root_lv and tmpfs mounted. our entries at fstab are like... (1 Reply)
Discussion started by: Ikaro0
1 Replies

5. AIX

SCSI PCI - X RAID Controller card RAID 5 AIX Disks disappeared

Hello, I have a scsi pci x raid controller card on which I had created a disk array of 3 disks when I type lspv ; I used to see 3 physical disks ( two local disks and one raid 5 disk ) suddenly the raid 5 disk array disappeared ; so the hardware engineer thought the problem was with SCSI... (0 Replies)
Discussion started by: filosophizer
0 Replies

6. Solaris

Solaris not recognizing RAID 5 disks

I've just installed Sol 10 Update 9 on a Sun 4140 server and have a RAID 1 configuration (2 136 Gb drives) for the OS and have created a RAID 5 array (6 136 GB) drives. When i log into the system I am unable to see the RAID 5 disks at all. I've tried using the devfsadm command but no luck and... (9 Replies)
Discussion started by: goose25
9 Replies

7. Red Hat

Device Mapper Notations and LVM

Hi, I had a doubt regarding device mapper notations and their corresponding LVM volumes. I have configured a volume group with two logical volumes in it as root and swap. The entries in the /etc/fstab file show the dm notations namely, /dev/mapper/VolGroup00-LogVol01... (2 Replies)
Discussion started by: kanna_geekworkz
2 Replies

8. Red Hat

device-mapper-multipath path [undef]

I have an HP blade with Qlogic HBA's connected to an EVA8000. I have downloaded the latest multipath.conf from HP's website. The drive presented to the server appears to be configured and working except the output of "multipath -l" shows for all paths. What is causing this output? mpath0... (2 Replies)
Discussion started by: manzier
2 Replies

9. Solaris

Move disks to different StorEdge, keeping RAID

Hi. I need to move a 5 disk RAID5 array from a SE3310 box to a different SE3310 array. After installing the disks in the "new" StorEdge device, I "would like" ;) to be able have access to the data which is on the RAID. Essentially, the quesion is, how can this be done? :confused: I checked... (5 Replies)
Discussion started by: alexs77
5 Replies

10. Red Hat

IBM RAID disks

We have a Red Hat linux server running on IBM x445 hardware. There are external disks in an IBM EXP300 disk enclosure. The system is running RAID 5. One of the four IBM disks (73.4 GB 10k FRU 06P5760) has become faulty. The system is still up and running OK because of the RAID. In that same EXP300... (3 Replies)
Discussion started by: pdudley
3 Replies
Login or Register to Ask a Question
MKINITRD(8)						      System Manager's Manual						       MKINITRD(8)

NAME
mkinitrd - creates initial ramdisk images for preloading modules SYNOPSIS
mkinitrd [--version] [-v] [-f] [--preload=module] [--omit-scsi-modules] [--omit-raid-modules] [--omit-lvm-modules] [--with=module] [--image-version] [--fstab=fstab] [--nocompress] [--builtin=module] [--nopivot] image kernel-version DESCRIPTION
mkinitrd creates filesystem images which are suitable for use as Linux initial ramdisk (initrd) images. Such images are often used for preloading the block device modules (such as IDE, SCSI or RAID) which are needed to access the root filesystem. mkinitrd automatically loads filesystem modules (such as ext3 and jbd), IDE modules, all scsi_hostadapter entries in /etc/modules.conf, and raid modules if the system's root partition is on raid, which makes it simple to build and use kernels using modular device drivers. Any module options specified in /etc/modules.conf are passed to the modules as they are loaded by the initial ramdisk. If the root device is on a loop device (such as /dev/loop0), mkinitrd will build an initrd which sets up the loopback file properly. To do this, the fstab must contain a comment of the form: # LOOP0: /dev/hda1 vfat /linux/rootfs LOOP0 must be the name of the loop device which needs to be configured, in all capital lettes. The parameters after the colon are the device which contains the filesystem with the loopback image on it, the filesystem which is on the device, and the full path to the loop- back image. If the filesystem is modular, initrd will automatically add the filesystem's modules to the initrd image. The root filesystem used by the kernel is specified in the boot configuration file, as always. The traditional root=/dev/hda1 style device specification is allowed. If a label is used, as in root=LABEL=rootPart the initrd will search all available devices for an ext2 or ext3 filesystem with the appropriate label, and mount that device as the root filesystem. OPTIONS
--builtin=module Act as if module is built into the kernel being used. mkinitrd will not look for this module, and will not emit an error if it does not exist. This option may be used multiple times. -f Allows mkinitrd to overwrite an existing image file. --fstab=fstab Use fstab to automatically determine what type of filesystem the root device is on. Normally, /etc/fstab is used. --image-version The kernel version number is appended to the initrd image path before the image is created. --nocompress Normally the created initrd image is compressed with gzip. If this option is specified, the compression is skipped. --nopivot Do not use the pivot_root system call as part of the initrd. This lets mkinitrd build proper images for Linux 2.2 kernels at the expense of some features. In particular, some filesystems (such as ext3) will not work properly and filesystem options will not be used to mount root. This option is not recommended, and will be removed in future versions. --omit-lvm-modules Do not load any lvm modules, even if /etc/fstab expects them. --omit-raid-modules Do not load any raid modules, even if /etc/fstab and /etc/raidtab expect them. --omit-scsi-modules Do not load any scsi modules, including 'scsi_mod' and 'sd_mod' modules, even if they are present. --preload=module Load the module module in the initial ramdisk image. The module gets loaded before any SCSI modules which are specified in /etc/mod- ules.conf. This option may be used as many times as necessary. -v Prints out verbose information while creating the image (normally the mkinitrd runs silently). --version Prints the version of mkinitrd that's being used and then exits. --with=module Load the modules module in the initial ramdisk image. The module gets loaded after any SCSI modules which are specified in /etc/mod- ules.conf. This option may be used as many times as necessary. FILES
/dev/loop* A block loopback device is used to create the image, which makes this script useless on systems without block loopback support available. /etc/modules.conf Specified SCSI modules to be loaded and module options to be used. SEE ALSO
fstab(5), insmod(1), kerneld(8), lilo(8) AUTHOR
Erik Troan <ewt@redhat.com> 4th Berkeley Distribution Sat Mar 27 1999 MKINITRD(8)