Sponsored Content
Top Forums UNIX for Advanced & Expert Users mdadm container! How does it work Post 302568992 by hytron on Saturday 29th of October 2011 12:42:58 AM
Old 10-29-2011
mdadm container! How does it work

Hi everyone,

I am not sure if I understand how mdadm --create /dev/md0 --level=container works?
A device called /dev/md0 appears in /proc/mdstat but I am not sure how to use that device?

I have 2 blank drives with 1 500GB partition on each. I would like to setup mirroring, but not in the way that you have to create partition of the size that you want mirror to be. I am not sure if container does that or not? I have looked all over the net and was unable to find any info.

If someone can give me more info on it, it would be greatly appreciated!
 

10 More Discussions You Might Find Interesting

1. Linux

mdadm - Swapping 500GB disks for 1TB

Hi, I have a three disk raid 5, with 500GB disks. This is close to being full, and whilst I can just add another disk and rebuild to add another 500GB, I would prefer to replace with 1TB disks. So i have some questions. Can I replace these disks one by one with bigger disks? I... (1 Reply)
Discussion started by: snoop2048
1 Replies

2. Virtualization and Cloud Computing

is mdadm --incremental --rebuild --run --scan destructive?

Hello Unix Community: My task to figure out how to add a 20G volume to an existing EBS Array (RAID0) at AWS. I haven't been told that growing the existing volumes isn't an option, or adding another larger volume to the existing array is the way to go. The client's existing data-store is... (0 Replies)
Discussion started by: Habitual
0 Replies

3. Solaris

solaris zones vs container..

kindly share what are difference b/w solaris zones and containers.... (8 Replies)
Discussion started by: Rajesh_Apple
8 Replies

4. Emergency UNIX and Linux Support

mdadm unable to fail a resyncing drive?

Hi All I have a RAID 5 array consisting of 4 drives that had a partial drive failure in one of the drives. Rebooting shows the faulty drive as background rebuilding and mdadm /dev/ARRAYID shows three drives as in sync with the fourth drive as spare rebuilding. However the array won't come... (9 Replies)
Discussion started by: Bashingaway
9 Replies

5. UNIX for Advanced & Expert Users

mdadm question

Hello, I have 4 drives (500G each) in a raid 10, I got a power failior and this is the result? cat /proc/mdstat Personalities : md126 : inactive sdb sdc sdd sde 1953536528 blocks super external:-md127/0 md127 : inactive sdd(S) sde(S) sdb(S) sdc(S) 9028 blocks super... (3 Replies)
Discussion started by: rmokros
3 Replies

6. Red Hat

mdadm for / and /boot

had this RHEL 5 installation with /dev/sda1 and /dev/sda2 running.. created two more partitions /dev/sdj1 and /dev/sdj2 , the same sized partition as /dev/sda trying to use mdadm to create RAID1 .. I cannot even do it in "rescue" mode, I wonder if it can be done.. it kept... (2 Replies)
Discussion started by: ppchu99
2 Replies

7. Solaris

Container

Hi, Can any one let me know in detail what is container and how it is different from zones. and also any process to create a container (6 Replies)
Discussion started by: chetansingh23
6 Replies

8. UNIX for Advanced & Expert Users

USB RAID 5 Problem on Joli OS 1.2 (Ubuntu) using mdadm

Hi All, I have been trying to create a USB RAID 5 using mdadm tool on Joli OS 1.2 (Ubuntu) but with no luck. I cannot even get pass the creation of array device (/dev/md0) and superblock. I am using 3 USB keys (2 16.4 GB kingston and 1 16GB sandisk). My steps are: ... (5 Replies)
Discussion started by: powelltallen
5 Replies

9. Filesystems, Disks and Memory

MDADM Failure - where it came from?

Hello, i have a system with 6 sata3 seagate st3000dm01 disks running on stable Debian with software raid mdadm. i have md0 for root and md1 for swap and md2 for the files. i now want to add one more disk = sdh4 for md2 but i got this errors: The new disk is connected to an 4 port sata... (7 Replies)
Discussion started by: Sunghost
7 Replies

10. UNIX for Advanced & Expert Users

How to fix mistake on raid: mdadm create instead of assemble?

Hi guys, I'm new to RAID although I've had a server running raid5 for a while. It was delivered preinstalled like this and I never really wondered how to monitor and maintain it. This quick introduction just to let you understand why I'm such an idiot asking such a silly question. Now what... (0 Replies)
Discussion started by: chebarbudo
0 Replies
CCDCONFIG(8)						    BSD System Manager's Manual 					      CCDCONFIG(8)

NAME
ccdconfig -- configuration utility for the concatenated disk driver SYNOPSIS
ccdconfig [-cv] ccd ileave [flags] dev ... ccdconfig -C [-v] [-f config_file] ccdconfig -u [-v] ccd ... ccdconfig -U [-v] [-f config_file] ccdconfig -g [ccd ...] DESCRIPTION
The ccdconfig utility is used to dynamically configure and unconfigure concatenated disk devices, or ccds. For more information about the ccd, see ccd(4). The options are as follows: -c Configure a ccd. This is the default behavior of ccdconfig. -C Configure all ccd devices listed in the ccd configuration file. -f config_file When configuring or unconfiguring all devices, read the file config_file instead of the default /etc/ccd.conf. -g Dump the current ccd configuration in a format suitable for use as the ccd configuration file. If no arguments are specified, every configured ccd is dumped. Otherwise, the configuration of each listed ccd is dumped. -u Unconfigure a ccd. -U Unconfigure all ccd devices listed the ccd configuration file. -v Cause ccdconfig to be verbose. A ccd is described on the command line and in the ccd configuration file by the name of the ccd, the interleave factor, the ccd configuration flags, and a list of one or more devices. The flags may be represented as a decimal number, a hexadecimal number, a comma-separated list of strings, or the word ``none''. The flags are as follows: CCDF_UNIFORM 0x02 Use uniform interleave CCDF_MIRROR 0x04 Support mirroring CCDF_NO_OFFSET 0x08 Do not use an offset CCDF_LINUX 0x0A Linux md(4) compatibility The format in the configuration file appears exactly as if it were entered on the command line. Note that on the command line and in the configuration file, the flags argument is optional. # # /etc/ccd.conf # Configuration file for concatenated disk devices # # ccd ileave flags component devices ccd0 16 none /dev/da2s1 /dev/da3s1 The component devices need to name partitions of type FS_BSDFFS (or ``4.2BSD'' as shown by disklabel(8)). If you want to use the Linux md(4) compatibility mode, please be sure to read the notes in ccd(4). FILES
/etc/ccd.conf default ccd configuration file EXAMPLES
A number of ccdconfig examples are shown below. The arguments passed to ccdconfig are exactly the same as you might place in the /etc/ccd.conf configuration file. The first example creates a 4-disk stripe out of four scsi disk partitions. The stripe uses a 64 sector interleave. The second example is an example of a complex stripe/mirror combination. It reads as a two disk stripe of da4 and da5 which is mirrored to a two disk stripe of da6 and da7. The last example is a simple mirror. The 2nd slice of /dev/da8 is mirrored with the 3rd slice of /dev/da9 and assigned to ccd0. # ccdconfig ccd0 64 none /dev/da0s1 /dev/da1s1 /dev/da2s1 /dev/da3s1 # ccdconfig ccd0 128 CCDF_MIRROR /dev/da4 /dev/da5 /dev/da6 /dev/da7 # ccdconfig ccd0 128 CCDF_MIRROR /dev/da8s2 /dev/da9s3 The following are matching commands in Linux and FreeBSD to create a RAID-0 in Linux and read it from FreeBSD. # Create a RAID-0 on Linux: mdadm --create --chunk=32 --level=0 --raid-devices=2 /dev/md0 /dev/hda1 /dev/hdb1 # Make the RAID-0 just created available on FreeBSD: ccdconfig -c /dev/ccd0 32 linux /dev/ada0s1 /dev/ada0s2 When you create a new ccd disk you generally want to fdisk(8) and disklabel(8) it before doing anything else. Once you create the initial label you can edit it, adding additional partitions. The label itself takes up the first 16 sectors of the ccd disk. If all you are doing is creating file systems with newfs, you do not have to worry about this as newfs will skip the label area. However, if you intend to dd(1) to or from a ccd partition it is usually a good idea to construct the partition such that it does not overlap the label area. For example, if you have A ccd disk with 10000 sectors you might create a 'd' partition with offset 16 and size 9984. # disklabel ccd0 > /tmp/disklabel.ccd0 # disklabel -Rr ccd0 /tmp/disklabel.ccd0 # disklabel -e ccd0 The disklabeling of a ccd disk is usually a one-time affair. If you reboot the machine and reconfigure the ccd disk, the disklabel you had created before will still be there and not require reinitialization. Beware that changing any ccd parameters: interleave, flags, or the device list making up the ccd disk, will usually destroy any prior data on that ccd disk. If this occurs it is usually a good idea to reini- tialize the label before [re]constructing your ccd disk. RECOVERY
An error on a ccd disk is usually unrecoverable unless you are using the mirroring option. But mirroring has its own perils: It assumes that both copies of the data at any given sector are the same. This holds true until a write error occurs or until you replace either side of the mirror. This is a poor-man's mirroring implementation. It works well enough that if you begin to get disk errors you should be able to backup the ccd disk, replace the broken hardware, and then regenerate the ccd disk. If you need more than this you should look into external hardware RAID SCSI boxes, RAID controllers (see GENERIC), or software RAID systems such as geom(8) and gvinum(8). SEE ALSO
dd(1), ccd(4), disklabel(8), fdisk(8), gvinum(8), rc(8) HISTORY
The ccdconfig utility first appeared in NetBSD 1.0A. BUGS
The initial disklabel returned by ccd(4) specifies only 3 partitions. One needs to change the number of partitions to 8 using ``disklabel -e'' to get the usual BSD expectations. BSD
October 1, 2013 BSD
All times are GMT -4. The time now is 08:17 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy