Sponsored Content
Operating Systems Solaris mirroring root disk using svm - but no free slices for metadb's Post 302305388 by badoshi on Wednesday 8th of April 2009 07:22:47 PM
Old 04-08-2009
mirroring root disk using svm - but no free slices for metadb's

Hi all,

we have an existing system that was configured using just one of the (two) internal disks. I want to mirror the disk using SVM, but have realised there is no free slice for creating the metadb's. Is there a workaround I can use for this?

In the past we have always kept slice 7 free - but this is not available:

Code:
AVAILABLE DISK SELECTIONS:
       0. c1t0d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424>
          /pci@1f,700000/scsi@2/sd@0,0
       1. c1t1d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424>
          /pci@1f,700000/scsi@2/sd@1,0
Specify disk (enter its number): ^D


Code:
# df -k
Filesystem            kbytes    used   avail capacity  Mounted on
/dev/dsk/c1t0d0s0    4133838 2560427 1532073    63%    /
/proc                      0       0       0     0%    /proc
mnttab                     0       0       0     0%    /etc/mnttab
fd                         0       0       0     0%    /dev/fd
/dev/dsk/c1t0d0s1    4133838  699047 3393453    18%    /var
swap                 11470176      48 11470128     1%    /var/run
swap                 11470128       0 11470128     0%    /tmp
/dev/dsk/c1t0d0s6    6198606  621245 5515375    11%    /opt
/dev/dsk/c1t0d0s3      46079    1041   40431     3%    /mdb
/dev/dsk/c1t0d0s5    5165838    5147 5109033     1%    /var/crash
/dev/dsk/c1t0d0s7    42645282 8677569 33541261    21%    /export/home

Code:
# cat /etc/vfstab
#device         device          mount           FS      fsck    mount   mount
#to mount       to fsck         point           type    pass    at boot options
#
fd      -       /dev/fd fd      -       no      -
/proc   -       /proc   proc    -       no      -
/dev/dsk/c1t0d0s4       -       -       swap    -       no      -
/dev/dsk/c1t0d0s0       /dev/rdsk/c1t0d0s0      /       ufs     1       no      -
/dev/dsk/c1t0d0s1       /dev/rdsk/c1t0d0s1      /var    ufs     1       no      -
/dev/dsk/c1t0d0s7       /dev/rdsk/c1t0d0s7      /export/home    ufs     2       yes     -
/dev/dsk/c1t0d0s3       /dev/rdsk/c1t0d0s3      /mdb    ufs     2       yes     -
/dev/dsk/c1t0d0s6       /dev/rdsk/c1t0d0s6      /opt    ufs     2       yes     -
/dev/dsk/c1t0d0s5       /dev/rdsk/c1t0d0s5      /var/crash      ufs     2       yes     -
swap    -       /tmp    tmpfs   -       yes     -

 

9 More Discussions You Might Find Interesting

1. Solaris

Solaris mirroring / non-root disk

Hi Guys, Need to add 2 disks into a JBOD array (3310). Does anyone see anything wrong with my Procedure / Doco below? 1> Logon to system, check system logs for abnormal entries. 2> Make backups of related system files: A>cp -p /etc/system /etc/system.backup.081505 B>cp -p /etc/vfstab... (3 Replies)
Discussion started by: BG_JrAdmin
3 Replies

2. Solaris

Veritas root disk mirroring

Hi there, My task is to replace the two 73 G disks with two 143 G disks , which has vxvm 4.1 running on it. I would like to know whether the steps iam following are correct. 1. Break the sub-disks, plexes of the root mirror. 2. Remove the sub-disks,plexes of the root mirror. 3. Remove one of... (10 Replies)
Discussion started by: Jartan
10 Replies

3. UNIX for Dummies Questions & Answers

Resize root disk slices

I have a Sun box running Solaris 9. My root disk was slices too small when it was installed and I am now at 99% capacity for my root partition. Is it possible, and if so how?, to increase the size of slice 0 and decrease the size of slice 7?? Thanks! Current partition table... (6 Replies)
Discussion started by: FredSmith
6 Replies

4. Solaris

SVM, mirroring

Really sorry for the long posting. But i would really want to clear all the doubts. I have 2 disk c0t0d0 & c0t1d0, i wanted to mirror c0t1d0 (mirror) to c0t0d0 (main). Creating state database replica: metadb -a -c3 -f c0t0d0s7 ... (3 Replies)
Discussion started by: beginningDBA
3 Replies

5. Solaris

Root Disk mirroring in SVM

Dear All, Please help me to configure root mirroring using SVM in Solaris 9. Thanks and Regards, Lakkireddy BR (3 Replies)
Discussion started by: lbreddy
3 Replies

6. Solaris

Disk mismatch while trying to zfs mirroring non-root disks

Hello All, I am trying to mirror two non-root hard drives using zfs. But "fmthard" fails when I try to copy the vtoc due to disk mismatch. Please help me. --- iostat command shows the disk to be similiar --- format command shows disk to be different :confused: --- c1t2d0 is the active... (8 Replies)
Discussion started by: pingmeback
8 Replies

7. Solaris

Root disk mirroring in SVM

I tried doing rootdisk mirroring in my local host , i added a new Ide disk in my system and copied the prtvtoc from root disk to the newly added disk, and then when i tried to add database replicas on both the disks, it was added for boot disk but for the newly added disk i gave the error, which... (6 Replies)
Discussion started by: Laxxi
6 Replies

8. Solaris

Root disk mirroring in Solaris 10

I would like to perform root disk mirroring task. Can someone please help me out on this. Thanks !! Regards, Rama (2 Replies)
Discussion started by: ramagore85
2 Replies

9. UNIX for Advanced & Expert Users

VCS triggerring panic on 1 node, root disk under SVM

We have two node cluster with OS disk mirrored under SVM. There is slight disk problem on one of the mirror disk causing cluster to panic. Failure of one mirror disk causing VCS to panic the node. Why VCS is not able to write /var filesystem, as one of the disk is healthy. ... (1 Reply)
Discussion started by: amlanroy
1 Replies
metadb(1M)						  System Administration Commands						metadb(1M)

NAME
metadb - create and delete replicas of the metadevice state database SYNOPSIS
/sbin/metadb -h /sbin/metadb [-s setname] /sbin/metadb [-s setname] -a [-f] [-k system-file] mddbnn /sbin/metadb [-s setname] -a [-f] [-k system-file] [-c number] [-l length] slice... /sbin/metadb [-s setname] -d [-f] [-k system-file] mddbnn /sbin/metadb [-s setname] -d [-f] [-k system-file] slice... /sbin/metadb [-s setname] -i /sbin/metadb [-s setname] -p [-k system-file] [mddb.cf-file] DESCRIPTION
The metadb command creates and deletes replicas of the metadevice state database. State database replicas can be created on dedicated slices, or on slices that will later become part of a simple metadevice (concatenation or stripe) or RAID5 metadevice. Do not place state database replicas on fabric-attached storage, SANs, or other storage that is not directly attached to the system and available at the same point in the boot process as traditional SCSI or IDE drives. See NOTES. The metadevice state database contains the configuration of all metadevices and hot spare pools in the system. Additionally, the metadevice state database keeps track of the current state of metadevices and hot spare pools, and their components. Solaris Volume Manager automati- cally updates the metadevice state database when a configuration or state change occurs. A submirror failure is an example of a state change. Creating a new metadevice is an example of a configuration change. The metadevice state database is actually a collection of multiple, replicated database copies. Each copy, referred to as a replica, is subject to strict consistency checking to ensure correctness. Replicated databases have an inherent problem in determining which database has valid and correct data. To solve this problem, Volume Man- ager uses a majority consensus algorithm. This algorithm requires that a majority of the database replicas be available before any of them are declared valid. This algorithm strongly encourages the presence of at least three initial replicas, which you create. A consensus can then be reached as long as at least two of the three replicas are available. If there is only one replica and the system crashes, it is possible that all metadevice configuration data can be lost. The majority consensus algorithm is conservative in the sense that it will fail if a majority consensus cannot be reached, even if one replica actually does contain the most up-to-date data. This approach guarantees that stale data will not be accidentally used, regardless of the failure scenario. The majority consensus algorithm accounts for the following: the system will stay running with exactly half or more replicas; the system will panic when less than half the replicas are available; the system will not reboot without one more than half the total replicas. When used with no options, the metadb command gives a short form of the status of the metadevice state database. Use metadb -i for an explanation of the flags field in the output. The initial state database is created using the metadb command with both the -a and -f options, followed by the slice where the replica is to reside. The -a option specifies that a replica (in this case, the initial) state database should be created. The -f option forces the creation to occur, even though a state database does not exist. (The -a and -f options should be used together only when no state databases exist.) Additional replicas beyond those initially created can be added to the system. They contain the same information as the existing replicas, and help to prevent the loss of the configuration information. Loss of the configuration makes operation of the metadevices impossible. To create additional replicas, use the metadb -a command, followed by the name of the new slice(s) where the replicas will reside. All repli- cas that are located on the same slice must be created at the same time. To delete all replicas that are located on the same slice, the metadb -d command is used, followed by the slice name. When used with the -i option, metadb displays the status of the metadevice state databases. The status can change if a hardware failure occurs or when state databases have been added or deleted. To fix a replica in an error state, delete the replica and add it back again. The metadevice state database (mddb) also contains a list of the replica locations for this set (local or shared diskset). The local set mddb can also contain host and drive information for each of the shared disksets of which this node is a member. Other than the diskset host and drive information stored in the local set mddb, the local and shared diskset mddbs are functionality identical. The mddbs are written to during the resync of a mirror or during a component failure or configuration change. A configuration change or failure can also occur on a single replica (removal of a mddb or a failed disk) and this causes the other replicas to be updated with this failure information. OPTIONS
Root privileges are required for all of the following options except -h and -i. The following options can be used with the metadb command. Not all the options are compatible on the same command line. Refer to the SYNOP- SIS to see the supported use of the options. -a Attach a new database device. The /kernel/drv/md.conf file is automatically updated with the new information and the /etc/lvm/mddb.cf file is updated as well. An alternate way to create replicas is by defining them in the /etc/lvm/md.tab file and specifying the assigned name at the command line in the form, mddbnn, where nn is a two-digit number given to the replica definitions. Refer to the md.tab(4) man page for instructions on setting up replicas in that file. -c number Specifies the number of replicas to be placed on each device. The default number of replicas is 1. -d Deletes all replicas that are located on the specified slice. The /kernel/drv/md.conf file is automatically updated with the new information and the /etc/lvm/mddb.cf file is updated as well. -f The -f option is used to create the initial state database. It is also used to force the deletion of replicas below the minimum of one. (The -a and -f options should be used together only when no state databases exist.) -h Displays a usage message. -i Inquire about the status of the replicas. The output of the -i option includes characters in front of the device name that represent the status of the state database. Explanations of the characters are displayed following the replica status and are as follows: d replica does not have an associated device ID. o replica active prior to last mddb configuration change u replica is up to date l locator for this replica was read successfully c replica's location was in /etc/lvm/mddb.cf p replica's location was patched in kernel m replica is master, this is replica selected as input r replica does not have device relocation information t tagged data is associated with the replica W replica has device write errors a replica is active, commits are occurring to this M replica had problem with master blocks D replica had problem with data blocks F replica had format problems S replica is too small to hold current database R replica had device read errors B tagged data associated with the replica is not valid -k system-file Specifies the name of the kernel file where the replica information should be written. The default system-file is /ker- nel/drv/md.conf. This option is for use with the local diskset only. -l length Specifies the size of each replica. The default length is 8192 blocks, which should be appropriate for most configurations. "Replica sizes of less than 128 blocks are not recommended. -p Specifies updating the system file (/kernel/drv/md.conf) with entries from the /etc/lvm/mddb.cf file. This option is nor- mally used to update a newly built system before it is booted for the first time. If the system has been built on a system other than the one where it will run, the location of the mddb.cf on the local machine can be passed as an argument. The system file to be updated can be changed using the -k option. This option is for use with the local diskset only. -s setname Specifies the name of the diskset on which the metadb command will work. Using the -s option will cause the command to per- form its administrative function within the specified diskset. Without this option, the command will perform its function on local database replicas. slice Specifies the logical name of the physical slice (partition), such as /dev/dsk/c0t0d0s3. EXAMPLES
Example 1: Creating Initial State Database Replicas The following example creates the initial state database replicas on a new system. # metadb -a -f c0t0d0s7 c0t1d0s3 c1t0d0s7 c1t1d0s3 The -a and -f options force the creation of the initial database and replicas. You could then create metadevices with these same slices, making efficient use of the system. Example 2: Adding Two Replicas on Two New Disks This example shows how to add two replicas on two new disks that have been connected to a system currently running Volume Manager. # metadb -a c0t2d0s3 c1t1d0s3 Example 3: Deleting Two Replicas This example shows how to delete two replicas from the system. Assume that replicas have been set up on /dev/dsk/c0t2d0s3 and /dev/dsk/c1t1d0s3. # metadb -d c0t2d0s3 c1t1d0s3 Although you can delete all replicas, you should never do so while metadevices still exist. Removing all replicas causes existing metade- vices to become inoperable. FILES
/etc/lvm/mddb.cf Contains the location of each copy of the metadevice state database. /etc/lvm/md.tab Workspace file for metadevice database configuration. /kernel/drv/md.conf Contains database replica information for all metadevices on a system. Also contains Solaris Volume Manager configuration information. EXIT STATUS
The following exit values are returned: 0 successful completion >0 an error occurred ATTRIBUTES
See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Availability |SUNWmdr | +-----------------------------+-----------------------------+ SEE ALSO
mdmonitord(1M), metaclear(1M), metadetach(1M), metahs(1M), metainit(1M), metaoffline(1M), metaonline(1M), metaparam(1M), metarecover(1M), metarename(1M), metareplace(1M), metaroot(1M), metaset(1M), metassist(1M), metastat(1M), metasync(1M), metattach(1M), md.tab(4), md.cf(4), mddb.cf(4), md.tab(4), attributes(5), md(7D) Solaris Volume Manager Administration Guide NOTES
Replicas cannot be stored on fabric-attached storage, SANs, or other storage that is not directly attached to the system. Replicas must be on storage that is available at the same point in the boot process as traditional SCSI or IDE drives. A replica can be stored on a: o Dedicated local disk partition o Local partition that will be part of a volume o Local partition that will be part of a UFS logging device SunOS 5.10 14 Sep 2004 metadb(1M)
All times are GMT -4. The time now is 10:27 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy