Sponsored Content
Full Discussion: Solaris Volume manager
Operating Systems Solaris Solaris Volume manager Post 302435789 by DukeNuke2 on Thursday 8th of July 2010 07:53:53 AM
Old 07-08-2010
this is (afaik) not possible without a new set of disks which are in raid5 configuration. so you need the raid1 and the raid5 to copy the data.
 

10 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

newfs hotspares: solaris volume manager

I'm running sun volume manager on solaris 9. I have two hotspares and are currently on standby. Both are not being utilized. Can I newfs both of them? Do I need to deleted the hostpares first, then newfs? hsp002: 2 hot spares Device Status Length Reloc ... (3 Replies)
Discussion started by: xnightcrawl
3 Replies

2. Solaris

Disk Mirror in Solaris 9 via Solaris Volume Manager

Hello, I am trying to do mirror in solaris 9. I have total 0-7 disks 4 5 6 7 0 1 2 3 Drive 0 and Drive 4 = Boot Drives Need to Mirror following drives. Drive 1 and Drive 5 = Need to mirror Drive 1 was mounted on: /prod1, /prod2, /prod3, /prod4, /prod5. Then i... (3 Replies)
Discussion started by: deal732
3 Replies

3. Solaris

How to resize mirror volume in veritas volume manager 3.5 on Solaris 9 OE

Hi all, I have a problem with vxvm volume which is mirror with two disks. when i am try to increase file system, it is throwing an ERROR: can not allocate 5083938 blocks, ERROR: can not able to run vxassist on this volume. Please find a sutable solutions. Thanks and Regards B. Nageswar... (0 Replies)
Discussion started by: nageswarb
0 Replies

4. Solaris

Need some E-Labs for solaris 10 & veritas volume manager

Hello friends, I need to test some of the solaris 10 concepts and veritas volume manager.i want to know ,where i get the testing labs or E-labs(online) for this practice.i ready to pay for this labs.pls kindly help me to get the details or website names...Thanks in Advance -Jay (6 Replies)
Discussion started by: rjay.com
6 Replies

5. Solaris

Check free space in Solaris Volume Manager

I 've 300Gb Lun assigned to my system which has soft partitions , is there any way to check how much the space i have used out of 300Gb ? and how may more soft partitions i can create and expand the existing ones ? (3 Replies)
Discussion started by: fugitive
3 Replies

6. Solaris

What is the command to see the veritas volume manager 3.0.4 license..in solaris 8?

what is the command to see the veritas volume manager 3.0.4 license..in solaris 8.I wnat to see the license number (3 Replies)
Discussion started by: tv.praveenkumar
3 Replies

7. Solaris

Veritas volume manager in solaris.

Can you please let me know the certification code for veritas volume manager in solaris ? Thanks in advance. (2 Replies)
Discussion started by: gowthamakanthan
2 Replies

8. Solaris

Solaris Volume Manager

Hello All, I have small doubt. It's related to SVM in solaris 10. I have created raid 0 (striping) using 3 slices of 500 MB size (default interlace value as 32KB) d2: Concat/Stripe ---- Interlace value is 32 KB Size: 2923830 blocks (1.4 GB) Stripe 0: (interlace: 64 blocks) Device Start... (1 Reply)
Discussion started by: vaibhav.kanchan
1 Replies

9. Solaris

root disk mirroring in solaris volume manager for solaris 10

Need a procedure document to do "root disk mirroring in solaris volume manager for solaris 10". I hope some one will help me asap. I need to do it production environment. Let me know if you need any deatils on this. Thanks, Rama (1 Reply)
Discussion started by: ramareddi16
1 Replies

10. Solaris

solaris volume manager- stripe?

Hello Admins.. I am going through solaris volume manager guide for RAID-0 concatenation and stripes, I do not understand the concept of stripe from following example of concatenation. There is an eample for concatenation: # metainit d25 1 1 c0t1d0s2 d25: Concat/Stripe is setup the... (5 Replies)
Discussion started by: snchaudhari2
5 Replies
volinfo(8)						      System Manager's Manual							volinfo(8)

NAME
volinfo - Print accessibility and usability of volumes SYNOPSIS
/usr/sbin/volinfo [-Vp] [-g diskgroup] [-U usetype] [-o useopt] [volume...] OPTIONS
The following options are recognized: Writes a list of utilities that would be called from volinfo, along with the arguments that would be passed. The -V performs a ``mock run'' so the utilities are not actually called. Reports the name and condition of each plex in each reported volume. Specifies the usage type for the operation. If no volume operands are specified, the output is restricted to volumes with this usage type. If volume operands are specified, this will result in a failure message for all named volumes that do not have the indi- cated usage type. Specifies the disk group for the operation, either by disk group ID or by disk group name. By default, the disk group is chosen based on the volume operands. If no volume operands are specified, the disk group defaults to rootdg. Passes in usage-type-specific options to the operation. This option is currently unsupported. DESCRIPTION
The volinfo utility reports a usage-type-dependent condition on one or more volumes in a disk group. A report for each volume specified by the volume operand is written to the standard output. If no volume operands are given, a volume condition report is provided for each vol- ume in the selected disk group. Each invocation can be applied to only one disk group at a time, due to internal implementation constraints. Any volume operands will be used to determine a default disk group, according to the standard disk group selection rules described in volintro(8). A specific disk group can be forced with -g diskgroup. Output Format Summary reports for each volume are printed in one-line output records. Each volume output line consists of blank-separated fields for the volume name, volume usage type, and volume condition. Each plex output line consists of blank-separated fields for the plex name and the plex condition. The following example shows the volume summary: # volinfo bigvol fsgen Startable vol2 fsgen Started brokenvol gen Unstartable The following example shows the plex summary, with the plex records accompanied by their volume records: # volinfo -p vol bigvol fsgen Startable plex bigvol-01 ACTIVE vol vol2 fsgen Started plex vol2-01 ACTIVE vol brokenvol gen Unstartable Volume Conditions The volume condition is a usage-type-dependent summary of the state of a volume. This condition is derived from the volume's kernel-enabled state and the usage-type-dependent states of the volume's plexes. Volume conditions for the fsgen and gen usage types are reported as follows: The volume is not enabled and at least one of the plexes has a reported condition of ACTIVE or CLEAN. A volume startall operation would likely succeed in starting a volume in this condition. The vol- ume is not enabled and fails to meet the criteria for being Startable. A volume in this condition is not started and may be configured incorrectly or prevented from automatic startup (with volume startall) because of errors or other conditions. The volume is enabled and at least one of the associated plexes is enabled in read-write mode (which is normal for enabled plexes in the ACTIVE and EMTPY conditions). A volume in this condition has been started and can be used. The volume is enabled, but the volume does not meet the criteria for being Started. A volume in this condition has been started, but is inaccessible because of errors that have occurred since the volume was started, or because of administrative actions, such as voldg -k rmdisk. Volume conditions for volumes of the raid5 usage type include the following conditions used for the fsgen and gen usage types: Startable, Unstartable, Started, Started Unusable Additional volume conditions for raid5 volumes are: The RAID-5 plex of the volume is in degraded mode due to the unavailability of a sub- disk in that plex. Some of the parity in the RAID-5 plex is stale and requires recovery. Plex Conditions The following plex conditions (reported with -p) are reported for the fsgen and gen usage types: No physical disk was found for one of the subdisks in the plex. This implies either that the physical disk failed, making it unrecognizable, or that the physical disk is no longer attached through a known access path. A physical disk used by one of the subdisks in the plex was removed through administrative action with voldg -k rmdisk. The plex was detached from use as a result of an uncorrectable I/O failure on one of the subdisks in the plex. The plex does not contain valid data, either as a result of a disk replacement affecting one of the subdisks in the plex, or as a result of an administrative action on the plex such as volplex det. The plex contains valid data and the volume was stopped cleanly. Either the volume is started and the plex is enabled, or the volume was not stopped cleanly and the plex was valid when the volume was stopped. The plex was disabled using the volmend off operation. The plex is part of a volume that has not yet been initialized. The plex is associated tempo- rarily as part of a current operation, such as volplex cp or volplex att. A system reboot or manual starting of a volume will dissociate the plex. The plex was created for temporary use by a current operation. A system reboot or manual starting of a volume will remove the plex. The plex and its subdisks were created for temporary use by a current operation. A system reboot or manual starting of the volume will remove the plex and all of its subdisks. The plex is being attached as part of a backup operation by the volassist snapstart opera- tion. When the attach is complete, the condition will change to SNAPDONE. A system reboot or manual starting of the volume will remove the plex and all of its subdisks. A volassist snapstart operation completed the process of attaching the plex. It is a candidate for selection by the volassist snapshot operation. A system reboot or manual starting of the volume will remove the plex and all of its subdisks. The plex is being attached as part of a backup operation by the volplex snapstart operation. When the attach is complete, the condition will change to SNAPDIS. A system reboot or manual starting of the volume will dissociate the plex. A volassist snapstart operation completed the process of attaching the plex. It is a candidate for selection by the volplex snapshot operation. A system reboot or manual starting of the volume will dissociate the plex. Plexes of raid5 volumes can be either data plexes (that is, RAID-5 plexes) or log plexes. Plex conditions for RAID-5 plexes and log plexes include the following conditions used for the fsgen and gen usage types: NODAREC, REMOVED, IOFAIL, CLEAN, ACTIVE, OFFLINE RAID-5 plexes can have these additional conditions: Due to subdisk failures, the plex is in degraded mode. This indicates a loss of data redundancy in the RAID-5 volume and any further failures could cause data loss. The parity is not in sync with the data in the plex. This indicates a loss of data redundancy in the RAID-5 volume and any further failures could case data loss. A double failure occurred within the plex. The plex is unusable due to subdisk failures and/or stale parity. Log plexes of RAID-5 volumes can have this additional condition: The contents of the plex are not usable as logging data. EXIT CODES
The volinfo utility exits with a nonzero status if the attempted operation fails. A nonzero exit code is not a complete indicator of the problems encountered, but rather denotes the first condition that prevented further execution of the utility. See volintro(8) for a list of standard exit codes. SEE ALSO
volintro(8), volassist(8), volmend(8), volplex(8), volsd(8), volume(8) volinfo(8)
All times are GMT -4. The time now is 08:02 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy