Sponsored Content
Operating Systems Solaris Convert from raw disk to solaris volume manager disk Post 302625691 by fretagi on Wednesday 18th of April 2012 06:22:04 AM
Old 04-18-2012
Convert from raw disk to solaris volume manager disk

I have a solaris 10 system configured using NetApp as its storage, and the file systems are already configured as you can see from the example below:
Code:
root@moneta # df -h
Filesystem             size   used  avail capacity  Mounted on
/dev/md/dsk/d0         9.8G   513M   9.3G     6%    /
/devices                 0K     0K     0K     0%    /devices
ctfs                     0K     0K     0K     0%    /system/contract
proc                     0K     0K     0K     0%    /proc
mnttab                   0K     0K     0K     0%    /etc/mnttab
swap                    21G   1.7M    21G     1%    /etc/svc/volatile
objfs                    0K     0K     0K     0%    /system/object
sharefs                  0K     0K     0K     0%    /etc/dfs/sharetab
/dev/md/dsk/d6         9.8G   4.0G   5.7G    42%    /usr
fd                       0K     0K     0K     0%    /dev/fd
/dev/md/dsk/d1         9.8G   3.5G   6.2G    36%    /var
swap                    21G   208K    21G     1%    /tmp
swap                    21G    96K    21G     1%    /var/run
/dev/dsk/c4t60A98000646F6172636F677178564852d0s0
                       197G   100G    95G    52%    /ora_moneta_oraarch
/dev/md/dsk/d30        550G    52G   493G    10%    /local_backup
/dev/dsk/c4t60A98000646F6172636F677231526977d0s0
                        43G   753M    42G     2%    /moneta_polled01
/dev/dsk/c4t60A98000646F6172636F677234564655d0s0
                        64G    13G    51G    21%    /moneta_parsed01
/dev/dsk/c4t60A98000646F6172636F67724B4C2D6Dd0s6
                       689G   548G   134G    81%    /moneta_collected02
/dev/dsk/c4t60A98000646F6172636F677231507347d0s0
                        64G    53G    10G    84%    /moneta_temp01
/dev/md/dsk/d5         9.8G    69M   9.7G     1%    /opt
/dev/dsk/c4t60A98000646F6172636F677253506852d0s0
                       584G   513G    65G    89%    /ora_data01
/dev/md/dsk/d4         192G   151G    39G    80%    /internaldisk1
/dev/dsk/c4t60A98000646F6172636F67717542506Ad0s0
                       5.9G   2.4G   3.5G    41%    /moneta_home
/dev/dsk/c4t60A98000646F6172636F67724D6B7548d0s6
                       591G   462G   122G    80%    /moneta_collected03
/dev/dsk/c4t60A98000646F6172636F677255764435d0s0
                       583G   505G    72G    88%    /ora_data04
/dev/dsk/c4t60A98000646F6172636F677176744B35d0s0
                        20G   9.8G   9.7G    51%    /oracle
/dev/dsk/c4t60A98000646F6172636F6772557A4931d0s0
                       584G   522G    56G    91%    /ora_data03
/dev/dsk/c4t60A98000646F6172636F6772554A706Dd0s0
                       584G   521G    57G    91%    /ora_data02

As you can some of these file systems have exceed the 90% mark, so we want to extend them. From my research I need to convert the
Code:
/dev/dsk/c4t60A98000646F6172636F6772554A706Dd0s0

into “/dev/dsk/md/d60” as an example, which is “solaris volume manager” nomenclature. Then I can use growfs -m to resize the partition.
But I need first to convert those disks into solaris volume manager disks. Please can you give me the steps to do that?
FR
 

10 More Discussions You Might Find Interesting

1. Solaris

Disk Mirror in Solaris 9 via Solaris Volume Manager

Hello, I am trying to do mirror in solaris 9. I have total 0-7 disks 4 5 6 7 0 1 2 3 Drive 0 and Drive 4 = Boot Drives Need to Mirror following drives. Drive 1 and Drive 5 = Need to mirror Drive 1 was mounted on: /prod1, /prod2, /prod3, /prod4, /prod5. Then i... (3 Replies)
Discussion started by: deal732
3 Replies

2. Filesystems, Disks and Memory

Volume Manager; importing a disk

System: Alpha with Tru64 5.1b Disk under LSM (Logical Storage Manager; essentially v2 of Veritas VxVM) control was generating disk errors. The disk was timing out a lot and generating a few disk errors. DBAs couldn't keep the oracle instance up on that node of the cluster. I contacted HP and got... (1 Reply)
Discussion started by: BOFH
1 Replies

3. Solaris

Reading raw disk on Solaris

Hello I wonder if someone could help me in reading a raw (non-Solaris) disk on a Solaris system... I have an IDE HDD in my Sun Blade and would like to read it (using C). It appears on the system and with the format command shows up as c0t1d0. I use the dd command to read the disk as such:... (19 Replies)
Discussion started by: son_t
19 Replies

4. Solaris

How to create new partitions in solaris,from the raw disk?

Hi all, I would like to know how to make new partitions.... I currently have allocated 60G for various slices (I have totally used 4 out of 7 available slices... I am running only solaris on my box. My plan is to have entire disk dedicated to solaris and run other OS from within... (19 Replies)
Discussion started by: wrapster
19 Replies

5. UNIX for Advanced & Expert Users

Veritas Volume Manager question (Disk layout with 4 plexes)

I am trying to build a veritas volume similar to an existing volume on another server. The output on source server is: usbtor12# vxprint -hrtg appdg v anvil_sqlVOL - ENABLED ACTIVE 629145600 SELECT - fsgen pl anvil_sqlVOL-01 anvil_sqlVOL ENABLED ACTIVE 629145600... (3 Replies)
Discussion started by: momin313
3 Replies

6. Solaris

Veritas Volume Manager: disk "failed was"

Hello there, I'm going to describe a situation I've got here... feel free to ask away questions and I'll provide what I can if it'll help me get this answered! When I do a vxdisk list, I see a disk that VxVM calls "disk4" that is listed as "failed was: c1t9d0s2". When I do a format, I can go... (3 Replies)
Discussion started by: kitykity
3 Replies

7. UNIX for Dummies Questions & Answers

VERITAS Volume Manager - mirror a disk/volume

I have a machine (5.10 Generic_142900-03 sun4u sparc SUNW,Sun-Fire-V210) that we are upgrading the storage and my task is to mirror what is already on the machine to the new disk. I have the disk, it is labeled and ready but I am not sure of the next steps to mirror the existing diskgroup and... (1 Reply)
Discussion started by: rookieuxixsa
1 Replies

8. Solaris

root disk mirroring in solaris volume manager for solaris 10

Need a procedure document to do "root disk mirroring in solaris volume manager for solaris 10". I hope some one will help me asap. I need to do it production environment. Let me know if you need any deatils on this. Thanks, Rama (1 Reply)
Discussion started by: ramareddi16
1 Replies

9. AIX

Regarding AIX volume manager & replacing a disk

First a little background: I'm working with an AIX 6.1 TL05 running two mirrored SAS disks (rootvg) and four SSDs (appvg) All four SSDs belong to appvg and are setup to mirror as follows: hdisk4 --> hdisk6 (containing application fs) hdisk5 --> hdisk7 (containing database fs) A few days... (1 Reply)
Discussion started by: Michael Mullig
1 Replies

10. Solaris

Convert from raw disk to solaris volume manager disk

I have a solaris 10 system configured using NetApp as its storage, and the file systems are already configured as you can see from the example below: root@moneta # df -h Filesystem size used avail capacity Mounted on /dev/md/dsk/d0 9.8G 513M 9.3G 6% / ... (0 Replies)
Discussion started by: fretagi
0 Replies
cmdisklock(1m)															    cmdisklock(1m)

NAME
cmdisklock - manage Serviceguard cluster lock devices. SYNOPSIS
cmdisklock check path cmdisklock [-f] reset path DESCRIPTION
cmdisklock is a tool to check the current state of a Serviceguard cluster lock device. It can also be used to reset the state of the clus- ter lock device. The need to reset the cluster lock device state could arise if the cluster lock device is replaced or becomes corrupt. A cluster lock device can be either an HP-UX LVM cluster lock or a cluster lock LUN device. HP-UX LVM cluster locks exist only on a disk in an LVM volume group. Cluster lock LUNs exist only on disks dedicated to cluster lock. cmdisklock is useful for checking either type of cluster lock and for re-initializing cluster lock LUN devices after a failure or corruption. NOTE To restore an HP-UX LVM cluster lock, use vgcfgrestore. cmdisklock will fail until vgcfgrestore is run, and cmdisklock is unnecessary as long as vgcfgbackup was done after the cluster lock was initialized. See the Managing Serviceguard manual for details. The syntax of the path option depends on the type of lock. For HP-UX LVM cluster lock disks, the syntax is VG:PV (for example: /dev/vglock:/dev/dsk/c0t0d2). For cluster lock LUN disks, the path is the disk device path. For example, /dev/sdd1 (on Linux) or /dev/dsk/c0t1d2 (on HP-UX). Options cmdisklock supports the following options: check Check the current state of the cluster lock device and report the results. reset Reset (initialize) the state of the cluster lock device. This operation should only be performed on a cluster lock LUN device. For HP-UX LVM cluster lock, use vgcfgrestore as documented in the Managing Serviceguard manual. After performing a reset, a check can be used to verify that the lock is cleared. EXAMPLES
If the cluster lock LUN device becomes corrupted and the cluster is up, messages like the following will appear in syslog. Mar 15 12:20:41 usb cmdisklockd[17599]: WARNING: Cluster lock LUN /dev/dsk/c0t1d2 is corrupt: bad label. Until this situation is cor- rected, a single failure could cause all nodes in the cluster to crash. Mar 15 12:20:41 usb cmdisklockd[17599]: After ensuring that all active nodes in the cluster have logged this message, run 'cmdisklock reset /dev/dsk/c0t1d2' to repair Mar 15 12:20:41 usb cmdisklockd[17599]: Cluster lock disk /dev/dsk/c0t1d2 is inaccessible Once the above messages appear in syslog on all running nodes, the following command will re-initialize the cluster lock LUN: ucd:/> cmdisklock reset /dev/dsk/c0t1d2 WARNING: Cluster lock LUN /dev/dsk/c0t1d2 is corrupt: bad label. Until this situation is corrected, a single failure could cause all nodes in the cluster to crash. After ensuring that all active nodes in the cluster have logged this message, run 'cmdisklock reset /dev/dsk/c0t1d2' to repair /dev/dsk/c0t1d2 is inaccessible Resetting cluster lock device /dev/dsk/c0t1d2 Cluster lock reset completed /dev/dsk/c0t1d2 is accessible cleared After the lock is restored, a message like the following appears in syslog: Mar 15 12:23:11 usb cmdisklockd[17599]: Cluster lock disk /dev/dsk/c0t1d2 is accessible WARNINGS
CAUTION For cluster lock LUN, reset is a potentially destructive operation. While cmdisklock checks for known volume manager and file system use (overridden by -f), it does not validate that the device to be reset is actually used by any cluster. If -f is used on the wrong device file, loss of data may result. CAUTION Care should be taken when doing a reset when the cluster is active as there is a remote possibility that the cluster will partition right when this command is run and both nodes could end up thinking they have successfully acquired the lock. To avoid this situation, make sure cmcld has logged a message in syslog on all running nodes saying the device is inaccessble, before performing a reset. Note that it is safe to run cmdisklock when the cluster is down. RETURN VALUE
cmdisklock returns the following values: 0 Successful completion. 1 The disk is inaccessible or is not recognized as a cluster lock. AUTHOR
cmdisklock was developed by HP. SEE ALSO
cmapplyconf(1m), cmviewcl(1m), vgcfgbackup(1m), vgcfgrestore(1m) Requires Optional Serviceguard Software cmdisklock(1m)
All times are GMT -4. The time now is 12:38 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy