Sponsored Content
Top Forums UNIX for Advanced & Expert Users live upgrade with raid0 soft partitions Post 302719023 by chaandana on Monday 22nd of October 2012 02:03:54 AM
Old 10-22-2012
live upgrade with raid0 soft partitions

Hi,

I have this mirrored system with soft-partitions.
I have a difficulty determining the lucreate cmd in this env.
Code:
#metastat -p
d0 -m d10 d20 1   
d10 1 1 c1t2d0s0
d20 1 1 c1t3d0s0
d1 -m d11 d21 1   
d11 1 1 c1t2d0s5
d21 1 1 c1t3d0s5
d100 -p d1 -o 58720384 -b 8388608   
d200 -p d1 -o 32 -b 41943040

#cat /etc/vfstab
Code:
/dev/md/dsk/d0  /dev/md/rdsk/d0 /       ufs     1       no      -
/dev/md/dsk/d100        /dev/md/rdsk/d100       /var    ufs    2   yes   logging
/dev/md/dsk/d200        /dev/md/rdsk/d200       /opt    ufs    2   yes   logging

And this is the lucreate cmd I'm thinking of:
Code:
lucreate -c old -n new -m /:/dev/md/dsk/d110:ufs,mirror
-m /:/dev/dsk/c1t3d0s0:detach,attach,preserve
-m /var:/dev/md/dsk/d111:ufs,mirror
-m /var:/dev/dsk/c1t3d0s5:detach,attach,preserve
-m /opt:/dev/md/dsk/d112:ufs,mirror
-m /opt:/dev/dsk/c1t3d0s5:detach,attach,preserve

Would this work ?

Thanks
Dahasak.

Last edited by Scrutinizer; 10-22-2012 at 03:35 AM.. Reason: code tags
 

8 More Discussions You Might Find Interesting

1. Solaris

Live upgrade Issue

Hi, I upgraded solaris 10 x86 from update 3 to update 7 with zones installed n UFS file system . The global zone was updated but the non global zone still shows update 3 what could be the reason for this and how can i update the local zones to update 7 (0 Replies)
Discussion started by: fugitive
0 Replies

2. Solaris

SVM Soft partitions in a metaset- how to migrate ?

Hi Friends, I need your assistance, could someone assist.... Scenario:- Solaris 10, with zones and mounted file systems, using SAN. SVM with metasets on it configured soft partitions on the meta device. These soft partitions are mounted and file systems are being accessed. We are now trying... (6 Replies)
Discussion started by: unahb1
6 Replies

3. Solaris

Replacement of metadevice with soft partitions

Hi guys, currently I'm working on one problem. I have Sol10 box where I need to replace old storage for the new one with minimal downtime. Layout of metadevice looks like this: d103 p 1.0GB d100 d102 p 2.0GB d100 d101 p 1.0GB d100 d100 s ... (3 Replies)
Discussion started by: brusell
3 Replies

4. Solaris

Replace LUN with soft partitions

Hi all, can somebody provide me with advice if is possible to replace physical LUN, with soft partitions (located directly on LUN, not on metadevice) by the new one. Current state looks like this: root@xxx:/etc/zones # metastat -c| egrep -i 900bec2d286 d504 p 159GB... (1 Reply)
Discussion started by: brusell
1 Replies

5. Solaris

Live upgrade question

I want to basically update an ABE that someone created a few months back. I'm sure stuff has changed since it was made, and I was going to delete it and create a new one. But from what I'm looking at, the lumake appears like it would be a faster approach. I want to use live upgrade to... (0 Replies)
Discussion started by: BG_JrAdmin
0 Replies

6. Solaris

Live Upgrade Issue

I tried a live upgrade for one my solaris 10u8 server which didnt go sucessfull and after that i now have following mounts in memory. df: cannot statvfs /.alt.sol10u8_2/var: No such file or directory df: cannot statvfs /.alt.sol10u8_2/var/run: No such file or directory df: cannot statvfs... (0 Replies)
Discussion started by: fugitive
0 Replies

7. Emergency UNIX and Linux Support

Live Upgrade Query

I am upgrading Solaris-9 Update-7 to Solaris-9 Update-9 through live upgrade. I am able to create another boot environment and have OS DVD inside server. But i am confused for, what command/path should I give for luupgrade. OS DVD is mounted on /mnt Boot environments are Solaris9old (Active now)... (2 Replies)
Discussion started by: solaris_1977
2 Replies

8. Solaris

Live upgrade first steps

Hello Guys, I am a little confused about the first step in the live upgrade process. I will be glad if someone can clarify this for me. The pre-live upgrade patch, when do you add this patch to the OS you want to upgrade? 1. before creating the new boot environment? or 2. after creating... (1 Reply)
Discussion started by: cjashu
1 Replies
volfs(7FS)							   File Systems 							volfs(7FS)

NAME
volfs - Volume Management file system DESCRIPTION
volfs is the Volume Management file system rooted at root_dir. The default location for root-dir is /vol, but this can be overridden using the -d option of vold (see vold(1M)). This file system is maintained by the Volume Management daemon, vold, and will be considered to be /vol for this description. Media can be accessed in a logical manner (no association with a particular piece of hardware), or a physical manner (associated with a particular piece of hardware). Logical names for media are referred to through /vol/dsk and /vol/rdsk. /vol/dsk provides block access to random access devices. /vol/rdsk provides character access to random access devices. The /vol/rdsk and /vol/dsk directories are mirrors of one another. Any change to one is reflected in the other immediately. The dev_t for a volume will be the same for both the block and character device. The default permissions for /vol are mode=0555, owner=root, group=sys. The default permissions for /vol/dsk and /vol/rdsk are mode=01777, owner=root, group=sys. Physical references to media are obtained through /vol/dev. This hierarchy reflects the structure of the /dev name space. The default per- missions for all directories in the /vol/dev hierarchy are mode=0555, owner=root, group=sys. mkdir(2), rmdir(2), unlink(2) (rm), symlink(2) (ln -s), link(2) (ln), and rename(2) (mv) are supported, subject to normal file and direc- tory permissions. The following system calls are not supported in the /vol filesystem: creat(2), only when creating a file, and mknod(2). If the media does not contain file systems that can be automatically mounted by rmmount(1M), users can gain access to the media through the following /vol locations: +-----------------------------------+------------------------------------+ | Location | State of Media | +-----------------------------------+------------------------------------+ |/vol/dev/diskette0/unnamed_floppy | formatted unnamed floppy-block | | | device access | +-----------------------------------+------------------------------------+ |/vol/dev/rdiskette0/unnamed_floppy | formatted unnamed floppy-raw | | | device access | +-----------------------------------+------------------------------------+ |/vol/dev/diskette0/unlabeled | unlabeled floppy-block device | | | access | +-----------------------------------+------------------------------------+ |/vol/dev/rdiskette0/unlabeled | unlabeled floppy-raw device access | +-----------------------------------+------------------------------------+ |/vol/dev/dsk/c0t6/unnamed_cdrom | CD-ROM-block device access | +-----------------------------------+------------------------------------+ |/vol/dev/rdsk/c0t6/unnamed_cdrom | CD-ROM-raw device access | +-----------------------------------+------------------------------------+ For more information on the location of CD-ROM and floppy media, see System Administration Guide: Basic Administration or rmmount(1M). Partitions Some media support the concept of a partition. If the label identifies partitions on the media, the name of the media becomes a directory with partitions under it. Only valid partitions are represented. Partitions cannot be moved out of a directory. For example, if disk volume 'foo' has three valid partitions, 0, 2, and 5, then: /vol/dsk/foo/s0 /vol/dsk/foo/s2 /vol/dsk/foo/s5 for block access and /vol/rdsk/foo/s0 /vol/rdsk/foo/s2 /vol/rdsk/foo/s5 for character access. If a volume is relabeled to reflect different partitions, the name space changes to reflect the new partition layout. A format program can check to see if there are others with the volume open and not allow the format to occur if it is. Volume Management, however, does not explicitly prevent the rewriting of a label while others have the volume open. If a partition of a volume is open, and the volume is relabeled to remove that partition, it will appear exactly as if the volume were missing. A notify event will be generated and the user may cancel the operation with volcancel(1), if desired. SEE ALSO
volcancel(1), volcheck(1), volmissing(1) rmmount(1M), vold(1M), rmmount.conf(4), vold.conf(4) System Administration Guide: Basic Administration SunOS 5.10 8 Feb 1995 volfs(7FS)
All times are GMT -4. The time now is 03:55 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy