Sponsored Content
Full Discussion: Create rpool manually
Operating Systems Solaris Create rpool manually Post 302911481 by achenle on Thursday 31st of July 2014 09:48:58 PM
Old 07-31-2014
Maybe lucreate. Use it to create a new ZFS boot environment.

I think you can use the disk you wanted to use as a mirror to your current root drive, switch to the new root disk, then use the current root disk as a mirror for the new boot environment's root disk.
 

10 More Discussions You Might Find Interesting

1. Solaris

rpool issue

My system is showing following .. how can i remove this rpool .. ? zpool status -v rpool pool: rpool state: UNAVAIL status: One or more devices could not be opened. There are insufficient replicas for the pool to continue functioning. action: Attach the missing device and... (0 Replies)
Discussion started by: fugitive
0 Replies

2. Solaris

"flar create" produces a tiny flar image of rpool

Hi, can someone help with the following? I got 1x146GB disk which is span by just one zfs rpool. When issue for example: as you can see in above output the flar is very small, just around 10KB rpool with all data sets has several GB so this can't be good image. This is on the Solaris 10... (3 Replies)
Discussion started by: TomSu
3 Replies

3. Solaris

How to change hostid saved in rpool?

Hi, We have a Sun M5000. I am now trying to boot the second system domain by using the boot disk (a mirrored boot disk actually) of the first domain (if succeed then no OS installation is needed for the second domain). I got the following errors: SPARC Enterprise M5000 Server, using Domain... (21 Replies)
Discussion started by: aixlover
21 Replies

4. Solaris

ZFS rpool physical disk move

I'd like to finish setting up this system and then move the secondary or primary disk to another system that is the exact same hardware. I've done things like this in the past with ufs and disk suite mirroring just fine. But I have yet to do it with a zfs root pool mirror. Are there any... (1 Reply)
Discussion started by: Metasin
1 Replies

5. Solaris

How to access content of problematic rpool ?

Hi guys, I have a Solaris 10 x86 server with 2 physical disks. It had a ZFS rpool on the first disk c0t0d0s0. The server stopped booting with some error like "Cannot find bootfile". I booted with Solaris DVD and re-installed the operating system on the second disk with a different pool... (0 Replies)
Discussion started by: ARPcPro
0 Replies

6. Solaris

Resize rpool

Hi, I have rpool about 500G. So i want to use 210G from rpool and assign mount point as /database. I seek in google and couldn't found it. Does anyone know how to achieve it? Thanks. (2 Replies)
Discussion started by: mzainal
2 Replies

7. Solaris

Zfs rpool size

Hi everyone, I am doing housekeeping of my Solaris 11 for zfs snapshot to reduce the snapshot size. I have already cleared the / file system, however the rpool size still not reduced. Filesystem Size Used Available Capacity Mounted on rpool/ROOT/solaris-2 98G 6.9G ... (2 Replies)
Discussion started by: freshmeat
2 Replies

8. Solaris

Help with mirroring rpool

Hello, I need some help mirroring my rpool. I have a 60gb ssd running the rpool alone and want to mirror it for redundancy so I bought a 120gb ssd that I found for a good price. 60gb drives aren't as easy to find at a good price anymore it seems. I (a bit naively) thought after reading... (9 Replies)
Discussion started by: Zorken
9 Replies

9. Solaris

Trying to create ZFS slice on rpool

I have a 240GB disk as rpool. I have installed Solaris 11.3 to a partition which is 110GB. Now I have another 130GB which is unallocated. I want to use that additional space as a temporary folder to be shared between Solaris and Linux. The additional space had no /dev/dsk/c2t4... entry so I used... (8 Replies)
Discussion started by: kebabbert
8 Replies

10. Solaris

How to increase the size of rpool in Solaris 11?

Hi Please let me know how to increase the size of rpool in solaris 11 in ldom. I know how to map the new LUN to LDOM after that please let me know the procedure to increase the rpool and how to identify new disk in OBP level of ldom as I need to set the new/lun to be my new boot device. (5 Replies)
Discussion started by: hb00
5 Replies
vx_emerg_start(1M)														vx_emerg_start(1M)

NAME
vx_emerg_start - start Veritas Volume Manager from recovery media SYNOPSIS
vx_emerg_start [-m] [-r root_daname] hostname DESCRIPTION
The vx_emerg_start utility can be used to start Veritas Volume Manager (VxVM) when a system is booted from alternate media, or when a sys- tem has been booted into Maintenance Mode Boot (MMB) mode. This allows a rootable VxVM configuration to be repaired in the event of a cat- astrophic failure. vx_emerg_start verifies that the /etc/vx/volboot file exists, and checks the command-line arguments against the contents of this file. OPTIONS
-m Mounts the root file system contained on the rootvol volume after VxVM has been started. Prior to being mounted, the rootvol volume is started and fsck is run on the root file system. -r root_daname Specifies the disk access name of one of the members of the root disk group that is to be imported. This option can be used to spec- ify the appropriate root disk group when multiple generations of the same root disk group exist on the system under repair. If this option is not specified, the desired root disk group may not be imported if multiple disk groups with the same name exist on the sys- tem, and if one of these disk groups has a more recent timestamp. ARGUMENTS
hostname Specifies the system name (nodename) of the host system being repaired. This name is used to allow the desired root disk group to be imported. It must match the name of the system being repaired, as it is unlikely to be recorded on the recovery media from which you booted the system. NOTES
HP-UX Maintenance Mode Boot (MMB) is intended for recovery from catastrophic failures that have prevented the target machine from booting. If a VxVM root volume is mirrored, only one mirror is active when the system is in MMB mode. Any writes that are made to the root file sys- tem in this mode can corrupt this file system when both mirrors are subsequently configured. The vx_emerg_start script allows VxVM to be started while a system is in MMB mode, and marks the non-boot mirror plexes as stale. This prevents corruption of the root volume or file system by forcing a subsequent recovery from the boot mirror to the non-boot mirrors to take place. USAGE
After VxVM has been started, various recovery options can be performed depending on the nature of the problem. It is recommended that you use the vxprint command to determine the state of the configuration. One common problem is when all the plexes of the root disk are stale as shown in the following sample output from vxprint: v rootvol root DISABLED 393216 - ACTIVE - pl rootvol-01 rootvol DISABLED 393216 - STALE - sd rootdisk01-02 rootvol-01 ENABLED 393216 0 - - pl rootvol-02 rootvol DISABLED 393216 - STALE - pl rootvol-02 rootvol DISABLED 393216 - STALE - sd rootdisk02-02 rootvol-02 ENABLED 393216 0 - - In this case, the volume can usually be repaired by using the vxvol command as shown here: vxvol -g 4.1ROOT -f start rootvol If the volume is mirrored, it is put in read-write-back recovery mode. As the command is run in the foreground, it does not exit until the recovery is complete. It is then recommended that you run fsck on the root file system, and mount it, before attempting to reboot the sys- tem: fsck -F vxfs -o full /dev/vx/rdsk/4.1ROOT/rootvol mkdir /tmp_mnt mount -F vxfs /dev/vx/dsk/4.1ROOT/rootvol /tmp_mnt SEE ALSO
fsck(1M), mkdir(1M), mount(1M), vxintro(1M), vxprint(1M), vxvol(1M) VxVM 5.0.31.1 24 Mar 2008 vx_emerg_start(1M)
All times are GMT -4. The time now is 03:46 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy