Sponsored Content
Operating Systems Solaris How to split a zfs/zpool mirror for backout purposes? Post 302954030 by os2mac on Thursday 3rd of September 2015 01:44:05 PM
Old 09-03-2015
I've successfully booted from it just by recreating the zpool from the detached volume. it's already got the boot block in it because it was part of the root pool.

---------- Post updated at 09:44 AM ---------- Previous update was at 09:40 AM ----------

so the basic procedure here is to detach the second volume. proceed with your patching and then if you want to recover, destroy the zpool with the remaining volume, and do a zpool import to rebuild the zpool from the detached volume.

If you reattach it... it will sync the bad data over the good. Which isn't really what you want unless you intend to destroy the reserved copy.
 

9 More Discussions You Might Find Interesting

1. Solaris

zfs mirror disk

Is it possible to create a Mirror with zfs ?? I'm experimented user with Solstice Disk suite. Or Sun Volume manager or veritas volume manager. But, i would like switch from Disksuite to Zfs. All my mirrored disks. (1 Reply)
Discussion started by: simquest
1 Replies

2. Solaris

ZFS Mirror versus Hardware Mirror

I've looked a little but haven't found a solid answer, assuming there is one. What's better, hardware mirroring or ZFS mirroring? Common practice for us was to use the raid controllers on the Sun x86 servers. Now we've been using ZFS mirroring since U6. Any performance difference? Any other... (3 Replies)
Discussion started by: Lespaul20
3 Replies

3. Solaris

Zfs::zpool.cache file

Hi All, I am trying to read zpool.cache file to find out pool information like pool name, devices it uses and all properties. File seems to be in packed format.I am not sure how to unpack it. But from opensolaris code base we can see that they have used libz for uncompromising this file, but... (0 Replies)
Discussion started by: shailesh_111
0 Replies

4. BSD

Unable to create zfs zpool in FreeBSD 8.2: no such pool or dataset

I am trying to test simple zfs functionality on a FreeBSD 8.2 VM. When I try to run a 'zpool create' I receive the following error: # zpool create zfspool /dev/da0s1a cannot create 'zfspool': no such pool or dataset # zpool create zfspool /dev/da0 cannot create 'zfspool': no such pool or... (3 Replies)
Discussion started by: bstring
3 Replies

5. Solaris

[Q] zpool mirror; read from preferred vdev(s)

Hallo, Using a zpool mirror; is there a way to set something like a primary storage / preferred vdevs / read-first devices like it was in disksuite using "metaparam -r " or the "vxvol rdpol " to define a preferred plex in VxVM? Will ZFS always use round-robin? In our example we want to use a... (3 Replies)
Discussion started by: pressy
3 Replies

6. Solaris

Zpool /ZFS recovery

running VM server for Sparc on a Solaris 11 server. I have a Guest LDOm that had two seperate zpools running, one for the zones and one for the OS. The OS was corrupted and had to be replaced. The zones zfs file system is intact I think. I still have access to the disk and can still see it in... (3 Replies)
Discussion started by: os2mac
3 Replies

7. Solaris

Upgrading Solaris - what happens to zpool/zfs versions

Hi everyone, I'm hoping someone can help me out here. I've googled lots and don't think I can find an easy answer to this. We're in the process of upgrading Solaris from v10 5/08 to v10 9/10. The zpools for luns are currently at version 10, and I understand Solaris v10 9/10 has support for... (3 Replies)
Discussion started by: badoshi
3 Replies

8. Solaris

Is there any GUI or web ui to administrate zfs/zpool

HI, Is there any GUI or web ui to administrate zfs/zpool. i want to monitor/expand/migrate zfs from one machine to another machine (0 Replies)
Discussion started by: bentech4u
0 Replies

9. Solaris

Restore of Netapp FC lun targets used as the disks for a zpool with exported zfs file systems

So, We have a Netapp storage solution. We have Sparc T4-4s running with LDOMS and client zones in the LDOMS, We are using FC for storage comms. So here's the basic setup FC luns are exported to the primary on the Sparc box. using LDM they are then exported to the LDOM using vdisk. at the... (4 Replies)
Discussion started by: os2mac
4 Replies
lvlnboot(1M)															      lvlnboot(1M)

NAME
lvlnboot - prepare LVM logical volume to be root, boot, primary swap, or dump volume SYNOPSIS
autobackup] { boot_lv | dump_lv | root_lv | | swap_lv }] [vg_name] Remarks cannot be performed if the volume group is activated in shared mode. DESCRIPTION
This command is supported for arguments belonging to volume groups version 1.0. Attempting this command with arguments belonging to volume groups versions 2.0 or higher will result in an error. The command updates all physical volumes in the volume group so that the logical volume becomes the root, boot, primary swap, or a dump volume when the system is next booted on the volume group. If a nonexistent logical volume is specified, this command fails. If a differ- ent logical volume is already linked to the root or primary swap, the command fails. The boot information stored on disks assumes a certain ordering of disks listed in lvlnboot(1M) command should be run in recovery mode option) when there is a change to the order of the disks in for any bootable volume group (see vgimport(1M) and vgscan(1M) for more infor- mation). Options and Arguments recognizes the following options and arguments: vg_name The path name of a volume group. Set automatic backup for this invocation of this command. autobackup can have one of the following values: Automatically back up configuration changes made to the logical volume. This is the default. After this command executes, the command (see vgcfgbackup(1M)) is executed for the volume group to which the logical volume belongs. Do not back up configuration changes this time. Define boot_lv to be the boot volume the next time the system is booted on the volume group. boot_lv must be the first logical volume on the physical volume. boot_lv must be contiguous, and must not allow bad block relocation. boot_lv is used to locate the boot file system during the boot process. The boot file system has the kernel which is read by the boot loader (see hpux(1M) for PA-RISC systems). This command updates the file with the location of the root volume in the currently booted volume group. The file is used during maintenance-mode boots to locate the root volume. During normal boots (versus maintenance-mode boots, see hpux(1M) for PA-RISC systems), this command is automati- cally executed by (see inittab(4)). Since this command is performed during boot, it does not need to be performed manually unless is missing (or alternatively, performing a normal reboot will recreate this file). During maintenance-mode boots, since the root volume group is not activated, does not update Maintenance-mode boot will fail if does not already exist with the correct location of the root volume. See When a new root volume group is created, the first boot must be a normal boot (versus a maintenance-mode boot), so that gets created. This option does not allow updating for any volume group other than the one that is booted. Define dump_lv to be one of the dump volumes the next time the system is booted on the volume group. dump_lv must be a contiguous logical volume and cannot have Bad Block Relocation enabled. The command updates the Boot Data Reserved Area of each bootable physical volume in the volume group (see pvcre- ate(1M)). The combined size of all the dump volumes should be at least 2048 bytes larger than the total memory of the sys- tem. Multiple dump devices can be configured, but each dump_lv must be entered with a separate command line. This option will be obsolete in the next HP-UX release. Define root_lv to be the root volume the next time the system is booted on this volume group. root_lv must be a con- tiguous logical volume and cannot have bad block relocation enabled. If root_lv is the first logical volume on the physical volume, then it is configured as the combined root-boot volume. Otherwise, root_lv is configured as the separate root volume in which case a separate boot volume needs to be configured using the option. Either the separate root or the separate boot volume can be configured first. The command updates the Boot Data Reserved Area of each bootable physical volume (see pvcreate(1M)) to enable the volume group to be used to locate the root file system. root_lv is also used as the root volume during a maintenance-mode boot (see hpux(1M) for PA-RISC systems). The physical volumes containing root_lv must have been created using the option (see pvcreate(1M)), indicating that that physical volume is to be used as a bootable physical volume. Also, the command (see mkboot(1M)) must have been run on the physical volume to create the LIF area at the top of the physical volume (see lif(4)). Recover any missing links to all of the logical volumes specified in the Boot Data Reserved Area and update the Boot Data Reserved Area of each bootable physical volume in the volume group (see pvcreate(1M)). Define swap_lv to be the primary swap volume the next time the system is booted on the volume group. swap_lv must be a contiguous logical volume, and a root logical volume must have been previously defined with this command. The command updates the Boot Data Reserved Area of each bootable physical volume in the volume group (see pvcre- ate(1M)). Any existing swap area previously defined must be removed via lvrmboot(1M). This option will be obsolete in the next HP-UX release. Print verbose messages. With no other arguments present, print information on root, boot, swap, and dump logical volumes. If a combined root-boot volume is configured, no information for the boot volume is displayed. EXTERNAL INFLUENCES
Environment Variables determines the language in which messages are displayed. If is not specified or is null, it defaults to "C" (see lang(5)). If any internationalization variable contains an invalid setting, all internationalization variables default to "C" (see environ(5)). EXAMPLES
The following examples show configuration of a combined root-boot volume. Create a root volume group, containing root, swap, and dump logical volumes. Assume that an appropriate directory called and a cor- responding file already exist (see lvm(7)). First, initialize the disk, say so that it can be used as an LVM boot disk. Place the LIF information on the disk using the command. Create the volume group Create a logical volume that is suitable for use as the root volume. This logical volume has to be the first in the volume group and should be a contiguous volume with bad block relocation turned off. Create a logical volume that will be used as primary swap. This volume should be contiguous. Create a logical volume that will be used as the dump volume. This volume should be contiguous with bad block relocation turned off. Specify that the logical volume, will be used as the root volume. Specify that the logical volume, will be used as the primary swap. Specify that the logical volume, will be used as the dump volume. Display the results of the previous operations. The following examples show configuration of separate root and boot volumes. Create a root volume group, containing root, boot, swap, and dump logical volumes. Assume that an appropriate directory called and a corresponding file already exist (see lvm(7)). First, initialize the disk, say so that it can be used as an LVM boot disk. Place the LIF information on the disk using the command. Create the volume group Create a logical volume that is suitable for use as the boot volume. This logical volume has to be the first in the volume group and should be a contiguous volume with bad block relocation turned off. Create a logical volume that is suitable for use as the root volume. This logical volume should be a contiguous volume with bad block relocation turned off. Create a logical volume that will be used as primary swap. This volume should be contiguous. Create a logical volume that will be used as the dump volume. This volume should be contiguous with bad block relocation turned off. Specify that the logical volume, will be used as the root volume. Specify that the logical volume, will be used as the boot volume. Specify that the logical volume, will be used as the primary swap. Specify that the logical volume, will be used as the dump volume. Display the results of the previous operations. The following example shows configuration of multiple dump volumes. Specify that logical volumes and should be used as the dump logical volumes and that should also be used as primary swap. Assume that the volume group and the logical volumes have been created and the logical volumes are contiguous. WARNINGS
Dump Volume Warnings At the HP-UX 11.00 release and forward, the command will support any size dump device depending upon the IODC firmware addressability of the system. If the configured dump logical volume is out of the range of what the firmware can address, the command will return an error message such as "Unable to configure dump logical volume. Dump logical volume size beyond the IODC max address." Separate Root/Boot Warnings Whenever mkboot(1M) is used to restore the LIF area of a damaged root physical volume, the boot_lv option of must be performed afterwards to record the boot volume information inside the new LIF (see lif(4)). Subsequent commands such as are dependent on the boot_lv informa- tion inside the LIF. If the option does not locate the boot volume boot_lv, and the root_lv has not yet been performed, then performing the root_lv option will enable the boot volume to be located. The command derives the location of boot volume from the location of the root volume. Maintenance-Mode Warnings When creating additional root volumes, a normal boot must be performed on each new root volume so that which is required for maintenance- mode boots (see hpux(1M) for PA-RISC systems), gets created for each new root volume. Mirrored root_lv volumes should start at the same offset on each physical volume so that the location stored in works for maintenance-mode boots off of any mirror. FILES
Contains the location of the root volume. Used during maintenance-mode boots (see hpux(1M) for PA-RISC systems) to locate the root volume. SEE ALSO
lvcreate(1M), lvrmboot(1M), lvmadm(1M), mkboot(1M), pvcreate(1M), vgcreate(1M), inittab(4), lif(4), lvm(7). lvlnboot(1M)
All times are GMT -4. The time now is 12:35 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy