How to split a zfs/zpool mirror for backout purposes?


 
Thread Tools Search this Thread
Operating Systems Solaris How to split a zfs/zpool mirror for backout purposes?
# 1  
Old 09-03-2015
How to split a zfs/zpool mirror for backout purposes?

Hello,

I am upgrading Veritas from 5.1 to 6.0.1/6.0.5 in a Solaris 10 u8 server with OS mirrored (rpool) in zfs/zpool configuration.

I need to split it to have a quick way to backout in case of failure (make splitted mirror side bootable for a quick revert, like booting from it). I remember that I did that in SVM or when OS drive was encapsulated in VxVM, but don't know how to do it in zfs/zpool.

Please your assistance.

Code:
$ zpool list
NAME    SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
rpool   136G  83.4G  52.6G    61%  ONLINE  -
$
$ zpool status
  pool: rpool
 state: ONLINE
 scrub: none requested
config:

        NAME          STATE     READ WRITE CKSUM
        rpool         ONLINE       0     0     0
          mirror      ONLINE       0     0     0
            c0t0d0s0  ONLINE       0     0     0
            c0t1d0s0  ONLINE       0     0     0

errors: No known data errors

# 2  
Old 09-03-2015
Zpool Detach

assuming you did
Code:
zpool attach

instead of
Code:
zpool add

it should just be a matter of doing
Code:
zpool detach

.
# 3  
Old 09-03-2015
Is it necessary to install a boot block prior detach or is it copied during the attach process (sync)?
# 4  
Old 09-03-2015
I've successfully booted from it just by recreating the zpool from the detached volume. it's already got the boot block in it because it was part of the root pool.

---------- Post updated at 09:44 AM ---------- Previous update was at 09:40 AM ----------

so the basic procedure here is to detach the second volume. proceed with your patching and then if you want to recover, destroy the zpool with the remaining volume, and do a zpool import to rebuild the zpool from the detached volume.

If you reattach it... it will sync the bad data over the good. Which isn't really what you want unless you intend to destroy the reserved copy.
Login or Register to Ask a Question

Previous Thread | Next Thread

9 More Discussions You Might Find Interesting

1. Solaris

Restore of Netapp FC lun targets used as the disks for a zpool with exported zfs file systems

So, We have a Netapp storage solution. We have Sparc T4-4s running with LDOMS and client zones in the LDOMS, We are using FC for storage comms. So here's the basic setup FC luns are exported to the primary on the Sparc box. using LDM they are then exported to the LDOM using vdisk. at the... (4 Replies)
Discussion started by: os2mac
4 Replies

2. Solaris

Is there any GUI or web ui to administrate zfs/zpool

HI, Is there any GUI or web ui to administrate zfs/zpool. i want to monitor/expand/migrate zfs from one machine to another machine (0 Replies)
Discussion started by: bentech4u
0 Replies

3. Solaris

Upgrading Solaris - what happens to zpool/zfs versions

Hi everyone, I'm hoping someone can help me out here. I've googled lots and don't think I can find an easy answer to this. We're in the process of upgrading Solaris from v10 5/08 to v10 9/10. The zpools for luns are currently at version 10, and I understand Solaris v10 9/10 has support for... (3 Replies)
Discussion started by: badoshi
3 Replies

4. Solaris

Zpool /ZFS recovery

running VM server for Sparc on a Solaris 11 server. I have a Guest LDOm that had two seperate zpools running, one for the zones and one for the OS. The OS was corrupted and had to be replaced. The zones zfs file system is intact I think. I still have access to the disk and can still see it in... (3 Replies)
Discussion started by: os2mac
3 Replies

5. Solaris

[Q] zpool mirror; read from preferred vdev(s)

Hallo, Using a zpool mirror; is there a way to set something like a primary storage / preferred vdevs / read-first devices like it was in disksuite using "metaparam -r " or the "vxvol rdpol " to define a preferred plex in VxVM? Will ZFS always use round-robin? In our example we want to use a... (3 Replies)
Discussion started by: pressy
3 Replies

6. BSD

Unable to create zfs zpool in FreeBSD 8.2: no such pool or dataset

I am trying to test simple zfs functionality on a FreeBSD 8.2 VM. When I try to run a 'zpool create' I receive the following error: # zpool create zfspool /dev/da0s1a cannot create 'zfspool': no such pool or dataset # zpool create zfspool /dev/da0 cannot create 'zfspool': no such pool or... (3 Replies)
Discussion started by: bstring
3 Replies

7. Solaris

Zfs::zpool.cache file

Hi All, I am trying to read zpool.cache file to find out pool information like pool name, devices it uses and all properties. File seems to be in packed format.I am not sure how to unpack it. But from opensolaris code base we can see that they have used libz for uncompromising this file, but... (0 Replies)
Discussion started by: shailesh_111
0 Replies

8. Solaris

ZFS Mirror versus Hardware Mirror

I've looked a little but haven't found a solid answer, assuming there is one. What's better, hardware mirroring or ZFS mirroring? Common practice for us was to use the raid controllers on the Sun x86 servers. Now we've been using ZFS mirroring since U6. Any performance difference? Any other... (3 Replies)
Discussion started by: Lespaul20
3 Replies

9. Solaris

zfs mirror disk

Is it possible to create a Mirror with zfs ?? I'm experimented user with Solstice Disk suite. Or Sun Volume manager or veritas volume manager. But, i would like switch from Disksuite to Zfs. All my mirrored disks. (1 Reply)
Discussion started by: simquest
1 Replies
Login or Register to Ask a Question