Sponsored Content
Full Discussion: ZFS Pool Mix-up
Operating Systems Solaris ZFS Pool Mix-up Post 302317877 by blicki on Wednesday 20th of May 2009 06:49:47 AM
Old 05-20-2009
Question ZFS Pool Mix-up

Hi all

I plan to install Solaris 10U6 on some SPARC server using ZFS as root pool, whereas I would like to keep the current setup done by VxVM:

- 2 internal disks: c0t0d0 and c0t1d0
- bootable root-volume (mirrored, both disks)
- 1 non-mirrored swap slice
- 1 non-mirrored slices for Live Upgrade

Is this possible using ZFS? Can I modify the properties of each volume after creating a mirrored pool oder have some volume mirrored afterwards?

Thanks for any help.

Kind regards, Mike
 

10 More Discussions You Might Find Interesting

1. Solaris

unable to import zfs pool

# zpool import pool: emcpool1 id: 5596268873059055768 state: UNAVAIL status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: Sun Message ID: ZFS-8000-3C config: emcpool1 ... (7 Replies)
Discussion started by: fugitive
7 Replies

2. Infrastructure Monitoring

zfs - migrate from pool to pool

Here are the details. cnjr-opennms>root$ zfs list NAME USED AVAIL REFER MOUNTPOINT openpool 20.6G 46.3G 35.5K /openpool openpool/ROOT 15.4G 46.3G 18K legacy openpool/ROOT/rds 15.4G 46.3G 15.3G / openpool/ROOT/rds/var 102M ... (3 Replies)
Discussion started by: pupp
3 Replies

3. Solaris

ZFS pool question

I created a pool the other day. I created a 10 gig files just for a test, then deleted it. I proceeded to create a few files systems. But for some reason the pool shows 10% full, but the files systems are both at 1%? Both files systems share the same pool. When I ls -al the pool I just... (6 Replies)
Discussion started by: mrlayance
6 Replies

4. Solaris

zfs pool migration

I need to migrate an existing raidz pool to a new raidz pool with larger disks. I need the mount points and attributes to migrate as well. What is the best procedure to accomplish this. The current pool is 6x36GB disks 202GB capacity and I am migrating to 5x 72GB disks 340GB capacity. (2 Replies)
Discussion started by: jac
2 Replies

5. Solaris

Best way to rename a ZFS Pool?

Other than export/import, is there a cleaner way to rename a pool without unmounting de FS? Something like, say "zpool rename a b"? Thanks. (2 Replies)
Discussion started by: verdepollo
2 Replies

6. Solaris

ZFS - overfilled pool

installed Solaris 11 Express on my server machine a while ago. I created a Z2 RAID over five HDDs and created a few ZFS filesystems on it. Once I (unintentionally) managed to fill the pool completely with data and (to my surprise) the filesystems stopped working - I could not read/delete any... (3 Replies)
Discussion started by: RychnD
3 Replies

7. Solaris

ZFS - Dataset / pool name are the same...cannot destroy

I messed up my pool by doing zfs send...recive So I got the following : zpool list NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT rpool 928G 17.3G 911G 1% 1.00x ONLINE - tank1 928G 35.8G 892G 3% 1.00x ONLINE - So I have "tank1" pool. zfs get all... (8 Replies)
Discussion started by: eladgrs
8 Replies

8. Solaris

Need to remove a disk from zfs pool

I accidently added a disk in different zpool instead of pool, where I want. root@prtdrd21:/# zpool status cvfdb2_app_pool pool: cvfdb2_app_pool state: ONLINE scan: none requested config: NAME STATE READ WRITE CKSUM cvfdb2_app_pool ONLINE 0 0 0... (1 Reply)
Discussion started by: solaris_1977
1 Replies

9. Solaris

Zfs send to compressed pool?

I have a newly created zpool, and I have set compression on, for the whole pool: # zfs set compression=on newPool Now I have zfs send | zfs receive lot of snapshots to my newPool, but the compression is gone. I was hoping that I would be able to send snapshots to the new pool (which is... (0 Replies)
Discussion started by: kebabbert
0 Replies

10. UNIX for Beginners Questions & Answers

Opening up ZFS pool as writable

I have installed FreeBSD onto a raw image file using QEMU Emulator successfully. I have formatted the image file using the ZFS file system (ZFS POOL). Using the following commands below I have successfully mounted the image file ready to be opened by zpool sudo losetup /dev/loop0 .img sudo... (1 Reply)
Discussion started by: alphatron150
1 Replies
lvextend(1M)															      lvextend(1M)

NAME
lvextend - increase space, increase mirrors for LVM logical volume SYNOPSIS
autobackup] le_number | lv_size | mirror_copies lv_path [pv_path ... | pvg_name ...] Remarks Mirrored disk operations require the installation of the optional HP MirrorDisk/UX software, which is not included in the standard HP-UX operating system. DESCRIPTION
The command can increase a logical volume's allocated extents, or increase its number of mirrored copies. Other logical volume characteristics can be modified with the and commands (see lvchange(1M) and lvreduce(1M)). To limit the allocation to specific physical volumes, specify the physical volume names as pv_path arguments or specify the physical volume group names as pvg_name arguments. Otherwise, all of the physical volumes in a volume group are available for allocating new physical extents. LVM always ensures that physical extent allocation can satisfy the current allocation policy or policies. If a physical volume is not suitable for use with a certain allocation policy, it is not used during physical extent allocation, even it is specified in a pv_path argument or indirectly in a pvg_name argument. The pvg_name argument is allowed only if one of the allocation policies of the logical volume is PVG-strict. Options and Arguments The option is only meaningful if the optional HP MirrorDisk/UX software has been installed. recognizes the following options and arguments: lv_path The block device path name of a logical volume. pv_path The block device path name of a physical volume. pvg_name The name of a physical volume group (see lvmpvg(4)). Set automatic backup for this invocation of this command. autobackup can have one of the following values: Automatically back up configuration changes made to the logical volume. This is the default. After this command executes, the command (see vgcfgbackup(1M)) is executed for the volume group to which the logical volume belongs. Do not back up configuration changes this time. Increase the space allocated to the logical volume, specified in logical extents. le_number is a decimal value greater than the current number of logical extents. le_number must be at least 1 and no greater than a volume group version-dependent maximum; use the command to determine the maximum number of logical extents for the volume group version. One, and only one, or option must be supplied. Increase the space allocated to the logical volume, specified in megabytes. lv_size is a decimal value greater than the current logical volume size. lv_size must be at least 1 and no greater than a volume group version-dependent maximum; use the command to determine the maximum logical volume size for the volume group version. lv_size is rounded up to the nearest multiple of the logical extent size, equivalent to the physical extent size defined for the volume group by the command (see vgcreate(1M)). One, and only one, or option must be specified. Set the number of mirror copies allocated for each logical extent. A mirror copy contains the same data as the original. mirror_copies must be at least 1 and no greater than a volume group version-dependent maximum; use the command to determine the maximum number of mirror copies for the volume group version. mirror_copies must be greater than the current value. Data in the new copies is synchronized unless the option is specified. The synchronization process can be time consuming, depending on hardware characteristics and the amount of data. One, and only one, or option must be specified. Do not synchronize the new mirror copies. This may affect data high availability so use or to synchronize the mirrors. The option must be specified along with this option. Striped Logical Volume considerations Striped and mirrored logical volumes are supported. An increase in size of a striped logical volume is done by increments of stripes logical extents. One increment corresponds to stripes physical extents if the volume is not mirrored or to stripes * (mirror_copies + 1) physical extents if the volume is mirrored. stripes is the number of disks the logical volume is striped across. It is set with the option stripes of the command. mirror_copies is the number of mirror copies allocated for each extent. It is set with the option of the and commands. LVM striped logical volumes are always allocated using the strict or PVG-strict allocation policies. Each physical extent of an increment is allocated on a different physical volume in the volume group. A size increase of a striped volume requires at least stripes (or stripes * (mirror_copies + 1) if the volume is mirrored) physical volumes with adequate free space and meeting the allocation policy. An increase of the number of mirror copies of a striped volume requires at least (stripes times the number of copies to add) physical vol- umes with adequate free space and meeting the allocation policy. Shared Volume Group Considerations For volume group version 1.0 and 2.0, cannot be used if the volume group is activated in shared mode. For volume groups version 2.1 (or higher), can be performed when activated in either shared, exclusive, or standalone mode. Note that the daemon must be running on all the nodes sharing a volume group activated in shared mode. See lvmpud(1M). If physical volume groups are passed as arguments, uses the physical volume group file of the system where the command is issued (the server). LVM shared mode is currently only available in Serviceguard clusters. EXTERNAL INFLUENCES
Environment Variables determines the language in which messages are displayed. If is not specified or is null, it defaults to "C" (see lang(5)). If any internationalization variable contains an invalid setting, all internationalization variables default to "C" (see environ(5)). EXAMPLES
Increase the number of the logical extents of a logical volume to 100: Increase the logical volume size to 400 MB: Allocate two mirrors (that is, two copies of the original) for each logical extent of a logical volume: Mirror a logical volume onto a particular physical volume. Allocate one mirror and do not synchronize the new mirror copy: Increase the size of a file system existing on a logical volume. First, increase the size of the logical volume. Unmount the file system. Extend the file system to occupy the entire (larger) logical volume. Remount the file system. WARNINGS
The creation of striped and mirrored logical volume(s) may prevent the import and activation of the volume group on an earlier HP-UX release. See lvcreate(1M) on the earlier release to see if it explicitly states that striping and mirroring is supported. If the striped and mirrored logical volumes of the volume group are removed or un-mirrored, the volume group becomes again compatible with the older HP-UX releases. SEE ALSO
lvchange(1M), lvcreate(1M), lvdisplay(1M), lvmadm(1M), lvmpud(1M), lvreduce(1M), lvsync(1M), pvchange(1M), pvdisplay(1M), vgsync(1M), intro(7), lvm(7). lvextend(1M)
All times are GMT -4. The time now is 08:13 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy