Sponsored Content
Full Discussion: ZFS pool question
Operating Systems Solaris ZFS pool question Post 302368260 by bartus11 on Wednesday 4th of November 2009 12:13:22 PM
Old 11-04-2009
If my previous command didn't work, try this:
Code:
zfs list -t filesystem,snapshot,volume

The thing is you probably have snapshot of that deleted file somewhere and that is what consumes space.
 

10 More Discussions You Might Find Interesting

1. Solaris

ZFS Pool Mix-up

Hi all I plan to install Solaris 10U6 on some SPARC server using ZFS as root pool, whereas I would like to keep the current setup done by VxVM: - 2 internal disks: c0t0d0 and c0t1d0 - bootable root-volume (mirrored, both disks) - 1 non-mirrored swap slice - 1 non-mirrored slices for Live... (1 Reply)
Discussion started by: blicki
1 Replies

2. Infrastructure Monitoring

zfs - migrate from pool to pool

Here are the details. cnjr-opennms>root$ zfs list NAME USED AVAIL REFER MOUNTPOINT openpool 20.6G 46.3G 35.5K /openpool openpool/ROOT 15.4G 46.3G 18K legacy openpool/ROOT/rds 15.4G 46.3G 15.3G / openpool/ROOT/rds/var 102M ... (3 Replies)
Discussion started by: pupp
3 Replies

3. Solaris

ZFS - list of disks used in a pool

Hi guys, We had created a pool as follows: zpool create filing_pool raidz c1t2d0 c1t3d0 ........ Due to some requirement, we need to destroy the pool and re-create another one. We wish to know now which disks have been included in the filing_pool, how do we list the disks used to create... (2 Replies)
Discussion started by: frum
2 Replies

4. Solaris

zfs pool migration

I need to migrate an existing raidz pool to a new raidz pool with larger disks. I need the mount points and attributes to migrate as well. What is the best procedure to accomplish this. The current pool is 6x36GB disks 202GB capacity and I am migrating to 5x 72GB disks 340GB capacity. (2 Replies)
Discussion started by: jac
2 Replies

5. Solaris

Best way to rename a ZFS Pool?

Other than export/import, is there a cleaner way to rename a pool without unmounting de FS? Something like, say "zpool rename a b"? Thanks. (2 Replies)
Discussion started by: verdepollo
2 Replies

6. Solaris

ZFS - overfilled pool

installed Solaris 11 Express on my server machine a while ago. I created a Z2 RAID over five HDDs and created a few ZFS filesystems on it. Once I (unintentionally) managed to fill the pool completely with data and (to my surprise) the filesystems stopped working - I could not read/delete any... (3 Replies)
Discussion started by: RychnD
3 Replies

7. Solaris

ZFS - Dataset / pool name are the same...cannot destroy

I messed up my pool by doing zfs send...recive So I got the following : zpool list NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT rpool 928G 17.3G 911G 1% 1.00x ONLINE - tank1 928G 35.8G 892G 3% 1.00x ONLINE - So I have "tank1" pool. zfs get all... (8 Replies)
Discussion started by: eladgrs
8 Replies

8. Solaris

reassign zfs pool lun

I have a branded zone txdjintra that utilizes a pool named Pool_djintra that is no longer required. There is a 150 Gig Lun assigned to the pool that I need to reassign to another branded zone txpsrsrv07 with a pool named Pool_txpsrsrv07 on the same sun blade. What is the process to do this? ... (0 Replies)
Discussion started by: jeffsr
0 Replies

9. Solaris

Zfs send to compressed pool?

I have a newly created zpool, and I have set compression on, for the whole pool: # zfs set compression=on newPool Now I have zfs send | zfs receive lot of snapshots to my newPool, but the compression is gone. I was hoping that I would be able to send snapshots to the new pool (which is... (0 Replies)
Discussion started by: kebabbert
0 Replies

10. UNIX for Beginners Questions & Answers

Opening up ZFS pool as writable

I have installed FreeBSD onto a raw image file using QEMU Emulator successfully. I have formatted the image file using the ZFS file system (ZFS POOL). Using the following commands below I have successfully mounted the image file ready to be opened by zpool sudo losetup /dev/loop0 .img sudo... (1 Reply)
Discussion started by: alphatron150
1 Replies
vxcached(1M)															      vxcached(1M)

NAME
vxcached - resize cache volumes when required SYNOPSIS
/etc/vx/bin/vxcached [-v] [mail-address ...] DESCRIPTION
The Veritas Volume Manager cache daemon, vxcached, is invoked by the VxVM kernel to monitor the cache objects that are used by space-opti- mized snapshots. When usage of a cache volume by space-optimized snapshots reaches the high watermark level that has been set for the cache, vxcached automatically grows the cache volume if this is required and configured. The -v option enables the daemon in verbose mode. If no mail addresses are given as arguments, mail is sent to root. If the autogrow feature has been enabled on a cache object, the values of the highwatermark, autogrowby and maxautogrow attributes for the cache object determine the behavior of vxcached: o When cache usage reaches the high watermark value, highwatermark, (default value is 90 percent), and the new required cache size would not exceed the value of maxautogrow (default value is twice the size of the cache volume in blocks), vxcached grows the size of the cache volume by the value of autogrowby (default value is 20% of the size of the cache volume in blocks). o When cache usage reaches the high watermark value, and the new required cache size would exceed the value of maxautogrow, vxcached deletes the oldest snapshot in the cache. If there are several snapshots with the same age, the largest of these is deleted. If the autogrow feature has been disabled on a cache object: o When cache usage reaches the high watermark value, vxcached deletes the oldest snapshot in the cache. If there are several snapshots with the same age, the largest of these is deleted. If there is only a single snapshot, the snapshot is detached and marked as invalid. The values of the highwatermark, autogrowby and maxautogrow attributes can be set when a cache object is created using vxmake. If neces- sary, you can use the vxcache set command to change the values of these attributes for an existing cache. NOTES
The vxcached daemon is started automatically if a license for the FastResync feature is already present on the system. If you add such a license while VxVM is running, you can either start vxcached manually, or shut down and reboot the system. Killing the vxcached process stops it from removing snapshots. The daemon can be prevented from being started by commenting out its entry in the startup script, /sbin/init.d/vxvm-recover. The vxcached daemon does not remove snapshots that are currently open, and it does not remove the last or only snapshot in the cache. If the cache space becomes exhausted, the snapshot is detached. If this happens, the snapshot is unrecoverable and must be removed manu- ally. Enabling the autogrow feature on the cache helps to avoid this situation occurring. However, for very small caches (of the order of a few megabytes), it is possible for the cache to become exhausted before the system has time to respond and grow the cache. In such cases, either increase the size of the cache manually, or reduce the value of highwatermark. You can use the maxautogrow attribute to limit the maximum size to which a cache can grow. To estimate this size, consider how much the contents of each source volume are likely to change between snapshot refreshes, and allow some additional space for contingency. SEE ALSO
vxcache(1M), vxintro(1M), vxmake(1M), vxsnap(1M) VxVM 5.0.31.1 24 Mar 2008 vxcached(1M)
All times are GMT -4. The time now is 02:29 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy