Sponsored Content
Operating Systems Solaris Using liveupgrade on single ZFS pool Post 302731541 by batas on Thursday 15th of November 2012 09:25:43 AM
Old 11-15-2012
Thanks all for the info Smilie
 

9 More Discussions You Might Find Interesting

1. Solaris

ZFS Pool Mix-up

Hi all I plan to install Solaris 10U6 on some SPARC server using ZFS as root pool, whereas I would like to keep the current setup done by VxVM: - 2 internal disks: c0t0d0 and c0t1d0 - bootable root-volume (mirrored, both disks) - 1 non-mirrored swap slice - 1 non-mirrored slices for Live... (1 Reply)
Discussion started by: blicki
1 Replies

2. Solaris

unable to import zfs pool

# zpool import pool: emcpool1 id: 5596268873059055768 state: UNAVAIL status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: Sun Message ID: ZFS-8000-3C config: emcpool1 ... (7 Replies)
Discussion started by: fugitive
7 Replies

3. Infrastructure Monitoring

zfs - migrate from pool to pool

Here are the details. cnjr-opennms>root$ zfs list NAME USED AVAIL REFER MOUNTPOINT openpool 20.6G 46.3G 35.5K /openpool openpool/ROOT 15.4G 46.3G 18K legacy openpool/ROOT/rds 15.4G 46.3G 15.3G / openpool/ROOT/rds/var 102M ... (3 Replies)
Discussion started by: pupp
3 Replies

4. Solaris

ZFS pool question

I created a pool the other day. I created a 10 gig files just for a test, then deleted it. I proceeded to create a few files systems. But for some reason the pool shows 10% full, but the files systems are both at 1%? Both files systems share the same pool. When I ls -al the pool I just... (6 Replies)
Discussion started by: mrlayance
6 Replies

5. Solaris

zfs pool migration

I need to migrate an existing raidz pool to a new raidz pool with larger disks. I need the mount points and attributes to migrate as well. What is the best procedure to accomplish this. The current pool is 6x36GB disks 202GB capacity and I am migrating to 5x 72GB disks 340GB capacity. (2 Replies)
Discussion started by: jac
2 Replies

6. Solaris

Best way to rename a ZFS Pool?

Other than export/import, is there a cleaner way to rename a pool without unmounting de FS? Something like, say "zpool rename a b"? Thanks. (2 Replies)
Discussion started by: verdepollo
2 Replies

7. Solaris

ZFS - overfilled pool

installed Solaris 11 Express on my server machine a while ago. I created a Z2 RAID over five HDDs and created a few ZFS filesystems on it. Once I (unintentionally) managed to fill the pool completely with data and (to my surprise) the filesystems stopped working - I could not read/delete any... (3 Replies)
Discussion started by: RychnD
3 Replies

8. Solaris

ZFS - Dataset / pool name are the same...cannot destroy

I messed up my pool by doing zfs send...recive So I got the following : zpool list NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT rpool 928G 17.3G 911G 1% 1.00x ONLINE - tank1 928G 35.8G 892G 3% 1.00x ONLINE - So I have "tank1" pool. zfs get all... (8 Replies)
Discussion started by: eladgrs
8 Replies

9. UNIX for Beginners Questions & Answers

Opening up ZFS pool as writable

I have installed FreeBSD onto a raw image file using QEMU Emulator successfully. I have formatted the image file using the ZFS file system (ZFS POOL). Using the following commands below I have successfully mounted the image file ready to be opened by zpool sudo losetup /dev/loop0 .img sudo... (1 Reply)
Discussion started by: alphatron150
1 Replies
DPM_ADDFS(3)						       DPM Library Functions						      DPM_ADDFS(3)

NAME
dpm_addfs - add a filesystem to a disk pool SYNOPSIS
#include <sys/types.h> #include "dpm_api.h" int dpm_addfs (char *poolname, char *server, char *fs, int status, int weight) DESCRIPTION
dpm_addfs adds a filesystem to a disk pool. poolname specifies the disk pool name previously defined using dpm_addpool. server specifies the host name of the disk server where this filesystem is mounted. fs specifies the mount point of the dedicated filesystem. status Initial status of this filesystem. It can be set to 0 or FS_DISABLED or FS_RDONLY. weight specifies the weight of the filesystem. This is used during the filesystem selection. The value must be positive. A negative value will tell the server to allocate the default weight value (1). It is recommended to use a value lower than 10. This function requires ADMIN privilege. RETURN VALUE
This routine returns 0 if the operation was successful or -1 if the operation failed. In the latter case, serrno is set appropriately. ERRORS
ENOENT Filesystem does not exist. EACCES The caller does not have ADMIN privilege. EFAULT poolname, server or fs is a NULL pointer. EEXIST this filesystem is already part of a pool. ENOMEM Memory could not be allocated for storing the filesystem definition. EINVAL The pool is unknown or the length of poolname exceeds CA_MAXPOOLNAMELEN or the length of server exceeds CA_MAXHOSTNAMELEN or the length of fs exceeds 79. SENOSHOST Host unknown. SEINTERNAL Database error. SECOMERR Communication error. LCG
$Date$ DPM_ADDFS(3)
All times are GMT -4. The time now is 11:29 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy