Sponsored Content
Operating Systems Solaris Splitting rpool mirror disk in ZFS Post 302825623 by solaris_1977 on Monday 24th of June 2013 07:34:51 PM
Old 06-24-2013
This function is starting with release 9, while my release is 7.
If I shut down server and remove disk (physically) from server and then bring up server, that should work ?
And if this works, how will I attach (resynch) both disk after my activity got completed and I reinserted disk inside server ?
 

9 More Discussions You Might Find Interesting

1. Solaris

zfs mirror disk

Is it possible to create a Mirror with zfs ?? I'm experimented user with Solstice Disk suite. Or Sun Volume manager or veritas volume manager. But, i would like switch from Disksuite to Zfs. All my mirrored disks. (1 Reply)
Discussion started by: simquest
1 Replies

2. Solaris

ZFS Mirror versus Hardware Mirror

I've looked a little but haven't found a solid answer, assuming there is one. What's better, hardware mirroring or ZFS mirroring? Common practice for us was to use the raid controllers on the Sun x86 servers. Now we've been using ZFS mirroring since U6. Any performance difference? Any other... (3 Replies)
Discussion started by: Lespaul20
3 Replies

3. Solaris

ZFS rpool physical disk move

I'd like to finish setting up this system and then move the secondary or primary disk to another system that is the exact same hardware. I've done things like this in the past with ufs and disk suite mirroring just fine. But I have yet to do it with a zfs root pool mirror. Are there any... (1 Reply)
Discussion started by: Metasin
1 Replies

4. AIX

Clone or mirror your AIX OS larger disk to smaller disk ?

hello folks, I have a 300GB ROOTVG volume groups with one filesystem /backup having 200GB allocated space Now, I cannot alt disk clone or mirrorvg this hdisk with another smaller disk. The disk size has to be 300GB; I tried alt disk clone and mirrorvg , it doesn't work. you cannot copy LVs as... (9 Replies)
Discussion started by: filosophizer
9 Replies

5. HP-UX

What is the difference between DRD and Root Mirror Disk using LVM mirror ?

what is the difference between DRD and Root Mirror Disk using LVM mirror ? (3 Replies)
Discussion started by: maxim42
3 Replies

6. Solaris

Zfs rpool size

Hi everyone, I am doing housekeeping of my Solaris 11 for zfs snapshot to reduce the snapshot size. I have already cleared the / file system, however the rpool size still not reduced. Filesystem Size Used Available Capacity Mounted on rpool/ROOT/solaris-2 98G 6.9G ... (2 Replies)
Discussion started by: freshmeat
2 Replies

7. Solaris

Need help to understand zfs rpool space allocation

Hi, I am unable to understand that, in one of my servers while df -kh Filesystem Size Used Available Capacity Mounted on rpool/ROOT/solaris-2 98G 11G 29G 29% / Even the Root FS filled on 40gb and system becomes unstable. it is showing... (4 Replies)
Discussion started by: anuragr
4 Replies

8. Solaris

ZFS: /system/zones "respawning" on rpool

Hi, I have this fresh installation of Solaris 11.3 sparc. I have two zfs pools both using two disks in mirroring mode, both are online. I want to move /system/zones, currently rpool/VARSHARE/zones, from rpool to the other zfs pool so my zones don't consume space on the disks allocated to... (1 Reply)
Discussion started by: X96
1 Replies

9. Solaris

Trying to create ZFS slice on rpool

I have a 240GB disk as rpool. I have installed Solaris 11.3 to a partition which is 110GB. Now I have another 130GB which is unallocated. I want to use that additional space as a temporary folder to be shared between Solaris and Linux. The additional space had no /dev/dsk/c2t4... entry so I used... (8 Replies)
Discussion started by: kebabbert
8 Replies
CFS(4)							     Kernel Interfaces Manual							    CFS(4)

NAME
cfs - cache file system SYNOPSIS
cfs -s [-rd] [-f partition] cfs -a netaddr [-rd] [-f partition] [mtpt] DESCRIPTION
Cfs is a user-level file server that caches information about remote files onto a local disk. It is normally started by the kernel at boot time, though users may start it manually. Cfs is interposed between the kernel and a network connection to a remote file server to improve the efficiency of access across slow network connections such as modem lines. On each open of a file cfs checks the consistency of cached information and discards any old information for that file. Cfs mounts onto mtpt (default /) after connecting to the file server. The options are: s the connection to the remote file server is on file descriptors 0 and 1. a netaddr dial the destination netaddr to connect to a remote file server. r reformat the cache disk partition. d turn on debugging f partition use file partition as the cache disk partition. All 9P messages except read, clone, and walk (see intro(5)) are passed through cfs unchanged to the remote server. A clone followed imme- diately by a walk is converted into a clwalk. If possible, a read is satisfied by cached data. Otherwise, the file server is queried for any missing data. FILES
/dev/hd0cache Default file used for storing cached data. SOURCE
/sys/src/cmd/cfs CFS(4)
All times are GMT -4. The time now is 04:38 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy