09-21-2012
Quote:
Originally Posted by
bstring
Also, does anyone know what filesystems are natively supported in FBSD 6.x and 8.x? I believe 6.x supports ufs and 8.x supports ufs and zfs, but I am not positive.
ZFS is a combined file system and logical volume manager originally designed by Sun Microsystems. It was ported to FreeBSD and has been part of the operating system since FreeBSD 7.0. FreeBSD 6.x use UFS.
The Z File System
10 More Discussions You Might Find Interesting
1. Solaris
Hi all
I plan to install Solaris 10U6 on some SPARC server using ZFS as root pool, whereas I would like to keep the current setup done by VxVM:
- 2 internal disks: c0t0d0 and c0t1d0
- bootable root-volume (mirrored, both disks)
- 1 non-mirrored swap slice
- 1 non-mirrored slices for Live... (1 Reply)
Discussion started by: blicki
1 Replies
2. Solaris
# zpool import
pool: emcpool1
id: 5596268873059055768
state: UNAVAIL
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
devices and try again.
see: Sun Message ID: ZFS-8000-3C
config:
emcpool1 ... (7 Replies)
Discussion started by: fugitive
7 Replies
3. Infrastructure Monitoring
Here are the details.
cnjr-opennms>root$ zfs list
NAME USED AVAIL REFER MOUNTPOINT
openpool 20.6G 46.3G 35.5K /openpool
openpool/ROOT 15.4G 46.3G 18K legacy
openpool/ROOT/rds 15.4G 46.3G 15.3G /
openpool/ROOT/rds/var 102M ... (3 Replies)
Discussion started by: pupp
3 Replies
4. Solaris
I need to migrate an existing raidz pool to a new raidz pool with larger disks. I need the mount points and attributes to migrate as well. What is the best procedure to accomplish this. The current pool is 6x36GB disks 202GB capacity and I am migrating to 5x 72GB disks 340GB capacity. (2 Replies)
Discussion started by: jac
2 Replies
5. Solaris
Other than export/import, is there a cleaner way to rename a pool without unmounting de FS?
Something like, say "zpool rename a b"?
Thanks. (2 Replies)
Discussion started by: verdepollo
2 Replies
6. Solaris
Hi All,
I want to write a script to create flar images on multiple servers. In non zfs filesystem I am using -X option to refer a file to exclude mounts on different servers.
but on ZFS -X option is not working. I want multiple mounts to be ignore on ZFS base system during flarecreate.
I... (0 Replies)
Discussion started by: uxravi
0 Replies
7. Solaris
installed Solaris 11 Express on my server machine a while ago. I created a Z2 RAID over five HDDs and created a few ZFS filesystems on it.
Once I (unintentionally) managed to fill the pool completely with data and (to my surprise) the filesystems stopped working - I could not read/delete any... (3 Replies)
Discussion started by: RychnD
3 Replies
8. Solaris
I messed up my pool by doing zfs send...recive So I got the following :
zpool list
NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
rpool 928G 17.3G 911G 1% 1.00x ONLINE -
tank1 928G 35.8G 892G 3% 1.00x ONLINE -
So I have "tank1" pool.
zfs get all... (8 Replies)
Discussion started by: eladgrs
8 Replies
9. Solaris
I have a single zpool with 3 2-way mirrors ( 3 x 2 way vdevs) it has a degraded disk in mirror-2, I know I can suffer a single drive failure, but looking at this how many drive failures can this suffer before it is no good? On the face of it, I thought that I could lose a further 2 drives in each... (4 Replies)
Discussion started by: fishface
4 Replies
10. Solaris
On an OmniOS server, I removed a single-disk pool I was using for testing.
Now, when I run zpool import it will show it as FAULTED, since that single disk not available anymore.
# zpool import
pool: fido
id: 7452075738474086658
state: FAULTED
status: The pool was last... (11 Replies)
Discussion started by: priyadarshan
11 Replies
LEARN ABOUT FREEBSD
cgread
CGREAD(3) BSD Library Functions Manual CGREAD(3)
NAME
cgread, cgread1, cgwrite1 -- read/write cylinder groups of UFS disks
LIBRARY
UFS File System Access Library (libufs, -lufs)
SYNOPSIS
#include <sys/param.h>
#include <sys/mount.h>
#include <ufs/ufs/ufsmount.h>
#include <ufs/ufs/dinode.h>
#include <ufs/ffs/fs.h>
#include <libufs.h>
int
cgread(struct uufsd *disk);
int
cgread1(struct uufsd *disk, int c);
int
cgwrite1(struct uufsd *disk, int c);
DESCRIPTION
The cgread() and cgread1() functions provide cylinder group reads for libufs(3) consumers. The cgread1() function reads from one cylinder
group, specified by c into the d_cg field of a userland UFS disk structure. It sets the d_lcg field to the cylinder group number c.
The cgread() function operates on sequential cylinder groups. Calling the cgread() function is equivalent to calling cgread1() with a cylin-
der group specifier equivalent to the value of the current d_ccg field, and then incrementing the d_ccg field.
The cgwrite1() function stores cylinder group specified by c from d_cg field of a userland UFS disk structure on disk.
RETURN VALUES
Both functions return 0 if there are no more cylinder groups to read, 1 if there are more cylinder groups, and -1 on error.
ERRORS
The function cgread() may fail and set errno for any of the errors specified for the library function bread(3).
The function cgread1() has semantically identical failure conditions to those of cgread().
The function cgwrite1() may fail and set errno for any of the errors specified for the library function bwrite(3).
SEE ALSO
bread(3), bwrite(3), libufs(3)
HISTORY
These functions first appeared as part of libufs(3) in FreeBSD 5.1.
AUTHORS
Juli Mallett <jmallett@FreeBSD.org>
BSD
June 4, 2003 BSD