04-19-2013
Great thanks for the quick reply.
9 More Discussions You Might Find Interesting
1. Filesystems, Disks and Memory
Hello, I have 4 disks to add on an Unix OS. 2 of them are already added. When I try to make the diskadd command, the system reply's that Diskadd was "Not Found". I already tried to go to the /etc/dskmgmt/ and call there the diskadd, but the system gives me the same error message.
Can anybody help... (4 Replies)
Discussion started by: luisjdm
4 Replies
2. Solaris
Hi, boss
i have a question about the system add the new disk. I add 6 disks in Sunfire 880, i want to put these 6 disks in to service but i do not want re-install the system, some guys can give me some advie on this case?
thanks very much
the follow is detail
0. c1t0d0 <SUN72G cyl... (3 Replies)
Discussion started by: surainbow
3 Replies
3. Solaris
Previously , i remove the disk by
#vxdg -g testdg -k rmdisk testdg02
But i got error when i -k adddisk
bash-2.03# vxdisk list
DEVICE TYPE DISK GROUP STATUS
c0t0d0s2 auto:none - - online invalid
c0t1d0s2 auto:none ... (1 Reply)
Discussion started by: waibabe
1 Replies
4. Filesystems, Disks and Memory
This may sound simple, but I cant find anything on Google for just adding a new disk.
This is a data disk, not boot or anything like that. I've got 6 disks in there, one of them almost dead, and I need to copy what I can from the dead one to the new one.
My problem is that when I go to... (3 Replies)
Discussion started by: LP4Ever
3 Replies
5. Solaris
Hi,
Anyone can help me, just want to confirm, if possible adding a disk storage that we have a RAWDATA with +ASM in our database. If possible, what would you recommend "workaround used" from us to do and the preparation. And also how long the downtime will take.
Our disk continuous to grow,... (0 Replies)
Discussion started by: fpalero
0 Replies
6. Solaris
I added a disk to a zpool using "zpool add diskname"
My intention was mirror a zpool disk that no mirror; that is a zpool with only one disk. I did not issue the right command. Now, the disk has been added successfully but I cannot remove nor detach it as Solaris 11 thinks it has data on it... (14 Replies)
Discussion started by: LittleLebowski
14 Replies
7. Solaris
Hello,
I need some help in recovering ZFS pool. Here is scenerio. There are two disks -
c0t0d0 - This is good disk. I cloned it from other server and boot server from this disk.
c0t1d0 - This is original disk of this server, having errors. I am able to mount it on /mnt. So that I can copy... (1 Reply)
Discussion started by: solaris_1977
1 Replies
8. Solaris
issue,
I had a zpool which was full
pool_temp1 199G 197G 1.56G 99% ONLINE -
pool_temp2 199G 196G 3.09G 98% ONLINE -
as you can see, full
so I replaced with a larger disk.
zpool replace pool_temp1 c3t600144F0FF8BA036000058CC1DB80008d0s0... (2 Replies)
Discussion started by: rrodgers
2 Replies
9. Solaris
On an OmniOS server, I removed a single-disk pool I was using for testing.
Now, when I run zpool import it will show it as FAULTED, since that single disk not available anymore.
# zpool import
pool: fido
id: 7452075738474086658
state: FAULTED
status: The pool was last... (11 Replies)
Discussion started by: priyadarshan
11 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)
NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS
--predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO
bup-midx(1), bup-save(1)
BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown- bup-margin(1)