09-02-2013
I don't have the original /etc/filesystems
I am contemplating recreating the VG (loosing all data)
10 More Discussions You Might Find Interesting
1. Programming
Hello,
I want to know what call to use to read the details of a process control block in solaris ?:) (2 Replies)
Discussion started by: hmurali
2 Replies
2. UNIX for Dummies Questions & Answers
Hmm, how to ask this without sounding too malicious...
How might one go about causing a disk corruption in OS X specifically or via the command line in UNIX in general?
Doesnt matter the severity of the problem, I just want to scare the person a little, then fix the problem for them.
Any... (1 Reply)
Discussion started by: Yummator
1 Replies
3. AIX
I don't understand the concept behind importvg and exportvg. What are some examples of when one would use this? (1 Reply)
Discussion started by: outtacontrol
1 Replies
4. Solaris
I need to corrupt a superblock of a mounted device in a soalris m/c and check recovery from an alternate superblock. How can this be done? (2 Replies)
Discussion started by: sujathan
2 Replies
5. AIX
Hello,
On Aix 5.3, during importvg, the varyonvg fails:
importvg -y vgtest hdisk20
0516-013 varyonvg: The volume group cannot be varied on because
there are no good copies of the descriptor area.
When i use manually the command varyonvg -u -b -t vgtest to force, the vg can... (3 Replies)
Discussion started by: dantares
3 Replies
6. AIX
Hi
How to check if a vg is exported from an LPAR using exportvg?
Is there a command or place where we can see the list of the vg's exported from the LPAR ?
Because if i run the command to varyonvg i get the below error :
root@cbspsrc02 #varyonvg old_rootvg
0516-082 varyonvg: Unable to... (7 Replies)
Discussion started by: newtoaixos
7 Replies
7. HP-UX
Hallo Friends,
I have application X running on hpux 11.11 and oracle 9i release 2. I recently had a hardware failure on disk /dev/dsk/c2t0d0
Below is the systemlog file :
root@a7dmc:/var/adm/syslog > /opt/resmon/bin/resdata -R 155713541 -r /storage/events/enclosures/gazemon/0_1_1_0.0.0... (11 Replies)
Discussion started by: kekanap
11 Replies
8. Shell Programming and Scripting
I don't know if I am asking this correctly, but I have a hard drive with some bad sectors and it appears that some of the data is corrupt. I am having allot of trouble copying the data to a new drive. The issue is not in copying files, but that the new drive to which files are copied is not acting... (17 Replies)
Discussion started by: LMHmedchem
17 Replies
9. AIX
Is there a way to find the volume group name after assigning hdisk to the machine for importing it ?
importvg -y <find_the_name_from_the_unknown_disk_> hdiskX
Code tags and no hijacking. You should be aware of that ;) (3 Replies)
Discussion started by: filosophizer
3 Replies
10. High Performance Computing
I am managing a linux cluster which has been build on Platform Cluster Manager PCM 1.2.1) from IBM Platform Computing. Unfortunately somebody deteled data files of postgresql from /var/lib directory. I somehow managed to start the postmaster service again, but all the administrative commands of... (2 Replies)
Discussion started by: ahsanpmd
2 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)
NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS
--predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO
bup-midx(1), bup-save(1)
BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown- bup-margin(1)