So this is confirmed, your file system is ext3 with block size 1 KB, and this limits file size to 16 GB.
You have to do your backup to another file system with a bigger block size. Or stream it to a tape.
--
Or recreate the /home file system with a bigger block size.
This would delete all your existing data, so you need a full backup before you do it!
Nowerdays mkfs should default to a 4 KB block size, and file size limit is 64 GB.
Or define explicit options as shown by
Hi all,
I want to set a size limitation for some user in the system, for an example, each user only have 5MB free space in the system. The user cannot user more than 5 MB space.
Is it possible to do this? Thanks! (1 Reply)
Hi All,
Can anyone please clarify me the following questions:
1. Is there any file size limitation in HP-UX 11i, that I can able to create upto certain size of file (say 2 GB) and not more then that????
2. At max. how many files we can able to keep inside a folder????
3. How many... (2 Replies)
hi ,
iam trying to sort millions of records which is delimited and i cant able to
use sort command more than 60 million..if i try to do so i got an message stating that "File size limit exceeded",Is there any file size limit for using sort command..
How can i solve this problem.
thanks
... (7 Replies)
Hi
I am trying to rcp a file from Solaris box to Linux.
When the file size is 2,205,255,047, the rcp fails with the message
Jan 10 01:11:53 hqsas167 rsh: pam_authenticate: error Authentication failed
However when I rcp a file with smaller size - 9,434,477 - the rcp completes with... (2 Replies)
Hi to every body there,
I am new this forum and this is my first post.
I am a new user of Unix,
is there any size limitation of files while creating tar file.
Thanks in advance (4 Replies)
Hi,
I am using fetchmail in my application so as to download mails to the localhost where the application is hosted from the mailserver.Fetchmail is configured as as to run as a daemon polling mails during an interval of 1sec.
So my concern here is, during each 2sec it is writing two... (10 Replies)
Hello All,
I am using a SunOS machine. My application creates output files for the downstream systems. However output files are restricted to 2GB of file size in SunOS due to which I am forced to create multiple files which is not supported by the downstream due to some limitations.
Is... (5 Replies)
Hi friends,
I tried to take a backup of my PC using tar command. But it ended with an error
tar: /home/backup/back.tar.gz: Cannot write: No space left on device
tar: Error is not recoverable: exiting now
But i checked the disk space and there is enough space is available.
]# df
Filesystem... (11 Replies)
Discussion started by: siva3492
11 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)