How do I find files in current folder only?
We are on AIX 5.3, so maxdepth is not supported.
I tried to do this
find /dir1/dir2/dir3/dir4 -prune -type f
to display all files in /dir1/dir2/dir3/dir4 only but it does not show any files.
Somehow the -prune option works for dir3 level... (7 Replies)
hi,
need help in writing a script in perl.
requirement :
1. Search for the files in a particular folder
2. search for a string in the file names
3. Delete the file which matches the string.
Ex: if the folder is C:\TEST
and the folder has 5 files like
2009ABCG.txt
2009MNO.txt... (2 Replies)
Hi all,
We have an existing script:
find /u03/oraprod/perpcomn/admin/out -type f -ctime +7 \
-exec cp {} "/u08/oraprod/backup/cout" \;
Which is to find all files more than 7 days and copy to another folder. However I would like to only list files with Sep 29, and cp to another folder.
... (2 Replies)
Hi all,
*I use Uwin and Cygwin emulator.
I´m trying to search for all text files in the current folder (C/Files) and its sub folders using
find -depth -name "*.txt"
The above command worked for me, but now I would like to copy all found text files to a new folder (C/Files/Text) with ... (4 Replies)
I have installed ubuntu. And I have create users ans groups.
Suppose if the user enter into through Putty SSH. He should have access only to home folder and cannot move to other than $HOME. User should not able to root files and /$ files.
Kindly provide solution.
Regards
Vasanth kumar (3 Replies)
To find a word from multiple level files:
"find . -type f -exec grep {} +" is working on UNIX machines but not working on Linux machine. What is the equivalent command on Linux to find the word from multiple level files?
Input is appreciated. (3 Replies)
Hi All,
I have a requirement to monitor the sub-directories under /home in a way that if the the folder size increases by 30 GB in a span of like an hour then it needs to send email alerts listing what as the actual size was and what's the current size which the subject listing the sub-directory... (25 Replies)
Hi All,
So I have another question. I'm trying to search for files with a certain extension and then move all of them up one level in the folder hierarchy.
So something like this:
original: /path/to/file/test.txt
after: /path/to/test.txt
I had some great help recently with another... (4 Replies)
Could it be possible to find common lines between all of the files in one folder? Just like comm -12 . So all of the files two at a time. I would like all of the outcomes to be written to a different files, and the file names could be simply numbers - 1 , 2 , 3 etc. All of the file names contain... (19 Replies)
Discussion started by: Eve
19 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)