04-01-2011
Hi!
Quote:
Originally Posted by
jim mcnamara
how about cksum
Well,
cksum is to slow. There can be files with > 2GB. And I want to scan also all subdirectories. The sum of the file size of all duplicated files is not important.
Dirk
10 More Discussions You Might Find Interesting
1. Shell Programming and Scripting
hello all
I like to make search on files , and the result need to be the files that are duplicated? (8 Replies)
Discussion started by: umen
8 Replies
2. Solaris
hi all,
in my server there are some specific application files which are spread through out the server... these are spread in folders..sub-folders..chid folders...
please help me, how can i find the total size of these specific files in the server... (3 Replies)
Discussion started by: abhinov
3 Replies
3. Shell Programming and Scripting
I have a directory with images:
-rw-r--r-- 1 root root 26216 Mar 19 21:00 020109.210001.jpg
-rw-r--r-- 1 root root 21760 Mar 19 21:15 020109.211502.jpg
-rw-r--r-- 1 root root 23144 Mar 19 21:30 020109.213002.jpg
-rw-r--r-- 1 root root 31350 Mar 20 00:45 020109.004501.jpg
-rw-r--r-- 1 root... (2 Replies)
Discussion started by: Ikon
2 Replies
4. Shell Programming and Scripting
What utility do you recommend for simply finding all duplicate files among all files? (4 Replies)
Discussion started by: kiasas
4 Replies
5. Shell Programming and Scripting
Hi,
Could anyone help me to solve this problem?
I have two files "f1" and "f2" having 2 fields in each, a) file size and b) file name. The data are almost same in both the files except for few and new additional lines. Now, I have to find out and print the output as, the difference in the... (3 Replies)
Discussion started by: royalibrahim
3 Replies
6. Shell Programming and Scripting
hi,
Please help me to write a command to delete duplicate lines from a file. And the size of file is 50 MB. How to remove duplicate lins from such a big file. (6 Replies)
Discussion started by: vsachan
6 Replies
7. Shell Programming and Scripting
I have more than 100 files like this:
SVEAVLTGPYGYT 2
SVEGNFEETQY 10
SVELGQGYEQY 28
SVERTGTGYT 6
SVGLADYNEQF 21
SVGQGYEQY 32
SVKTVLGYEQF 2
SVNNEQF 12
SVRDGLTNSPLH 3
SVRRDREGLEQF 11
SVRTSGSYEQY 17
SVSVSGSPLQETQY 78
SVVHSTSPEAF 59
SVVPGNGYT 75 (4 Replies)
Discussion started by: xshang
4 Replies
8. Shell Programming and Scripting
Hi !
I wonder if anyone can help on this : I have a directory: /xyz that has the following files:
chsLog.107.20130603.gz
chsLog.115.20130603
chsLog.111.20130603.gz
chsLog.107.20130603
chsLog.115.20130603.gz
As you ca see there are two files that are the same but only with a minor... (10 Replies)
Discussion started by: fretagi
10 Replies
9. Shell Programming and Scripting
Hi champs,
I have one of the requirement, where I need to compare two files line by line and ignore duplicates. Note, I hav files in sorted order.
I have tried using the comm command, but its not working for my scenario.
Input file1
srv1..development..employee..empname,empid,empdesg... (1 Reply)
Discussion started by: Selva_2507
1 Replies
10. Shell Programming and Scripting
Hello,
I have a huge directory (with millions of files) and need to find out duplicates based on BOTH file name and File size.
I know fdupes but it calculates MD5 which is very time-consuming and especially it takes forever as I have millions of files.
Can anyone please suggest a script or... (7 Replies)
Discussion started by: prvnrk
7 Replies
sum(1) General Commands Manual sum(1)
NAME
sum - print checksum and block or byte count of file(s)
SYNOPSIS
[file ...]
Remarks
is obsolescent and should not be used in new applications that are intended to be portable between systems. Use instead (see cksum(1)).
DESCRIPTION
calculates and prints to standard output a checksum for each named file, and also prints the size of the file in 512 byte blocks, rounded
up.
The default algorithm is a 16-bit sum of the bytes in which overflow is ignored. Alternate algorithms can be selected with the and
options.
Standard input is used if no file names are given.
is typically used to verify data integrity when copying files between systems.
Options
recognizes the following options:
Use an alternate algorithm in which the 16-bit sum is right rotated
with each byte in computing the checksum.
Use the 32-bit cyclical redundancy check (CRC) algorithm used by
RETURN VALUE
returns the following values upon completion:
All files were processed successfully.
One or more files could not be read or some other error occurred.
If an inaccessible file is encountered, continues processing any remaining files, but the final exit status is affected.
DIAGNOSTICS
Read error conditions are indistinguishable from end of file on most devices; check the block or byte count.
WARNINGS
This command is likely to be withdrawn from X/Open standards. Applications using this command might not be portable to other vendors'
platforms. The usage of cksum(1) is recommended.
SEE ALSO
cksum(1), wc(1).
STANDARDS CONFORMANCE
sum(1)