03-17-2009
I'm doing that now. Does it do a better job of compressing than compress?
10 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hello Everyone,
I can't transfer a large file (~15GB TAR Archive) from one linux machine to another via FTP.
I have tried the following:
1) Normal FTP the whole 15GB. This stops when it gets to about 2GB and doesn't go any further.
2) Split the 15GB file into 500MB pieces using the... (1 Reply)
Discussion started by: VVV
1 Replies
2. Shell Programming and Scripting
Hi Friends,
Getting an error while processing a very large file using an sqlloader........
The file is larger than 2 GB. Now need to change the compiler to 64-bit so that the file can be processed.
Is there any command for the same.
Thanks in advance. (1 Reply)
Discussion started by: Rohini Vijay
1 Replies
3. Shell Programming and Scripting
Hi All,
Following is the sample file
and following is the op desired
that is the last entry of each unique first field is required.
My solution is as follows
However the original file has around a million entries and around a 100,000 uniques first fields, so this soln.... (6 Replies)
Discussion started by: gauravgoel
6 Replies
4. UNIX for Dummies Questions & Answers
Hi Friends,
I'mfacing a problem while doing ftp of a large file.The control session is getting closed after sometime.But data session transfers the file successfully even when the control seeion is lost.I need to make the control session available as long as data session is active.
How can i... (1 Reply)
Discussion started by: rprajendran
1 Replies
5. UNIX for Advanced & Expert Users
We are experiencing a problem on a lengthy data transfer by FTP through a firewall. Since there are two ports in use on a ftp transfer (data and control), one sits idle while the other's transfering data. The idle port (control) will get timed out and the data transfer won't know that it's... (3 Replies)
Discussion started by: rprajendran
3 Replies
6. UNIX for Dummies Questions & Answers
Hi folks,
I have a big problem.... and need help from your experience/knowledge.
I previously install and use FREEBSD 7.0 release on my storage/backup
file server, for some reason, I can not transfer any files that is bigger
than 1GB. If I transfer it to Freebsd file server, the system... (2 Replies)
Discussion started by: bsdme2
2 Replies
7. UNIX for Dummies Questions & Answers
We have 3 Unix servers all running SVR4 Unix 1.4. I have no problems copying files to and from 2 of the servers using either the rcp command or ftp but when i come to transfer large files to the third server the copy gives up part way through and crashes this server. Copying smaller files using RCP... (7 Replies)
Discussion started by: coatesd
7 Replies
8. UNIX for Advanced & Expert Users
Hi ,
I want to transfer one file having 6GB(after compression) which is in .cpk format from one server to other server.
I tried scp command as well as FTP and also split the file then transfer the files thru scp command. At last i am facing the data lost and connection lost issue.
Generally it... (2 Replies)
Discussion started by: Sumit sarangi
2 Replies
9. Shell Programming and Scripting
hello all,
kindly i need your help, i made a script to print a specific lines from a huge file about 3 million line. the output of the script will be about 700,000 line...the problem is the script is too slow...it kept working for 5 days and the output was only 200,000 lines !!!
the script is... (16 Replies)
Discussion started by: m_wassal
16 Replies
10. Shell Programming and Scripting
Hi Experts,
I have to split huge file based on the pattern to create smaller files. The pattern which is expected in the file is:
Master.....
First...
second....
second...
third..
third...
Master...
First..
second...
third...
Master...
First...
second..
second..
second..... (2 Replies)
Discussion started by: saisanthi
2 Replies
LEARN ABOUT MINIX
dh_compress
DH_COMPRESS(1) Debhelper DH_COMPRESS(1)
NAME
dh_compress - compress files and fix symlinks in package build directories
SYNOPSIS
dh_compress [debhelperoptions] [-Xitem] [-A] [file...]
DESCRIPTION
dh_compress is a debhelper program that is responsible for compressing the files in package build directories, and makes sure that any
symlinks that pointed to the files before they were compressed are updated to point to the new files.
By default, dh_compress compresses files that Debian policy mandates should be compressed, namely all files in usr/share/info,
usr/share/man, files in usr/share/doc that are larger than 4k in size, (except the copyright file, .html and other web files, image files,
and files that appear to be already compressed based on their extensions), and all changelog files. Plus PCF fonts underneath
usr/share/fonts/X11/
FILES
debian/package.compress
These files are deprecated.
If this file exists, the default files are not compressed. Instead, the file is ran as a shell script, and all filenames that the shell
script outputs will be compressed. The shell script will be run from inside the package build directory. Note though that using -X is a
much better idea in general; you should only use a debian/package.compress file if you really need to.
OPTIONS
-Xitem, --exclude=item
Exclude files that contain item anywhere in their filename from being compressed. For example, -X.tiff will exclude TIFF files from
compression. You may use this option multiple times to build up a list of things to exclude.
-A, --all
Compress all files specified by command line parameters in ALL packages acted on.
file ...
Add these files to the list of files to compress.
CONFORMS TO
Debian policy, version 3.0
SEE ALSO
debhelper(7)
This program is a part of debhelper.
AUTHOR
Joey Hess <joeyh@debian.org>
11.1.6ubuntu2 2018-05-10 DH_COMPRESS(1)