8 More Discussions You Might Find Interesting
1. UNIX for Beginners Questions & Answers
Dear Users,
Appreciate your help if you could help me with splitting a large file > 1 million lines with sed or awk. below is the text in the file
input file.txt
scaffold1 928 929 C/T +
scaffold1 942 943 G/C +
scaffold1 959 960 C/T +... (6 Replies)
Discussion started by: kapr0001
6 Replies
2. Linux
Hi,
Anyone can help, I have a large textfile (one file), and I need to split into multiple file to break each file into ^L.
My textfile
==========
abc company
abc address
abc contact
^L
my company
my address
my contact
my skills
^L
your company
your address
========== (3 Replies)
Discussion started by: fspalero
3 Replies
3. UNIX for Dummies Questions & Answers
Hello....
I have two servers, one has an empty / and the other has a subdirectory with a large number (4 gig) with many, many files. I need a way to transfer the files en masse from the server with the large number of files to the one that is essentially blank.
I don't have space on the used... (16 Replies)
Discussion started by: blaine.miller
16 Replies
4. UNIX for Advanced & Expert Users
Hi ,
I want to transfer one file having 6GB(after compression) which is in .cpk format from one server to other server.
I tried scp command as well as FTP and also split the file then transfer the files thru scp command. At last i am facing the data lost and connection lost issue.
Generally it... (2 Replies)
Discussion started by: Sumit sarangi
2 Replies
5. Shell Programming and Scripting
Hello Gurus,
We are facing some performance issue in UNIX. If someone had faced such kind of issue in past please provide your suggestions on this .
Problem Definition:
/Few of load processes of our Finance Application are facing issue in UNIX when they uses a shell script having below... (19 Replies)
Discussion started by: KRAMA
19 Replies
6. UNIX for Dummies Questions & Answers
Hi folks,
I have a big problem.... and need help from your experience/knowledge.
I previously install and use FREEBSD 7.0 release on my storage/backup
file server, for some reason, I can not transfer any files that is bigger
than 1GB. If I transfer it to Freebsd file server, the system... (2 Replies)
Discussion started by: bsdme2
2 Replies
7. Filesystems, Disks and Memory
Hi, Im trying to take a database backup. one of the files is 26 GB. I am using cp -pr to create a backup copy of the database. after the copying is complete, if i do du -hrs on the folders i saw a difference of 2GB.
The weird fact is that the BACKUP folder was 2 GB more than the original one!
... (1 Reply)
Discussion started by: 0ktalmagik
1 Replies
8. UNIX for Advanced & Expert Users
Hello Everyone,
I can't transfer a large file (~15GB TAR Archive) from one linux machine to another via FTP.
I have tried the following:
1) Normal FTP the whole 15GB. This stops when it gets to about 2GB and doesn't go any further.
2) Split the 15GB file into 500MB pieces using the... (1 Reply)
Discussion started by: VVV
1 Replies