12-03-2009
thanks a lot, skmdu
Your perl script worked perfectly for my input file
Thanks a lot.
Besides that, can I ask you that how can I edit my perl script to let it only run my "program_name_count_length" program to archive the same goal?
Thanks again, skmdu ^^
10 More Discussions You Might Find Interesting
1. Shell Programming and Scripting
Friends,
I have to write a shell script,the description is----
i Have to check the uniqueness of the numbers in a file.
A file is containing 200thousand tickets and a ticket have 15 numbers in asecending order.And there is a strip that is having 6 tickets that means 90 numbers.I... (7 Replies)
Discussion started by: namishtiwari
7 Replies
2. Shell Programming and Scripting
Input file data contents:
>seq_1
MSNQSPPQSQRPGHSHSHSHSHAGLASSTSSHSNPSANASYNLNGPRTGGDQRYRASVDA
>seq_2
AGAAGRGWGRDVTAAASPNPRNGGGRPASDLLSVGNAGGQASFASPETIDRWFEDLQHYE
>seq_3
ATLEEMAAASLDANFKEELSAIEQWFRVLSEAERTAALYSLLQSSTQVQMRFFVTVLQQM
ARADPITALLSPANPGQASMEAQMDAKLAAMGLKSPASPAVRQYARQSLSGDTYLSPHSA... (7 Replies)
Discussion started by: patrick87
7 Replies
3. Shell Programming and Scripting
Hello Everyone,
I have a perl script that reads two types of data files (txt and XML). These data files are huge and large in number. I am using something like this :
foreach my $t (@text)
{
open TEXT, $t or die "Cannot open $t for reading: $!\n";
while(my $line=<TEXT>){
... (4 Replies)
Discussion started by: ad23
4 Replies
4. Shell Programming and Scripting
I’m new to Linux script and not sure how to filter out bad records from huge flat files (over 1.3GB each). The delimiter is a semi colon “;”
Here is the sample of 5 lines in the file:
Name1;phone1;address1;city1;state1;zipcode1
Name2;phone2;address2;city2;state2;zipcode2;comment... (7 Replies)
Discussion started by: lv99
7 Replies
5. AIX
Hi
When i copy 300GB of data from one filesystem to the other filesystem in AIX I get the error :
tar: 0511-825 The file 'SAPBRD.dat' is too large.
The command I used is :
# tar -cf - . | (cd /sapbackup ; tar -xf - )
im copying as root
The below is my ulimit -a output :
... (3 Replies)
Discussion started by: samsungsamsung
3 Replies
6. Shell Programming and Scripting
into small files. i need to add a head.txt and tail.txt into small files at the begin and end, and give a name as q1.xml q2.xml q3.xml ....
thank you very much. (2 Replies)
Discussion started by: dtdt
2 Replies
7. UNIX for Dummies Questions & Answers
Hi,
I have a Huge 7 GB file which has around 1 million records, i want to split this file into 4 files to contain around 250k messages each.
Please help me as Split command cannot work here as it might miss tags..
Format of the file is as below
<!--###### ###### START-->... (6 Replies)
Discussion started by: KishM
6 Replies
8. Shell Programming and Scripting
We have a folder XYZ with large number of files (>350,000). how can i split the folder and create say 10 of them XYZ1 to XYZ10 with 35,000 files each. (doesnt matter which files go where). (12 Replies)
Discussion started by: AlokKumbhare
12 Replies
9. UNIX for Advanced & Expert Users
Optimization shell/awk script to aggregate (sum) for all the columns of Huge data file
File delimiter "|"
Need to have Sum of all columns, with column number : aggregation (summation) for each column
File not having the header
Like below -
Column 1 "Total
Column 2 : "Total
...
...... (2 Replies)
Discussion started by: kartikirans
2 Replies
10. Solaris
Gents
I have huge NAS File System as /sys with size 10 TB and I want to Split each 1TB in spirit File System to be mounted in the server.
How to can I do that without changing anything in the source.
Please your support. (1 Reply)
Discussion started by: AbuAliiiiiiiiii
1 Replies
LEARN ABOUT DEBIAN
speedy_backend
SPEEDY_BACKEND(1p) SPEEDY_BACKEND(1p)
NAME
speedy_backend - the backend process for a persistent Perl interpreter
SYNOPSIS
none ; this program is not meant to be called directly.
DESCRIPTION
speedy, short for SpeedyCGI, is a way to run perl scripts persistently, which can make them run much more quickly.
After the script is initially run, instead of exiting, the perl interpreter is kept running inside a backend process, speedy_backend. Dur-
ing subsequent runs, this interpreter is used to handle new executions instead of starting a new perl interpreter each time. A very fast
frontend program, speedy, written in C, is executed for each request. This fast frontend then contacts the persistent Perl process, which
is usually already running, to do the work and return the results.
Each perl script runs in its own Unix process, so one perl script can't interfere with another. Command line options can also be used to
deal with programs that have memory leaks or other problems that might keep them from otherwise running persistently.
The speedy front end connects to the back end process, speedy_backend, via a Unix socket in /tmp. A queue is kept in a shared file in /tmp
that holds an entry for each process. In that queue are the pids of the perl processes waiting for connections. The frontend pulls a
process out of this queue, connects to its socket, sends over the environment and argv, and then uses this socket for stdin/stdout to the
perl process.
FILES
/tmp/speedy* A unix socket used to connect to the frontend process.
AUTHOR
Sam Horrocks
http://daemoninc.com
sam@daemoninc.com
NOTES
This manual page was created by Niko Tyni <ntyni@iki.fi> for Debian GNU/Linux, because the original program does not have one. It is based
on the original and more complete CGI::SpeedyCGI(3pm) manual page.
SEE ALSO
perl(1), CGI::SpeedyCGI(3pm), speedy(1)
SPEEDY_BACKEND(1p)