I made a program that extracts quotes while retaining special inner quotes (in this case an 'x' followed by an apostrophe). The original program is far more complicated than this, but I wanted to make it simple to troubleshoot.
I want to take these two perl commands and have the first command's results be piped into the second commands input but while only running perl once:
Input: ['Say, x'Hix'']
Output: Say, x'Hix'
Taking the output of the first command:
Result (proper): Say, 'Hi'
I realise that I could easily just run perl twice and pipe them into eachother, but running perl twice seems inefficient; especially given that this command is ran thousands of times:
I've tried combining both commands in the code below, but it doesn't seem to be taking the output of the first command as input for the second:
Result (not what I want):
What is the correct format for a single command that would combine portions of 2 different lines in the command history? I'm using a C shell. Here's a simplified command history to clarify:
4 rm file1
5 ls -ld file2 file3 file4
6 cat file 5
With the above history, what would be the... (5 Replies)
I would like to change the lines:
originalline1
originalline2
to:
originalline1new
originalline1newline
originalline2new
originalline2newline
To do this, id like to combine the commands:
sed 's/^/&new/g' file > newfile1
and
sed '/^/ a\\
newline\\
\\ (2 Replies)
Hello all,
I am trying to list and count all the files of a particular type in any given directory. I can use the commands separately but when I combine them they do not give an output.
The command for counting the files is ls -1 | wc -l and for listing all the file of particular type say... (2 Replies)
Is there anyway to achieve "find /home -name "*.bashrc" 2>/dev/null" and "PS1="\n>"" in the same command? I just wanna add a line to the previous command to change the PS1 variable to ">". (1 Reply)
Hi Guys,
I am looking to optimze these 5 SSH lines to a single SSH to get my machine to not hang! lol!
cat hosts.lst | xargs -n1 -t -i echo 'home/util/timeout 6 0 ssh -q {} top -b > util/{}.top &' >> r_query_info
cat hosts.lst | xargs -n1 -t -i echo 'home/util/timeout 6 0 ssh -q {} uname -r... (5 Replies)
hey can anyone tell me how can i combine these two commands so that it is executed only once, but gives me both the results.
IDLE=`sar 30 6 | grep Average | awk '{print $1 $5}' `
sar 30 120 | awk '{print $1" "$5}' >> mailx -m -s "$MSG" xyz@abc.com. (5 Replies)
Hi,
I have a directory with some XML files in it. I can use wildcards to get the list of XMLs I want
say I have following XMLs in same dir
Employee1.xml
Employee2.xml
Employee3.xml
and
Salary1.xml
Salary2.xml
Salary3.xml
apart from other .txt .dat files etc
I want to write a unix... (7 Replies)
I have a directory of 3000 files without extensions (Solaris 5.10).
I would like to iterate the file names through the 'file' command and output their mime types (most are pdf or jpg, but a very few might be psd or swf which show simply as 'data')
So, I would like the output of the 'ls'... (2 Replies)
Hello all,
I need to send an attachment and text in the body, both in the same Email.
Below are two cammand that send the required data in separate Emails. I need to combine them so that I get just 1 Email containing the attachment & text in the body.
uuencode ${filename} "${file_}" |... (6 Replies)
Discussion started by: Junaid Subhani
6 Replies
LEARN ABOUT DEBIAN
speedy_backend
SPEEDY_BACKEND(1p)SPEEDY_BACKEND(1p)NAME
speedy_backend - the backend process for a persistent Perl interpreter
SYNOPSIS
none ; this program is not meant to be called directly.
DESCRIPTION
speedy, short for SpeedyCGI, is a way to run perl scripts persistently, which can make them run much more quickly.
After the script is initially run, instead of exiting, the perl interpreter is kept running inside a backend process, speedy_backend. Dur-
ing subsequent runs, this interpreter is used to handle new executions instead of starting a new perl interpreter each time. A very fast
frontend program, speedy, written in C, is executed for each request. This fast frontend then contacts the persistent Perl process, which
is usually already running, to do the work and return the results.
Each perl script runs in its own Unix process, so one perl script can't interfere with another. Command line options can also be used to
deal with programs that have memory leaks or other problems that might keep them from otherwise running persistently.
The speedy front end connects to the back end process, speedy_backend, via a Unix socket in /tmp. A queue is kept in a shared file in /tmp
that holds an entry for each process. In that queue are the pids of the perl processes waiting for connections. The frontend pulls a
process out of this queue, connects to its socket, sends over the environment and argv, and then uses this socket for stdin/stdout to the
perl process.
FILES
/tmp/speedy* A unix socket used to connect to the frontend process.
AUTHOR
Sam Horrocks
http://daemoninc.com
sam@daemoninc.com
NOTES
This manual page was created by Niko Tyni <ntyni@iki.fi> for Debian GNU/Linux, because the original program does not have one. It is based
on the original and more complete CGI::SpeedyCGI(3pm) manual page.
SEE ALSO perl(1), CGI::SpeedyCGI(3pm), speedy(1)SPEEDY_BACKEND(1p)