Here is a file I want to process
For each instance, there are 12 parameters.
Now my script:
And its execution:
My example only uses two of the variables I am reading in, but you should get the idea. I have a text file with paramaters, and then read through the file executing commands on my read variables.
I think I see what you're saying here. But i didn't clarify that I won't be using unix to read the file. All I'm doing is using shell script to see if the file exists on a directory. If it does, the I'll call an external program to execute (extract and load data from) the file.
As such, I'm looking to see how I should handle multiple (source) files in the directory, all called by the same program (different call command parameters per different target structures). Do I make sense?
Hi,
i have a script 'a.sh' that should be called only by certain scripts like b.sh, c.sh
Inside a.sh, how can i determine
1) if this script was run directly from command prompt (or scheduler)
2) if called via other scripts?
Is there an easy way to get parent process name (not just pid),... (2 Replies)
I wrote a .sh script to run 5 php scripts. The problem is that it's running 1 then 2 then 3 in that order .... I want it to execute all 5 at ONCE....
nohup php /home/script1/script1.php && nohup php /home/script2/script2.php && nohup php /home/script3/script3.php && nohup php... (1 Reply)
I'm pretty new to scripting in Korn shell so please forgive me...
What I'm trying to do is to create a script that calls multiple other ksh scripts and defines variables for text files.
I need it to define my user defined variables (file paths, date & time stamps, etc that are currently in... (1 Reply)
How can I run the following command on multiple files and print out the corresponding multiple files.
perl script.pl genome.gff 1.txt > 1.gff
However, there are multiples files of 1.txt, from 1----100.txt
Thank you so much.
No duplicate posting! Continue here. (0 Replies)
How can I Run one script on multiple files and print out multiple files.
FOR EXAMPLE
i want to run script.pl on 100 files named 1.txt ....100.txt under same directory and print out corresponding file 1.gff ....100.gff.THANKS (4 Replies)
Hi Folks,
Please help me ,I need a unix shell script to check for multiple jobs running. if there are multiple backup Jobs running then it should be trigger an email .
Thanks,
Anand T (1 Reply)
I have multiple jobs and each job dependent on other job.
Each Job generates a log and If job completed successfully log file end's with JOB ENDED SUCCESSFULLY message and if it failed then it will end with JOB ENDED with FAILURE.
I need an help how to start.
Attaching the JOB dependency... (3 Replies)
Hi ,
Can someone help!
I need a shell script to run multiple scripts by using single shell script,
incase any one of the scripts fails, it should get exit and after trouble shooting if we re-execute it, it should start from the failed script.
I have a written a scripting till the... (1 Reply)
I have two scripts which I'm tying to run one after the other- this is what I've tried:
00 14 * * * /path/one.sh && /path/two.sh
I've also tried putting each script on a different line:
00 14 * * * /path/one.sh
00 14 * * * /path/two.sh
Can this be done? (1 Reply)
I have a txt file
line1
line2
line3
$!/bin/sh
cat /tmp/lus.txt | while read line
do
esxcli storage vmfs unmap -u $lin -n 4000
done
this works but does in one line at a time.
how do I do all lines at once simutaeously?
Please use CODE tags as required by forum rules! (4 Replies)
Discussion started by: tdubb123
4 Replies
LEARN ABOUT DEBIAN
cd-hit-para
CD-HIT-PARA.PL(1) User Commands CD-HIT-PARA.PL(1)NAME
cd-hit-para.pl - divide a big clustering job into pieces to run cd-hit or cd-hit-est jobs
SYNOPSIS
cd-hit-para.pl options
DESCRIPTION
This script divide a big clustering job into pieces and submit jobs to remote computers over a network to make it parallel. After
all the jobs finished, the script merge the clustering results as if you just run a single cd-hit or cd-hit-est.
You can also use it to divide big jobs on a single computer if your computer does not have enough RAM (with -L option).
Requirements:
1 When run this script over a network, the directory where you
run the scripts and the input files must be available on all the remote hosts with identical path.
2 If you choose "ssh" to submit jobs, you have to have
passwordless ssh to any remote host, see ssh manual to know how to set up passwordless ssh.
3 I suggest to use queuing system instead of ssh,
I currently support PBS and SGE
4 cd-hit cd-hit-2d cd-hit-est cd-hit-est-2d
cd-hit-div cd-hit-div.pl must be in same directory where this script is in.
Options
-i input filename in fasta format, required
-o output filename, required
--P program, "cd-hit" or "cd-hit-est", default "cd-hit"
--B filename of list of hosts,
requred unless -Q or -L option is supplied
--L number of cpus on local computer, default 0
when you are not running it over a cluster, you can use this option to divide a big clustering jobs into small pieces, I suggest you
just use "--L 1" unless you have enough RAM for each cpu
--S Number of segments to split input DB into, default 64
--Q number of jobs to submit to queue queuing system, default 0
by default, the program use ssh mode to submit remote jobs
--T type of queuing system, "PBS", "SGE" are supported, default PBS
--R restart file, used after a crash of run
-h print this help
More cd-hit/cd-hit-est options can be speicified in command line
Questions, bugs, contact Weizhong Li at liwz@sdsc.edu
cd-hit-para.pl 4.6-2012-04-25 April 2012 CD-HIT-PARA.PL(1)