Sponsored Content
Top Forums UNIX for Advanced & Expert Users Submitting multiple qsub jobs Post 302959284 by sea on Saturday 31st of October 2015 12:39:58 AM
Old 10-31-2015
You dont need the () there, unless you encapsule ls like $(ls run_*.sh), which would be unnecesary.
Not sure if it would require a leading ./run_*.sh.

hth
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

.netrc multiple ftp jobs to same machine

I have an ftp user, which has been setup to run ftp jobs to a specific machine (different jobs). for the first job i created .netrc in the ftp users home directory and added the appropriate commands machine FTPBOX01 login user1 password xxx macdef init etc etc get file bye I use the... (3 Replies)
Discussion started by: hcclnoodles
3 Replies

2. UNIX for Advanced & Expert Users

Submitting jobs remotely...Experts help reqd

Is there any way I can submit a job to a remote machine and return immediately without withing for the job to finish? What I mean is this...using rsh I can submit a job to a remote machine like this: rsh remotemac1 job.sh But this doesn't return untill the job has finished and as a... (3 Replies)
Discussion started by: arijit
3 Replies

3. Shell Programming and Scripting

monitoring multiple jobs Unix

I have a list of jobs. their expected start time and their expected duration. my needs are monitor multiple jobs (starting at different time and their execution time also differs) Need to mail if any of the job running longer than its expected duration. Whats the efficient way to scripting... (1 Reply)
Discussion started by: vikram3.r
1 Replies

4. Shell Programming and Scripting

Multiple jobs in perl

Hello, I would like to write a perl script which executes several jobs. The key thing is I only want 4 jobs to be executed at one time (that's because my machine as 4 cpu, and I want one job per cpu). Is there any way that I can get perl to co-operate with me in this? Thanks! (1 Reply)
Discussion started by: amcrisan
1 Replies

5. UNIX for Advanced & Expert Users

Problem Running qsub multiple jobs

Hello, I have a perl script that takes in one file of input and outputs to a file. I would like to run this script on several input files, and I would like to run it with qsub, something like this: Input files: FileListDssp.txt.numaa FileListDssp.txt.numab FileListDssp.txt.numac etc.. ... (1 Reply)
Discussion started by: InfoSeeker
1 Replies

6. Shell Programming and Scripting

Multiple jobs reading from same file

I have a sequence of tasks that I routinely run and I'm trying to parallelize certain portions of the sequence. Specifically, there are 3 tasks which all read from the same file, each performing different operations and writing to their own seperate file. I was wondering if I could execute these... (3 Replies)
Discussion started by: erichpowell
3 Replies

7. Shell Programming and Scripting

How to View multiple Cron jobs

Hi, I ran two crontab commands using: crontab program1 crontab program2 However when I type crontab -l only the second cron job shows up, how do I see all cron jobs running and how do I edit all at the same time Thanks in Advance S:D (10 Replies)
Discussion started by: walforum
10 Replies

8. Shell Programming and Scripting

Help submitting jobs to cluster

Hi I am new to submitting jobs. I am trying to submit my perl file to the cluster. This is what my shell file looks like (shell1.sh): #!/bin/sh #$ -S /bin/sh cd data/projects/mydir/abbc perl autocorro.pl followed by qsub shell1.sh It takes the qsub, but does nothing. I check... (1 Reply)
Discussion started by: theawknewbie
1 Replies

9. Shell Programming and Scripting

Shell script to run multiple jobs and it's dependent jobs

I have multiple jobs and each job dependent on other job. Each Job generates a log and If job completed successfully log file end's with JOB ENDED SUCCESSFULLY message and if it failed then it will end with JOB ENDED with FAILURE. I need an help how to start. Attaching the JOB dependency... (3 Replies)
Discussion started by: santoshkumarkal
3 Replies

10. UNIX for Beginners Questions & Answers

How to run multiple cron jobs?

I have two scripts which I'm tying to run one after the other- this is what I've tried: 00 14 * * * /path/one.sh && /path/two.sh I've also tried putting each script on a different line: 00 14 * * * /path/one.sh 00 14 * * * /path/two.sh Can this be done? (1 Reply)
Discussion started by: $shell_Learner
1 Replies
KEEPALIVED(8)						      System Manager's Manual						     KEEPALIVED(8)

NAME
keepalived - load-balancing and high-availability service SYNOPSIS
keepalived [-f|--use-file=FILE] [-P|--vrrp] [-C|--check] [-l|--log-console] [-D|--log-detail] [-S|--log-facility={0-7}] [-V|--dont-release-vrrp] [-I|--dont-release-ipvs] [-R|--dont-respawn] [-n|--dont-fork] [-d|--dump-conf] [-p|--pid=FILE] [-r|--vrrp_pid=FILE] [-c|--checkers_pid=FILE] [-v|--version] [-h|--help] DESCRIPTION
Keepalived provides simple and robust facilities for load-balancing and high-availability. The load-balancing framework relies on well-known and widely used Linux Virtual Server (IPVS) kernel module providing Layer4 load-balancing. Keepalived implements a set of checkers to dynamically and adaptively maintain and manage load-balanced server pool according their health. Keepalived also implements the VRRPv2 protocol to achieve high-availability with director failover. OPTIONS
-f, --use-file=FILE Use the specified configuration file. The default configuration file is "/etc/keepalived/keepalived.conf". -P, --vrrp Only run the VRRP subsystem. This is useful for configurations that do not use IPVS load balancer. -C, --check Only run the healthcheck subsystem. This is useful for configurations that use the IPVS load balancer with a single director with no failover. -l, --log-console Log messages to the local console. The default behavior is to log messages to syslog. -D, --log-detail Detailed log messages. -S, --log-facility=[0-7] Set syslog facility to LOG_LOCAL[0-7]. The default syslog facility is LOG_DAEMON. -V, --dont-release-vrrp Don't remove VRRP VIPs and VROUTEs on daemon stop. The default behavior is to remove all VIPs and VROUTEs when keepalived exits -I, --dont-release-ipvs Don't remove IPVS topology on daemon stop. The default behavior it to remove all entries from the IPVS virtual server table on when keepalived exits. -R, --dont-respawn Don't respawn child processes. The default behavior is to restart the VRRP and checker processes if either process exits. -n, --dont-fork Don't fork the daemon process. This option will cause keepalived to run in the foreground. -d, --dump-conf Dump the configuration data. -p, --pid=FILE Use specified pidfile for parent keepalived process. The default pidfile for keepalived is "/var/run/keepalived.pid". -r, --vrrp_pid=FILE Use specified pidfile for VRRP child process. The default pidfile for the VRRP child process is "/var/run/keepalived_vrrp.pid". -c, --checkers_pid=FILE Use specified pidfile for checkers child process. The default pidfile for the checker child process is "/var/run/keepalived_checkers.pid". -v, --version Display the version and exit. -h, --help Display this help message and exit. SEE ALSO
keepalived.conf(5), ipvsadm(8) AUTHOR
This man page was written by Ryan O'Hara <rohara@redhat.com> November 2012 KEEPALIVED(8)
All times are GMT -4. The time now is 08:47 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy