09-18-2001
Zombies Processes
Last week end I tried to kill some Zombie processes that dont let me shutted down the server as a matter of fact I tell ya, from my point of view there is no solution to kill a zombie process but a reboot. I know this is not a polite way to handle a server but even kill -9 or fuser -ck wasnt enough. I really wonder some one who know how to eliminate such processes.
8 More Discussions You Might Find Interesting
1. AIX
Ok, somehow i've managed to create two .ksh files with the same name. Impossible i know but somehow i did it by mistake...
I was actually copying a file and renaming it as something else but as i was typing the copy name i hit the delete key by mistake and got the ^? characters in the file name... (9 Replies)
Discussion started by: Jazmania
9 Replies
2. AIX
Hi
When i use "w" command, It list some users with "-" command. That means these use already logout,but still in the system somewhere, no process but list under the "w" and "who" command.How can i get rid of these user. Can anybody help me out.thanks a lot
xiko (2 Replies)
Discussion started by: xiko
2 Replies
3. Shell Programming and Scripting
i need to execute 5 jobs at a time in background and need to get the exit status of all the jobs i wrote small script below , i'm not sure this is right way to do it.any ideas please help.
$cat run_job.ksh
#!/usr/bin/ksh
####################################
typeset -u SCHEMA_NAME=$1
... (1 Reply)
Discussion started by: GrepMe
1 Replies
4. Shell Programming and Scripting
Hello,
I am running GNU bash, version 3.2.39(1)-release (x86_64-pc-linux-gnu). I have a specific question pertaining to waiting on jobs run in sub-shells, based on the max number of parallel processes I want to allow, and then wait... (1 Reply)
Discussion started by: srao
1 Replies
5. Red Hat
Hi everyone,
Got an interesting one (well, interesting to me)
I have a box with a 5Gb / mount point. Checking for large files I found nothing and in fact when I did a full du I found that there was only 1.6Gb in use! And yet 100% used in /
So there's an unaccounted 3.4Gb somewhere!
The... (3 Replies)
Discussion started by: keefbaker
3 Replies
6. UNIX for Dummies Questions & Answers
I've been trying to set up the phantom protocol just to try it out. I compiled it fine, but when I ran it I got an error that the configuration file wouldn't load. I found that file didn't exist, so I created it with a blank file, but got this:
./phantom
Loading configuration file... (4 Replies)
Discussion started by: Azrael
4 Replies
7. Programming
Having issue with an oracle stored procedure that fetches 5k array size to an down stream application using oracle client interface. It is creating phantom arrays and keeps sending arrays that do not exist to begin with and congesting the connections. This happened when we upgraded from oracle... (1 Reply)
Discussion started by: mrn6430
1 Replies
8. Shell Programming and Scripting
I have multiple jobs and each job dependent on other job.
Each Job generates a log and If job completed successfully log file end's with JOB ENDED SUCCESSFULLY message and if it failed then it will end with JOB ENDED with FAILURE.
I need an help how to start.
Attaching the JOB dependency... (3 Replies)
Discussion started by: santoshkumarkal
3 Replies
LEARN ABOUT HPUX
queuedefs
queuedefs(4) Kernel Interfaces Manual queuedefs(4)
NAME
queuedefs - queue description file for at, batch, and crontab
SYNOPSIS
DESCRIPTION
The file describes the characteristics of the queues managed by (see cron(1M)). Each non-comment line in this file describes one queue.
The format of the lines are as follows:
[njob[nice[nwait
The fields in this line are:
q The name of the queue, such that is the default queue for jobs started by (see at(1)), is the queue for jobs started by
(see at(1)), and is the queue for jobs run from a file (see crontab(1)). Queue names through designate user-defined
queues.
njob The maximum number of jobs that can be run simultaneously in that queue. Although any number can be specified here, (see
cron(1M)) by default limits the number of jobs that can be run on all the queues to 100. This limitation can be removed
by setting the variable to 1 in the file.
nice The value to give to all jobs in that queue that are not run with a user ID of super-user (see nice(1)). The default
value is 2.
nwait The number of seconds to wait before rescheduling a job that was deferred because more than njob jobs were running in that
job's queue, or because more than 100 jobs were running in all the queues (see njob above).
EXAMPLES
Consider the following file:
The file is interpreted as follows:
The queue, for jobs (see at(1)), can have up to 4 jobs running simultaneously, and those jobs will be run with a value of
1.
Since no nwait value is given, if a job cannot be run because too many other jobs are running, will wait 60 seconds
before trying again to run it (see cron(1M)).
The queue, for jobs (see at(1)), can have up to 2 jobs running simultaneously. Those jobs will be run with a value of 2.
If a job cannot be run because too many other jobs are running, will wait 90 seconds before trying again to run it.
All other queues can have up to 100 jobs running simultaneously. They will be run with a value of 2, and if a job cannot be run because
too many other jobs are running, will wait 60 seconds before trying again to run it.
SEE ALSO
at(1), nice(1), crontab(1), cron(1M), proto(4).
STANDARDS CONFORMANCE
queuedefs(4)