Sponsored Content
Top Forums UNIX for Advanced & Expert Users Kill phantom jobs automatically? Post 7112 by LivinFree on Wednesday 19th of September 2001 01:27:43 AM
Old 09-19-2001
I read an interesting description of zombies a while back. I can't remember where - maybe someone else can.
Have you ever seen the movie "The Princess Bride"? If so, you remeber that the main character was not dead - just "mostly dead". Zombie processes are like that. You can't kill them because they don't really exist as processes.

The only way of killing them, that I'm aware is to kill the parent - and not with -9 either. Usually, when a server shuts down ( from what I've seen ), it send various kill signals to all processes, starting with a simple request, then progressing (eventually) to SIGKILL (-9), to make sure all proceeses stop. Is this happening on your server? Also, are you running any sort of large database, or disk intensive programs? If so, it can take over 20-30 minutes to write all data to disk before it will exit. Make sure you are not kill -9'ing those!
 

8 More Discussions You Might Find Interesting

1. AIX

I created a phantom file now i can't delete it?

Ok, somehow i've managed to create two .ksh files with the same name. Impossible i know but somehow i did it by mistake... I was actually copying a file and renaming it as something else but as i was typing the copy name i hit the delete key by mistake and got the ^? characters in the file name... (9 Replies)
Discussion started by: Jazmania
9 Replies

2. AIX

w shows phantom user

Hi When i use "w" command, It list some users with "-" command. That means these use already logout,but still in the system somewhere, no process but list under the "w" and "who" command.How can i get rid of these user. Can anybody help me out.thanks a lot xiko (2 Replies)
Discussion started by: xiko
2 Replies

3. Shell Programming and Scripting

background jobs exit status and limit the number of jobs to run

i need to execute 5 jobs at a time in background and need to get the exit status of all the jobs i wrote small script below , i'm not sure this is right way to do it.any ideas please help. $cat run_job.ksh #!/usr/bin/ksh #################################### typeset -u SCHEMA_NAME=$1 ... (1 Reply)
Discussion started by: GrepMe
1 Replies

4. Shell Programming and Scripting

waiting on jobs in bash, allowing limited parallel jobs at one time, and then for all to finish

Hello, I am running GNU bash, version 3.2.39(1)-release (x86_64-pc-linux-gnu). I have a specific question pertaining to waiting on jobs run in sub-shells, based on the max number of parallel processes I want to allow, and then wait... (1 Reply)
Discussion started by: srao
1 Replies

5. Red Hat

Phantom space usage in /

Hi everyone, Got an interesting one (well, interesting to me) I have a box with a 5Gb / mount point. Checking for large files I found nothing and in fact when I did a full du I found that there was only 1.6Gb in use! And yet 100% used in / So there's an unaccounted 3.4Gb somewhere! The... (3 Replies)
Discussion started by: keefbaker
3 Replies

6. UNIX for Dummies Questions & Answers

Phantom Protocol Configuration File

I've been trying to set up the phantom protocol just to try it out. I compiled it fine, but when I ran it I got an error that the configuration file wouldn't load. I found that file didn't exist, so I created it with a blank file, but got this: ./phantom Loading configuration file... (4 Replies)
Discussion started by: Azrael
4 Replies

7. Programming

Phantom Arrays in PL/SQL

Having issue with an oracle stored procedure that fetches 5k array size to an down stream application using oracle client interface. It is creating phantom arrays and keeps sending arrays that do not exist to begin with and congesting the connections. This happened when we upgraded from oracle... (1 Reply)
Discussion started by: mrn6430
1 Replies

8. Shell Programming and Scripting

Shell script to run multiple jobs and it's dependent jobs

I have multiple jobs and each job dependent on other job. Each Job generates a log and If job completed successfully log file end's with JOB ENDED SUCCESSFULLY message and if it failed then it will end with JOB ENDED with FAILURE. I need an help how to start. Attaching the JOB dependency... (3 Replies)
Discussion started by: santoshkumarkal
3 Replies
queuedefs(4)                                                       File Formats                                                       queuedefs(4)

NAME
queuedefs - queue description file for at, batch, and cron SYNOPSIS
/etc/cron.d/queuedefs DESCRIPTION
The queuedefs file describes the characteristics of the queues managed by cron(1M). Each non-comment line in this file describes one queue. The format of the lines are as follows: q.[njobj][nicen][nwaitw] The fields in this line are: q The name of the queue. a is the default queue for jobs started by at(1); b is the default queue for jobs started by batch (see at(1)); c is the default queue for jobs run from a crontab(1) file. njob The maximum number of jobs that can be run simultaneously in that queue; if more than njob jobs are ready to run, only the first njob jobs will be run, and the others will be run as jobs that are currently running terminate. The default value is 100. nice The nice(1) value to give to all jobs in that queue that are not run with a user ID of super-user. The default value is 2. nwait The number of seconds to wait before rescheduling a job that was deferred because more than njob jobs were running in that job's queue, or because the system-wide limit of jobs executing has been reached. The default value is 60. Lines beginning with # are comments, and are ignored. EXAMPLES
Example 1: A sample file. # # a.4j1n b.2j2n90w This file specifies that the a queue, for at jobs, can have up to 4 jobs running simultaneously; those jobs will be run with a nice value of 1. As no nwait value was given, if a job cannot be run because too many other jobs are running cron will wait 60 seconds before trying again to run it. The b queue, for batch(1) jobs, can have up to 2 jobs running simultaneously; those jobs will be run with a nice(1) value of 2. If a job cannot be run because too many other jobs are running, cron(1M) will wait 90 seconds before trying again to run it. All other queues can have up to 100 jobs running simultaneously; they will be run with a nice value of 2, and if a job cannot be run because too many other jobs are running cron will wait 60 seconds before trying again to run it. FILES
/etc/cron.d/queuedefs queue description file for at, batch, and cron. SEE ALSO
at(1), crontab(1), nice(1), cron(1M) SunOS 5.10 1 Mar 1994 queuedefs(4)
All times are GMT -4. The time now is 11:23 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy