Linux and UNIX Man Pages

Linux & Unix Commands - Search Man Pages

condor_transfer_data(1) [debian man page]

condor_transfer_data(1) 				      General Commands Manual					   condor_transfer_data(1)

Name
       condor_transfer_data transfer - spooled data

Synopsis
       condor_transfer_data [-help -version]

       condor_transfer_data[-pool   centralmanagerhostname[:portnumber]-name  scheddname][-addr  "<a.b.c.d:port>"]  cluster...	cluster.process...
       user... -constraint expression...

       condor_transfer_data[-pool centralmanagerhostname[:portnumber]-name scheddname][-addr "<a.b.c.d:port>"] -all

Description
       condor_transfer_data causes Condor to transfer spooled data. It is meant to be used in conjunction with the -spooloption of  condor_submit,
       as in

       condor_submit -spool mysubmitfile

       Submission  of  a  job with the -spooloption causes Condor to spool all input files, the user log, and any proxy across a connection to the
       machine where the condor_schedddaemon is running. After spooling these files, the machine from which the job is	submitted  may	disconnect
       from the network or modify its local copies of the spooled files.

       When  the job finishes, the job has  JobStatus = 4, meaning that the job has completed. The output of the job is spooled, and condor_trans-
       fer_dataretrieves the output of the completed job.

Options
       -help

	  Display usage information

       -version

	  Display version information

       -pool centralmanagerhostname[:portnumber]

	  Specify a pool by giving the central manager's host name and an optional port number

       -name scheddname

	  Send the command to a machine identified by scheddname

       -addr <a.b.c.d:port>

	  Send the command to a machine located at "<a.b.c.d:port>"

       cluster

	  Transfer spooled data belonging to the specified cluster

       cluster.process

	  Transfer spooled data belonging to a specific job in the cluster

       user

	  Transfer spooled data belonging to the specified user

       -constraint expression

	  Transfer spooled data for jobs which match the job ClassAd expression constraint

       -all

	  Transfer all spooled data

Exit Status
       condor_transfer_datawill exit with a status value of 0 (zero) upon success, and it will exit with the value 1 (one) upon failure.

Author
       Condor Team, University of Wisconsin-Madison

Copyright
       Copyright (C) 1990-2012 Condor Team, Computer Sciences Department, University of  Wisconsin-Madison,  Madison,  WI.  All  Rights  Reserved.
       Licensed under the Apache License, Version 2.0.

       See the Condor Version 7.8.2 Manualor http://www.condorproject.org/licensefor additional notices. condor-admin@cs.wisc.edu

								  September 2012					   condor_transfer_data(1)

Check Out this Related Man Page

condor_rm(1)						      General Commands Manual						      condor_rm(1)

Name
       condor_rm remove - jobs from the Condor queue

Synopsis
       condor_rm [-help -version]

       condor_rm[-debug]   [-forcex]   [-pool	centralmanagerhostname[:portnumber]-name   scheddname][-addr  "<a.b.c.d:port>"]  cluster...  clus-
       ter.process... user... -constraint expression...

       condor_rm[-debug] [-pool centralmanagerhostname[:portnumber]-name scheddname][-addr "<a.b.c.d:port>"] -all

Description
       condor_rmremoves one or more jobs from the Condor job queue. If the -nameoption is specified, the named condor_scheddis targeted  for  pro-
       cessing.  Otherwise, the local condor_scheddis targeted. The jobs to be removed are identified by one or more job identifiers, as described
       below. For any given job, only the owner of the job or one of the queue super users (defined by the  QUEUE_SUPER_USERS  macro)  can  remove
       the job.

       When  removing  a grid job, the job may remain in the ``X'' state for a very long time. This is normal, as Condor is attempting to communi-
       cate with the remote scheduling system, ensuring that the job has been properly cleaned up. If it takes too long, or in rare  circumstances
       is never removed, the job may be forced to leave the job queue by using the -forcexoption. This forcibly removes jobs that are in the ``X''
       state without attempting to finish any clean up at the remote scheduler.

Options
       -help

	  Display usage information

       -version

	  Display version information

       -pool centralmanagerhostname[:portnumber]

	  Specify a pool by giving the central manager's host name and an optional port number

       -name scheddname

	  Send the command to a machine identified by scheddname

       -addr <a.b.c.d:port>

	  Send the command to a machine located at "<a.b.c.d:port>"

       -debug

	  Causes debugging information to be sent to  stderr , based on the value of the configuration variable  TOOL_DEBUG

       -forcex

	  Force the immediate local removal of jobs in the 'X' state (only affects jobs already being removed)

       cluster

	  Remove all jobs in the specified cluster

       cluster.process

	  Remove the specific job in the cluster

       user

	  Remove jobs belonging to specified user

       -constraint expression

	  Remove all jobs which match the job ClassAd expression constraint

       -all

	  Remove all the jobs in the queue

General Remarks
       Use the -forcexargument with caution, as it will remove jobs from the local queue immediately, but can orphan parts of  the  job  that  are
       running remotely and have not yet been stopped or removed.

Examples
       For a user to remove all their jobs that are not currently running:

       % condor_rm  -constraint 'JobStatus =!= 2'

Exit Status
       condor_rmwill exit with a status value of 0 (zero) upon success, and it will exit with the value 1 (one) upon failure.

Author
       Condor Team, University of Wisconsin-Madison

Copyright
       Copyright  (C)  1990-2012  Condor  Team,  Computer  Sciences Department, University of Wisconsin-Madison, Madison, WI. All Rights Reserved.
       Licensed under the Apache License, Version 2.0.

       See the Condor Version 7.8.2 Manualor http://www.condorproject.org/licensefor additional notices. condor-admin@cs.wisc.edu

								  September 2012						      condor_rm(1)
Man Page