Provided by: condor_23.6.2+dfsg-2build1_amd64 bug

NAME

       condor_submit - HTCondor Manual

       Queue jobs for execution under HTCondor

SYNOPSIS

       condor_submit  [-terse  ]  [-verbose  ]  [-unused  ]  [-file  submit_file]  [-name  schedd_name] [-remote
       schedd_name] [-addr <ip:port>] [-pool pool_name] [-disable ] [-password passphrase]  [-debug  ]  [-append
       command   ...][-batch-name   batch_name]   [-spool   ]  [-dump  filename]  [-interactive  ]  [-factory  ]
       [-allow-crlf-script    ]    [-dry-run    file]    [-maxjobs    number-of-jobs]     [-single-cluster     ]
       [<submit-variable>=<value> ] [submit description file ] [-queue queue_arguments]

DESCRIPTION

       condor_submit is the program for submitting jobs for execution under HTCondor. condor_submit requires one
       or  more  submit description commands to direct the queuing of jobs. These commands may come from a file,
       standard input, the command line, or from some combination of these. One submit description  may  contain
       specifications  for  the  queuing of many HTCondor jobs at once. A single invocation of condor_submit may
       cause one or more clusters. A cluster is a set of jobs specified in the submit description between  queue
       commands  for which the executable is not changed. It is advantageous to submit multiple jobs as a single
       cluster because the schedd uses much less memory to hold the jobs.

       Multiple clusters may be specified within a single submit  description.   Each  cluster  must  specify  a
       single executable.

       The job ClassAd attribute ClusterId identifies a cluster.

       The  submit  description  file argument is the path and file name of the submit description file. If this
       optional argument is the dash character (-), then the commands are taken from standard  input.  If  -  is
       specified  for  the  submit  description  file, -verbose is implied; this can be overridden by specifying
       -terse.

       If no submit description file argument is given, and no -queue argument  is  given,  commands  are  taken
       automatically from standard input.

       Note  that  submission  of  jobs  from a Windows machine requires a stashed password to allow HTCondor to
       impersonate the user submitting the job.  To stash a password, use the condor_store_cred command. See the
       manual page for details.

       For lengthy lines within the submit description file, the backslash (\) is a line continuation character.
       Placing the backslash at the end of a line causes the current line's command to  be  continued  with  the
       next  line  of  the  file. Submit description files may contain comments. A comment is any line beginning
       with a pound character (#).

OPTIONS

          -terse Terse output - display JobId ranges only.

          -verbose
                 Verbose output - display the created job ClassAd

          -unused
                 As a default, causes no warnings to be issued about user-defined macros not being  used  within
                 the  submit  description  file.  The meaning reverses (toggles) when the configuration variable
                 WARN_ON_UNUSED_SUBMIT_FILE_MACROS is set to the  non  default  value  of  False.  Printing  the
                 warnings  can  help  identify spelling errors of submit description file commands. The warnings
                 are sent to stderr.

          -file submit_file
                 Use submit_file as the submit description file. This is equivalent to providing submit_file  as
                 an argument without the preceding -file.

          -name schedd_name
                 Submit  to the specified condor_schedd. Use this option to submit to a condor_schedd other than
                 the default local one.  schedd_name is the value of the Name ClassAd attribute on  the  machine
                 where the condor_schedd daemon runs.

          -remote schedd_name
                 Submit  to  the  specified  condor_schedd,  spooling  all required input files over the network
                 connection. schedd_name is the value of the Name ClassAd attribute on  the  machine  where  the
                 condor_schedd daemon runs. This option is equivalent to using both -name and -spool.

          -addr <ip:port>
                 Submit  to  the  condor_schedd  at  the IP address and port given by the sinful string argument
                 <ip:port>.

          -pool pool_name
                 Look in the specified pool for the condor_schedd to submit to.  This option is used with  -name
                 or -remote.

          -disable
                 Disable  file  permission checks when submitting a job for read permissions on all input files,
                 such as those defined by commands input and transfer_input_files, as well as  write  permission
                 to  output  files,  such  as  a log file defined by log and output files defined with output or
                 transfer_output_files.

          -debug Cause debugging information to be sent to stderr, based  on  the  value  of  the  configuration
                 variable TOOL_DEBUG.

          -append command
                 Augment  the  commands in the submit description file with the given command. This command will
                 be considered to immediately precede the queue command within the submit description file,  and
                 come  after  all  other  previous commands. If the command specifies a queue command, as in the
                 example

                 condor_submit mysubmitfile -append "queue input in A, B, C"

                 then the entire -append command line option and its arguments are converted to

                 condor_submit mysubmitfile -queue input in A, B, C

                 The submit description file is not modified. Multiple  commands  are  specified  by  using  the
                 -append option multiple times. Each new command is given in a separate -append option. Commands
                 with spaces in them will need to be enclosed in double quote marks.

          -batch-name batch_name
                 Set  the  batch  name  for  this  submit. The batch name is displayed by condor_q -batch. It is
                 intended for use by users to give meaningful names to their jobs and to influence how  condor_q
                 groups  jobs  for display. Use of this argument takes precedence over a batch name specified in
                 the submit description file itself.

          -spool Spool all required  input  files,  job  event  log,  and  proxy  over  the  connection  to  the
                 condor_schedd.  After submission, modify local copies of the files without affecting your jobs.
                 Any output files for completed jobs need to be retrieved with condor_transfer_data.

          -dump filename
                 Sends all ClassAds to the specified file, instead of to the condor_schedd.

          -interactive
                 Indicates that the user wants to run an interactive shell on an execute machine  in  the  pool.
                 This  is  equivalent to creating a submit description file of a vanilla universe sleep job, and
                 then running condor_ssh_to_job by hand. Without any additional  arguments,  condor_submit  with
                 the -interactive flag creates a dummy vanilla universe job that sleeps, submits it to the local
                 scheduler, waits for the job to run, and then launches condor_ssh_to_job to run a shell. If the
                 user  would  like  to  run  the  shell  on  a  machine  that  matches a particular requirements
                 expression, the submit description file is specified, and it will contain the expression.  Note
                 that  all  policy  expressions  specified  in  the submit description file are honored, but any
                 executable or universe commands are overwritten to  be  sleep  and  vanilla.  The  job  ClassAd
                 attribute  InteractiveJob  is set to True to identify interactive jobs for condor_startd policy
                 usage.

          -factory
                 Sends all of the jobs as a late materialization job factory.   A  job  factory  consists  of  a
                 single  cluster  classad  and a digest containing the submit commands necessary to describe the
                 differences between jobs.  If the Queue statement has itemdata, then the itemdata will be sent.
                 Using this option is equivalent to using the max_materialize submit command.

          -allow-crlf-script
                 Changes the check for an invalid line ending on the executable script's #! line from  an  ERROR
                 to  a WARNING. The #! line will be ignored by Windows, so it won't matter if it is invalid; but
                 Unix and Linux will not run a script that has a Windows/DOS line ending on the  first  line  of
                 the  script.  So  condor_submit  will  not  allow  such  a  script to be submitted as the job's
                 executable unless this option is supplied.

          -dry-run file
                 Parse the submit description file, sending the resulting job ClassAd to the file given by file,
                 but do not submit the job(s). This  permits  observation  of  the  job  specification,  and  it
                 facilitates debugging the submit description file contents. If file is -, the output is written
                 to stdout.

          -maxjobs number-of-jobs
                 If  the  total number of jobs specified by the submit description file is more than the integer
                 value given by number-of-jobs, then no jobs are submitted for execution and an error message is
                 generated.  A 0 or negative value for the number-of-jobs causes no limit to be imposed.

          -single-cluster
                 If the jobs specified by the submit description file causes more than a single cluster value to
                 be assigned, then no jobs are submitted for execution and an error message is generated.

          <submit-variable>=<value>
                 Defines a submit command or submit variable with a value, and parses it as if it was placed  at
                 the  beginning  of  the submit description file. The submit description file is not changed. To
                 correctly parse the condor_submit command line, this option must  be  specified  without  white
                 space  characters before and after the equals sign (=), or the entire option must be surrounded
                 by double quote marks.

          -queue queue_arguments
                 A command line specification of how many jobs to queue, which is only permitted if  the  submit
                 description  file  does  not  have  a queue command. The queue_arguments are the same as may be
                 within a submit description file. The parsing of the queue_arguments finishes at the end of the
                 line or when a dash character (-) is encountered. Therefore,  its  best  placement  within  the
                 command line will be at the end of the command line.

                 On a Unix command line, the shell expands file globs before parsing occurs.

SUBMIT DESCRIPTION FILE COMMANDS

       Note: more information on submitting HTCondor jobs can be found here: Submitting a Job.

       The  condor_submit  language  supports  multi-line  values  in  commands.  The  syntax is the same as the
       configuration language  (see  more  details  here:  admin-manual/introduction-to-configuration:multi-line
       values).

       Each  submit  description  file  describes  one  or  more  clusters  of jobs to be placed in the HTCondor
       execution pool. All jobs in a cluster must share the same executable, but they may have  different  input
       and  output  files,  and  different  program arguments. The submit description file is generally the last
       command-line argument to condor_submit. If the submit description file argument is omitted, condor_submit
       will read the submit description from standard input.

       The submit description file must contain at least one executable command and at least one queue  command.
       All of the other commands have default actions.

       Note  that  a  submit  file that contains more than one executable command will produce multiple clusters
       when submitted. This is not generally recommended, and is not allowed for submit files that  are  run  as
       DAG node jobs by condor_dagman.

       The  commands  which  can  appear  in  the  submit description file are numerous. They are listed here in
       alphabetical order by category.

       BASIC COMMANDS

          arguments = <argument_list>
                 List of arguments to be supplied to the executable as part of the command line.

                 In the java universe, the first argument must be the name of the class containing main.

                 There are two permissible formats for specifying arguments, identified as the  old  syntax  and
                 the new syntax. The old syntax supports white space characters within arguments only in special
                 circumstances;  when  used,  the  command  line  arguments  are  represented in the job ClassAd
                 attribute Args. The new syntax supports  uniform  quoting  of  white  space  characters  within
                 arguments;  when  used, the command line arguments are represented in the job ClassAd attribute
                 Arguments.

                 Old Syntax

                 In the old syntax, individual  command  line  arguments  are  delimited  (separated)  by  space
                 characters.  To  allow a double quote mark in an argument, it is escaped with a backslash; that
                 is, the two character sequence \" becomes a single double quote mark within an argument.

                 Further interpretation of the argument string differs depending on  the  operating  system.  On
                 Windows,  the  entire  argument string is passed verbatim (other than the backslash in front of
                 double quote marks) to the Windows application. Most Windows  applications  will  allow  spaces
                 within  an  argument  value  by surrounding the argument with double quotes marks. In all other
                 cases, there is no further interpretation of the arguments.

                 Example:

                     arguments = one \"two\" 'three'

                 Produces in Unix vanilla universe:

                     argument 1: one
                     argument 2: "two"
                     argument 3: 'three'

                 New Syntax

                 Here are the rules for using the new syntax:

                 1. The entire string representing the command line arguments  is  surrounded  by  double  quote
                    marks. This permits the white space characters of spaces and tabs to potentially be embedded
                    within a single argument. Putting the double quote mark within the arguments is accomplished
                    by escaping it with another double quote mark.

                 2. The white space characters of spaces or tabs delimit arguments.

                 3. To  embed  white  space  characters of spaces or tabs within a single argument, surround the
                    entire argument with single quote marks.

                 4. To insert a literal single quote mark, escape it within an  argument  already  delimited  by
                    single quote marks by adding another single quote mark.

                 Example:

                     arguments = "3 simple arguments"

                 Produces:

                     argument 1: 3
                     argument 2: simple
                     argument 3: arguments

                 Another example:

                     arguments = "one 'two with spaces' 3"

                 Produces:

                     argument 1: one
                     argument 2: two with spaces
                     argument 3: 3

                 And yet another example:

                     arguments = "one ""two"" 'spacey ''quoted'' argument'"

                 Produces:

                     argument 1: one
                     argument 2: "two"
                     argument 3: spacey 'quoted' argument

                 Notice  that  in  the  new  syntax,  the  backslash  has  no  special meaning.  This is for the
                 convenience of Windows users.

          environment = <parameter_list>
                 List of environment variables.

                 There are two different formats for specifying the environment variables: the  old  format  and
                 the  new  format.  The  old  format  is  retained for backward-compatibility. It suffers from a
                 platform-dependent syntax and  the  inability  to  insert  some  special  characters  into  the
                 environment.

                 The new syntax for specifying environment values:

                 1. Put  double quote marks around the entire argument string. This distinguishes the new syntax
                    from the old. The old syntax does not have double quote marks around it. Any literal  double
                    quote marks within the string must be escaped by repeating the double quote mark.

                 2. Each environment entry has the form

                        <name>=<value>

                 3. Use white space (space or tab characters) to separate environment entries.

                 4. To  put  any  white  space  in  an  environment entry, surround the space and as much of the
                    surrounding entry as desired with single quote marks.

                 5. To insert a literal single quote mark, repeat the single quote mark  anywhere  inside  of  a
                    section surrounded by single quote marks.

                 Example:

                     environment = "one=1 two=""2"" three='spacey ''quoted'' value'"

                 Produces the following environment entries:

                     one=1
                     two="2"
                     three=spacey 'quoted' value

                 Under   the   old  syntax,  there  are  no  double  quote  marks  surrounding  the  environment
                 specification. Each environment entry remains of the form

                     <name>=<value>

                 Under Unix, list multiple environment entries by separating them with a  semicolon  (;).  Under
                 Windows, separate multiple entries with a vertical bar (|). There is no way to insert a literal
                 semicolon  under  Unix  or a literal vertical bar under Windows. Note that spaces are accepted,
                 but rarely desired, characters within parameter names and values, because they are  treated  as
                 literal  characters,  not  separators or ignored white space. Place spaces within the parameter
                 list only if required.

                 A Unix example:

                     environment = one=1;two=2;three="quotes have no 'special' meaning"

                 This produces the following:

                     one=1
                     two=2
                     three="quotes have no 'special' meaning"

                 If the environment is set with the environment command and getenv is also set, values specified
                 with environment override values in the submitter's environment (regardless of the order of the
                 environment and getenv commands).

          error = <pathname>
                 A path and file name used by HTCondor to capture any error messages the program would  normally
                 write  to  the  screen (that is, this file becomes stderr). A path is given with respect to the
                 file system of the machine on which the job is submitted. The file is written (by the  job)  in
                 the  remote scratch directory of the machine where the job is executed. When the job exits, the
                 resulting file is transferred back to the machine where the job was submitted, and the path  is
                 utilized  for  file placement.  If you specify a relative path, the final path will be relative
                 to the job's initial working directory, and HTCondor will create directories  as  necessary  to
                 transfer  the file.  If not specified, the default value of /dev/null is used for submission to
                 a Unix machine. If not specified, error messages  are  ignored  for  submission  to  a  Windows
                 machine. More than one job should not use the same error file, since this will cause one job to
                 overwrite  the errors of another. If HTCondor detects that the error and output files for a job
                 are the same, it will run the job such that the output and error data is merged.

          executable = <pathname>
                 An optional path and a required file name of the executable file for this job cluster. Only one
                 executable command within a submit description file is guaranteed to work properly.  More  than
                 one often works.

                 If  no  path or a relative path is used, then the executable file is presumed to be relative to
                 the current working directory of the user as the condor_submit command is issued.

          batch_name = <batch_name>
                 Set the batch name for this submit. The batch name is  displayed  by  condor_q  -batch.  It  is
                 intended  for use by users to give meaningful names to their jobs and to influence how condor_q
                 groups jobs for display. This value in a submit  file  can  be  overridden  by  specifying  the
                 -batch-name argument on the condor_submit command line.

          getenv = <<matchlist> | True | False>
                 If getenv is set to
                  True,  then  condor_submit  will copy all of the user's current shell environment variables at
                 the time of job submission into the job ClassAd. The job will therefore execute with  the  same
                 set  of environment variables that the user had at submit time. Defaults to False.  A wholesale
                 import of the user's environment is very likely to lead to problems  executing  the  job  on  a
                 remote  machine  unless  there  is a shared file system for users' home directories between the
                 access point and execute machine.  So rather than setting getenv to True, it is much better  to
                 set it to a list of environment variables to import.

                 Matchlist  is a comma, semicolon or space separated list of environment variable names and name
                 patterns that match or reject names.  Matchlist members are matched case-insensitively to  each
                 name  in  the environment and those that match are imported. Matchlist members can contain * as
                 wildcard character which matches anything at that position.  Members can have two *  characters
                 if  one  of them is at the end. Members can be prefixed with !  to force a matching environment
                 variable to not be imported.  The order of members in  the  Matchlist  has  no  effect  on  the
                 result.  getenv = true is equivalent to getenv = *

                 Prior to HTCondor 8.9.7 getenv allows only True or False as values.

                 Examples:

                     # import everything except PATH and INCLUDE (also path, include and other case-variants)
                     getenv = !PATH, !INCLUDE

                     # import everything with CUDA in the name
                     getenv = *cuda*

                     # Import every environment variable that starts with P or Q, except PATH
                     getenv = !path, P*, Q*

                 If the environment is set with the environment command and getenv is also set, values specified
                 with environment override values in the submitter's environment (regardless of the order of the
                 environment and getenv commands).

          input = <pathname>
                 HTCondor  assumes  that  its  jobs  are  long-running,  and  that the user will not wait at the
                 terminal for their completion. Because of this, the standard files which  normally  access  the
                 terminal,  (stdin, stdout, and stderr), must refer to files. Thus, the file name specified with
                 input should contain any keyboard input the  program  requires  (that  is,  this  file  becomes
                 stdin).  A  path  is  given  with respect to the file system of the machine on which the job is
                 submitted. The file is transferred before execution to the  remote  scratch  directory  of  the
                 machine where the job is executed. If not specified, the default value of /dev/null is used for
                 submission  to  a  Unix machine. If not specified, input is ignored for submission to a Windows
                 machine.

                 Note that this command does not refer  to  the  command-line  arguments  of  the  program.  The
                 command-line arguments are specified by the arguments command.

          log = <pathname>
                 Use  log  to specify a file name where HTCondor will write a log file of what is happening with
                 this job cluster, called a job event log. For example, HTCondor will place  a  log  entry  into
                 this  file  when  and  where  the  job  begins  running, when it transfers files, if the job is
                 evicted, and when the job completes. Most users find specifying a log file to be handy; its use
                 is recommended. If no log entry is specified, HTCondor does not create a log for this  cluster.
                 If  a relative path is specified, it is relative to the current working directory as the job is
                 submitted or the directory specified by submit command initialdir on the access point.

          notification = <Always | Complete | Error | Never>
                 Owners of HTCondor jobs are notified by e-mail when certain events occur. If defined by  Always
                 or Complete, the owner will be notified when the job terminates. If defined by Error, the owner
                 will  only  be notified if the job terminates abnormally, (as defined by JobSuccessExitCode, if
                 defined) or if the job is placed on hold because of a failure, and  not  by  user  request.  If
                 defined  by  Never (the default), the owner will not receive e-mail, regardless to what happens
                 to the job. The HTCondor User's manual documents statistics included in the e-mail.

          notify_user = <email-address>
                 Used to specify the e-mail address to use when HTCondor  sends  e-mail  about  a  job.  If  not
                 specified, HTCondor defaults to using the e-mail address defined by

                     job-owner@UID_DOMAIN

                 where  the  configuration variable  UID_DOMAIN is specified by the HTCondor site administrator.
                 If UID_DOMAIN has not been specified, HTCondor sends the e-mail to:

                     job-owner@submit-machine-name

          output = <pathname>
                 The output file captures any information the program would ordinarily write to the screen (that
                 is, this file becomes stdout). A path is given with respect to the file system of  the  machine
                 on which the job is submitted. The file is written (by the job) in the remote scratch directory
                 of the machine where the job is executed. When the job exits, the resulting file is transferred
                 back  to  the machine where the job was submitted, and the path is utilized for file placement.
                 If you specify a relative path, the final path will be relative to the  job's  initial  working
                 directory,  and  HTCondor  will  create  directories as necessary to transfer the file.  If not
                 specified, the default value of /dev/null is used for submission to  a  Unix  machine.  If  not
                 specified,  output is ignored for submission to a Windows machine. Multiple jobs should not use
                 the same output file, since this will cause one job to overwrite  the  output  of  another.  If
                 HTCondor  detects  that  the error and output files for a job are the same, it will run the job
                 such that the output and error data is merged.

                 Note that if a program explicitly opens and writes to a file, that file should not be specified
                 as the output file.

          priority = <integer>
                 An HTCondor job priority can be any integer,  with  0  being  the  default.  Jobs  with  higher
                 numerical  priority will run before jobs with lower numerical priority. Note that this priority
                 is on a per user basis. One user with many jobs may use this command to order his/her own jobs,
                 and this will have no effect on whether or not these jobs will  run  ahead  of  another  user's
                 jobs.

                 Note  that  the priority setting in an HTCondor submit file will be overridden by condor_dagman
                 if the submit file is used for a node in a DAG, and the priority of the node within the DAG  is
                 non-zero (see  Node Priorities for more details).

          queue [<int expr> ]
                 Places zero or more copies of the job into the HTCondor queue.

          queue  [<int expr> ] [<varname> ] in [slice ] <list of items> Places zero or more copies of the job in
                 the queue based on items in a <list of items>

          queue  [<int  expr>  ]  [<varname>  ]  matching  [files  |  dirs  ]  [slice ] <list of items with file
                 globbing>] Places zero or more copies of the job in the queue based on files that match a <list
                 of items with file globbing>

          queue  [<int expr> ] [<list of varnames> ] from [slice ] <file name> | <list of items>] Places zero or
                 more copies of the job in the queue based on lines from the submit file or from <file name>

                 The optional argument <int expr> specifies how many times to repeat the job  submission  for  a
                 given set of arguments. It may be an integer or an expression that evaluates to an integer, and
                 it  defaults to 1. All but the first form of this command are various ways of specifying a list
                 of items. When these forms are used <int expr> jobs will be queued for each item in  the  list.
                 The in, matching and from keyword indicates how the list will be specified.

                 • in  The  list  of  items  is an explicit comma and/or space separated <list of items>. If the
                   <list of items> begins with an open paren, and the close paren is not on the same line as the
                   open, then the list continues until a line that begins with a close paren is  read  from  the
                   submit file.

                 • matching  Each  item  in  the  <list of items with file globbing> will be matched against the
                   names of files and directories relative to the current directory, the set of  matching  names
                   is the resulting list of items.

                   • files Only filenames will matched.

                   • dirs Only directory names will be matched.

                 • from  <file name> | <list of items> Each line from <file name> or <list of items> is a single
                   item, this allows for multiple variables to be set for each item. Lines from <file  name>  or
                   <list  of  items>  will be split on comma and/or space until there are values for each of the
                   variables specified in <list of varnames>. The last variable will contain  the  remainder  of
                   the line. When the <list of items> form is used, the list continues until the first line that
                   begins  with  a close paren, and lines beginning with pound sign ('#') will be skipped.  When
                   using the <file name> form, if the <file name> ends with |, then it will  be  executed  as  a
                   script whatever the script writes to stdout will be the list of items.

                 The optional argument <varname> or <list of varnames> is the name or names of of variables that
                 will be set to the value of the current item when queuing the job. If no <varname> is specified
                 the  variable  ITEM  will  be  used.  Leading  and trailing whitespace be trimmed. The optional
                 argument <slice> is a python style slice selecting only some of the items in the list of items.
                 Negative step values are not supported.

                 A submit file may contain more than one queue statement, and if desired, any  commands  may  be
                 placed  between  subsequent  queue  commands,  such as new input, output, error, initialdir, or
                 arguments commands.  This is handy when submitting multiple runs  into  one  cluster  with  one
                 submit description file.

          universe = <vanilla | scheduler | local | grid | java | vm | parallel | docker | container>
                 Specifies which HTCondor universe to use when running this job. The HTCondor universe specifies
                 an HTCondor execution environment.

                 The  vanilla  universe is the default (except where the configuration variable DEFAULT_UNIVERSE
                 defines it otherwise).

                 The scheduler universe is for a job that is to run on the machine where the job  is  submitted.
                 This universe is intended for a job that acts as a metascheduler and will not be preempted.

                 The  local universe is for a job that is to run on the machine where the job is submitted. This
                 universe runs the job immediately and will not preempt the job.

                 The grid universe forwards the job to an external job management system. Further  specification
                 of the grid universe is done with the grid_resource command.

                 The java universe is for programs written to the Java Virtual Machine.

                 The vm universe facilitates the execution of a virtual machine.

                 The  parallel  universe is for parallel jobs (e.g. MPI) that require multiple machines in order
                 to run.

                 The docker universe runs a docker container as an HTCondor job.

                 The container universe runs a container as an HTCondor job using a supported container  runtime
                 system on the Execution Point.

          max_materialize = <limit>
                 Submit  jobs as a late materialization factory and instruct the condor_schedd to keep the given
                 number of jobs materialized.  Use this option to reduce the  load  on  the  condor_schedd  when
                 submitting  a  large  number of jobs.  The limit can be an expression but it must evaluate to a
                 constant at submit time.  A limit less than 1 will be treated as unlimited.  The  condor_schedd
                 can  be configured to have a materialization limit as well, the lower of the two limits will be
                 used.  (see  users-manual/submitting-a-job:submitting lots of jobs for more details).

          max_idle = <limit>
                 Submit jobs as a late materialization factory and instruct the condor_schedd to keep the  given
                 number  of  non-running  jobs  materialized.   Use  this  option  to  reduce  the  load  on the
                 condor_schedd when submitting a large number of jobs.  The limit may be an  expression  but  it
                 must  evaluate  to a constant at submit time.  Jobs in the Held state are considered to be Idle
                 for this limit.  A limit of less than 1 will prevent jobs from being materialized although  the
                 factory      will      still      be      submitted     to     the     condor_schedd.      (see
                 users-manual/submitting-a-job:submitting lots of jobs for more details).

       COMMANDS FOR MATCHMAKING

          rank = <ClassAd Float Expression>
                 A ClassAd Floating-Point expression that states how to rank machines which have already met the
                 requirements expression. Essentially, rank expresses preference. A higher numeric value  equals
                 better rank. HTCondor will give the job the machine with the highest rank.  For example,

                     request_memory = max({60, Target.TotalSlotMemory})
                     rank = Memory

                 asks  HTCondor to find all available machines with more than 60 megabytes of memory and give to
                 the job the machine with the most  amount  of  memory.  The  HTCondor  User's  Manual  contains
                 complete  information  on  the  syntax and available attributes that can be used in the ClassAd
                 expression.

          request_cpus = <num-cpus>
                 A requested number of CPUs (cores). If not specified,  the  number  requested  will  be  1.  If
                 specified, the expression

                     && (RequestCpus <= Target.Cpus)

                 is appended to the requirements expression for the job.

                 For  pools that enable dynamic condor_startd provisioning, specifies the minimum number of CPUs
                 requested for this job, resulting in a dynamic slot being created with this many cores.

          request_disk = <quantity>
                 The requested amount of disk space in KiB requested for this job. If not specified, it will  be
                 set to the job ClassAd attribute DiskUsage. The expression

                     && (RequestDisk <= Target.Disk)

                 is appended to the requirements expression for the job.

                 For  pools  that enable dynamic condor_startd provisioning, a dynamic slot will be created with
                 at least this much disk space.

                 Characters may be appended to a numerical value to indicate units.  K or KB indicates KiB,  210
                 numbers  of  bytes.  M  or  MB  indicates MiB, 220 numbers of bytes. G or GB indicates GiB, 230
                 numbers of bytes. T or TB indicates TiB, 240 numbers of bytes.

          request_gpus = <num-gpus>
                 A requested number of GPUs. If not specified, no GPUs will be requested.  If specified  one  of
                 the expressions below

                     && (TARGET.GPUs >= RequestGPUs)
                        or
                     && (countMatches(MY.RequireGPUs,TARGET.AvailableGPUs) >= RequestGPUs)

                 is appended to the requirements expression for the job. The first expression above is used when
                 there  is  no  constraint  on  the  GPU  properties  specified  in the submit file.  The second
                 expression is  used  when  require_gpus  or  one  of  the  GPU  property  constraints  such  as
                 gpus_minimum_memory is used.

                 For  pools  that  enable dynamic condor_startd provisioning, request_gpus specifies the minimum
                 number of GPUs requested for this job, resulting in a dynamic slot being created with this many
                 GPUs with the required properties.

          require_gpus = <constraint-expression>
                 A constraint on the properties of GPUs when used with  a  non-zero  request_gpus  value.   This
                 expression  will  be  combined  with  constraints  generated  by  the  use  of  one  or more of
                 gpus_minimum_capability, gpus_minimum_memory, gpus_minimum_runtime, or gpus_maximum_capability.
                 If not specified, no  constraint  on  GPUs  will  be  added  to  the  job.   If  specified  and
                 request_gpus is non-zero, the expression

                     && (countMatches(MY.RequireGPUs, TARGET.AvailableGPUs) >= RequestGPUs)

                 is appended to the requirements expression for the job.  This expression cannot be evaluated by
                 HTCondor  prior  to  version  9.8.0.  A  warning  to  this  will  effect  will  be printed when
                 condor_submit detects this condition.

                 For pools that enable dynamic condor_startd provisioning and are at least  version  9.8.0,  the
                 constraint  will  be  tested  against the properties of AvailableGPUs and only those that match
                 will be assigned to the dynamic slot.

          request_memory = <quantity>
                 The required amount of memory in MiB that this job needs to avoid excessive  swapping.  If  not
                 specified and the submit command vm_memory is specified, then the value specified for vm_memory
                 defines request_memory , If neither request_memory nor vm_memory is specified, the value is set
                 by  the  configuration variable JOB_DEFAULT_REQUESTMEMORY The actual amount of memory used by a
                 job is represented by the job ClassAd attribute MemoryUsage.

                 For pools that enable dynamic condor_startd provisioning, a dynamic slot will be  created  with
                 at least this much RAM.

                 The expression

                     && (RequestMemory <= Target.Memory)

                 is appended to the requirements expression for the job.

                 Characters  may be appended to a numerical value to indicate units.  K or KB indicates KiB, 210
                 numbers of bytes. M or MB indicates MiB, 220 numbers of bytes.  G  or  GB  indicates  GiB,  230
                 numbers of bytes. T or TB indicates TiB, 240 numbers of bytes.

                 request_GPUs

          request_<name> = <quantity>
                 The  required  amount  of the custom machine resource identified by <name> that this job needs.
                 The custom machine resource is defined in  the  machine's  configuration.  Machines  that  have
                 available GPUs will define <name> to be GPUs.  <name> must be at least two characters, and must
                 not  begin  with  _.   If <name> is either Cpu or Gpu a warning will be printed since these are
                 common typos.

          gpus_minimum_capability = <version> gpus_maximum_capability = <version>
                 The mininum or maximum required Capability value of the GPU, inclusive. Specified as a floating
                 point value (for example 8.5).  Use of one or more of these commands will create or modify  the
                 require_gpus   expression   unless   that   expression  already  references  the  GPU  property
                 Capabilities.  When request_gpus is not used, these commands are ignored.

          gpus_minimum_memory = <quantity>
                 The mininum quantity of GPU memory in MiB that a GPU must have in order to run the job.

                 Characters may be appended to a numerical value to indicate units.  K or KB indicates KiB,  210
                 numbers  of  bytes.  M  or  MB  indicates MiB, 220 numbers of bytes. G or GB indicates GiB, 230
                 numbers of bytes. T or TB indicates TiB, 240 numbers of bytes.

                 Use of this command will create or modify the require_gpus expression  unless  that  expression
                 already  references  the  GPU  property  GlobalMemoryMB.   When  request_gpus is not used, this
                 command is ignored.

          gpus_minimum_runtime = <version>
                 The version of the GPU (usually CUDA) runtime used  or  required  by  this  job,  specified  as
                 <major>.<minor>  (for example, 9.1).  If the minor version number is zero, you may specify only
                 the major version number.  A single version number of 1000 or  higher  is  assumed  to  be  the
                 integer-coded  version number using the Nvida convention of (major * 1000 + (minor * 10)).  Use
                 of this command will create or  modify  the  require_gpus  expression  unless  that  expression
                 already  references  the GPU property MaxSupportedVersion.  When request_gpus is not used, this
                 command is ignored.

          cuda_version = <version>
                 The version of the  CUDA  runtime,  if  any,  used  or  required  by  this  job,  specified  as
                 <major>.<minor>  (for example, 9.1).  If the minor version number is zero, you may specify only
                 the major version number.  A single version number of 1000 or  higher  is  assumed  to  be  the
                 integer-coded version number (major * 1000 + (minor % 100)).

                 This  command  has been superceeded by gpus_minimum_runtime, but it will still work if machines
                 advertise the CUDAMaxSupportedVersion attribute.  This does not arrange for the CUDA runtime to
                 be present, only for the job to run on a machine whose driver supports the specified version.

          requirements = <ClassAd Boolean Expression>
                 The requirements command is a boolean ClassAd expression which uses C-like operators. In  order
                 for  any  job  in  this  cluster  to  run on a given machine, this requirements expression must
                 evaluate to true on the given machine.

                 For scheduler and local universe jobs, the requirements expression  is  evaluated  against  the
                 Scheduler  ClassAd  which  represents the the condor_schedd daemon running on the access point,
                 rather than a remote machine. Like all commands in the submit  description  file,  if  multiple
                 requirements  commands are present, all but the last one are ignored. By default, condor_submit
                 appends the following clauses to the requirements expression:

                 1. Arch and OpSys are set equal to the Arch and OpSys of the submit machine.  In  other  words:
                    unless  you  request  otherwise,  HTCondor  will  give  your  job  machines  with  the  same
                    architecture and operating system version as the machine running condor_submit.

                 2. Cpus >= RequestCpus, if the job ClassAd attribute RequestCpus is defined.

                 3. Disk >= RequestDisk, if the job ClassAd attribute RequestDisk is defined. Otherwise, Disk >=
                    DiskUsage is appended to the requirements. The DiskUsage attribute  is  initialized  to  the
                    size  of  the  executable  plus  the  size  of any files specified in a transfer_input_files
                    command. It exists to ensure there is enough disk space on the target machine  for  HTCondor
                    to  copy over both the executable and needed input files. The DiskUsage attribute represents
                    the maximum amount  of  total  disk  space  required  by  the  job  in  kilobytes.  HTCondor
                    automatically  updates  the DiskUsage attribute approximately every 20 minutes while the job
                    runs with the amount of space being used by the job on the execute machine.

                 4. Memory >= RequestMemory, if the job ClassAd attribute RequestMemory is defined.

                 5. If Universe is set  to  Vanilla,  FileSystemDomain  is  set  equal  to  the  access  point's
                    FileSystemDomain.

                 View  the  requirements  of  a job which has already been submitted (along with everything else
                 about the job ClassAd) with the command condor_q -l; see the command  reference  for  condor_q.
                 Also,  see  the  HTCondor  Users  Manual  for  complete information on the syntax and available
                 attributes that can be used in the ClassAd expression.

       FILE TRANSFER COMMANDS

          dont_encrypt_input_files = < file1,file2,file... >
                 A comma and/or space separated list of input files that are not to be  network  encrypted  when
                 transferred  with  the file transfer mechanism. Specification of files in this manner overrides
                 configuration that would use encryption. Each input file must also be  in  the  list  given  by
                 transfer_input_files.   When  a path to an input file or directory is specified, this specifies
                 the path to the file on the submit side. A single wild card character (*) may be used  in  each
                 file name.

          dont_encrypt_output_files = < file1,file2,file... >
                 A  comma  and/or space separated list of output files that are not to be network encrypted when
                 transferred back with the file transfer  mechanism.  Specification  of  files  in  this  manner
                 overrides  configuration  that  would use encryption. The output file(s) must also either be in
                 the list given by transfer_output_files or be discovered and to be transferred  back  with  the
                 file  transfer  mechanism.  When  a  path  to  an  output  file or directory is specified, this
                 specifies the path to the file on the execute side. A single wild card  character  (*)  may  be
                 used in each file name.

          encrypt_execute_directory = <True | False>
                 Defaults  to  False.  If  set to True, HTCondor will encrypt the contents of the remote scratch
                 directory of the machine where the job is executed. This encryption is transparent to  the  job
                 itself,  but  ensures  that files left behind on the local disk of the execute machine, perhaps
                 due to a system crash, will remain private. In addition, condor_submit will append to the job's
                 requirements expression

                     && (TARGET.HasEncryptExecuteDirectory)

                 to ensure the job is matched to a machine that is capable of encrypting  the  contents  of  the
                 execute  directory.  This support is limited to Windows platforms that use the NTFS file system
                 and Linux platforms with the ecryptfs-utils package installed.

          encrypt_input_files = < file1,file2,file... >
                 A comma and/or space separated list of input files  that  are  to  be  network  encrypted  when
                 transferred  with the file transfer mechanism.  Specification of files in this manner overrides
                 configuration that would not use encryption. Each input file must also be in the list given  by
                 transfer_input_files.   When  a path to an input file or directory is specified, this specifies
                 the path to the file on the submit side. A single wild card character (*) may be used  in  each
                 file name. The method of encryption utilized will be as agreed upon in security negotiation; if
                 that  negotiation  failed,  then  the  file  transfer  mechanism must also fail for files to be
                 network encrypted.

          encrypt_output_files = < file1,file2,file... >
                 A comma and/or space separated list of output files that  are  to  be  network  encrypted  when
                 transferred  back  with  the  file  transfer  mechanism.  Specification of files in this manner
                 overrides configuration that would not use encryption. The output file(s) must also  either  be
                 in the list given by transfer_output_files or be discovered and to be transferred back with the
                 file  transfer  mechanism.  When  a  path  to  an  output  file or directory is specified, this
                 specifies the path to the file on the execute side. A single wild card  character  (*)  may  be
                 used  in  each  file name. The method of encryption utilized will be as agreed upon in security
                 negotiation; if that negotiation failed, then the file transfer mechanism must  also  fail  for
                 files to be network encrypted.

          erase_output_and_error_on_restart = < true | false>
                 If  false,  and when_to_transfer_output is ON_EXIT_OR_EVICT, HTCondor will append to the output
                 and error logs rather than erase (truncate) them when the job restarts.

          max_transfer_input_mb = <ClassAd Integer Expression>
                 This integer expression specifies the maximum allowed total size in MiB of the input files that
                 are transferred for a job. This expression does not apply to grid universe or files transferred
                 via file transfer plug-ins. The expression may refer to attributes  of  the  job.  The  special
                 value  -1  indicates  no  limit.  If  not  defined,  the  value  set  by configuration variable
                 MAX_TRANSFER_INPUT_MB is used. If the observed size of all input files at submit time is larger
                 than the limit, the job will be immediately placed on hold with a HoldReasonCode value  of  32.
                 If  the  job  passes  this initial test, but the size of the input files increases or the limit
                 decreases so that the limit is violated, the job will be placed on hold at the  time  when  the
                 file transfer is attempted.

          max_transfer_output_mb = <ClassAd Integer Expression>
                 This  integer  expression  specifies  the maximum allowed total size in MiB of the output files
                 that are transferred for a job. This expression does  not  apply  to  grid  universe  or  files
                 transferred  via file transfer plug-ins. The expression may refer to attributes of the job. The
                 special value -1 indicates no limit. If not  set,  the  value  set  by  configuration  variable
                 MAX_TRANSFER_OUTPUT_MB  is  used. If the total size of the job's output files to be transferred
                 is larger than the limit, the job will be placed on hold with a HoldReasonCode value of 33. The
                 output will be transferred up to the point when the limit is hit, so some files  may  be  fully
                 transferred, some partially, and some not at all.

          output_destination = <destination-URL>
                 When present, defines a URL that specifies both a plug-in and a destination for the transfer of
                 the  entire  output  sandbox  or  a  subset  of output files as specified by the submit command
                 transfer_output_files.  The plug-in does the transfer of files, and no files are sent  back  to
                 the access point. The HTCondor Administrator's manual has full details.

          should_transfer_files = <YES | NO | IF_NEEDED >
                 The  should_transfer_files  setting  is used to define if HTCondor should transfer files to and
                 from the remote machine where the job runs.  The file transfer mechanism is used to run jobs on
                 machines which do not have a shared file system with the submit machine.  should_transfer_files
                 equal to YES will cause HTCondor to always transfer files for the job. NO  disables  HTCondor's
                 file  transfer mechanism. IF_NEEDED will not transfer files for the job if it is matched with a
                 resource in the same FileSystemDomain as the access point (and therefore, on a machine with the
                 same shared file system). If the  job  is  matched  with  a  remote  resource  in  a  different
                 FileSystemDomain, HTCondor will transfer the necessary files.

                 For  more  information  about  this  and  other settings related to transferring files, see the
                 HTCondor User's manual section on the file transfer mechanism.

                 Note that should_transfer_files is not supported for jobs submitted to the grid universe.

          skip_filechecks = <True | False>
                 When True, file permission checks  for  the  submitted  job  are  disabled.  When  False,  file
                 permissions  are  checked;  this is the behavior when this command is not present in the submit
                 description file. File permissions are checked for read permissions on all input files, such as
                 those defined by commands input and transfer_input_files, and for write  permission  to  output
                 files,  such  as  a  log  file  defined  by  log  and  output  files  defined  with  output  or
                 transfer_output_files.

          stream_error = <True | False>
                 If True, then stderr is streamed back to the machine from  which  the  job  was  submitted.  If
                 False,  stderr  is  stored locally and transferred back when the job completes. This command is
                 ignored if the job ClassAd attribute TransferErr is False.  The default value  is  False.  This
                 command must be used in conjunction with error, otherwise stderr will sent to /dev/null on Unix
                 machines and ignored on Windows machines.

          stream_input = <True | False>
                 If  True,  then  stdin is streamed from the machine on which the job was submitted. The default
                 value is False. The command is only  relevant  for  jobs  submitted  to  the  vanilla  or  java
                 universes,  and  it  is  ignored by the grid universe. This command must be used in conjunction
                 with input, otherwise stdin will be /dev/null on Unix machines and ignored on Windows machines.

          stream_output = <True | False>
                 If True, then stdout is streamed back to the machine from  which  the  job  was  submitted.  If
                 False,  stdout  is  stored locally and transferred back when the job completes. This command is
                 ignored if the job ClassAd attribute TransferOut is False.  The default value  is  False.  This
                 command  must  be  used  in conjunction with output, otherwise stdout will sent to /dev/null on
                 Unix machines and ignored on Windows machines.

          transfer_executable = <True | False>
                 This  command  is  applicable  to  jobs  submitted  to  the  grid  and  vanilla  universes.  If
                 transfer_executable  is  set  to  False,  then  HTCondor looks for the executable on the remote
                 machine, and does not transfer the executable over. This is useful for  an  already  pre-staged
                 executable; HTCondor behaves more like rsh. The default value is True.

          transfer_input_files = < file1,file2,file... >
                 A  comma-delimited  list  of  all  the files and directories to be transferred into the working
                 directory for the job, before the job is  started.  By  default,  the  file  specified  in  the
                 executable  command  and  any  file  specified  in  the  input command (for example, stdin) are
                 transferred.

                 When a path to an input file or directory is specified, this specifies the path to the file  on
                 the  submit  side.  The  file is placed in the job's temporary scratch directory on the execute
                 side,  and  it  is  named  using  the  base  name  of   the   original   path.   For   example,
                 /path/to/input_file becomes input_file in the job's scratch directory.

                 When  a  directory  is  specified,  the  behavior  depends  on whether there is a trailing path
                 separator character.  When a directory is specified with a trailing path separator, it is as if
                 each of the items within the directory were  listed  in  the  transfer  list.   Therefore,  the
                 contents  are  transferred,  but  the  directory  itself is not. When there is no trailing path
                 separator, the directory itself is  transferred  with  all  of  its  contents  inside  it.   On
                 platforms  such  as  Windows  where  the  path separator is not a forward slash (/), a trailing
                 forward slash is treated as equivalent to a trailing path separator.  An example  of  an  input
                 directory specified with a trailing forward slash is input_data/.

                 For  grid  universe  jobs  other  than HTCondor-C, the transfer of directories is not currently
                 supported.

                 Symbolic links to files are transferred as the files they point to.  Transfer of symbolic links
                 to directories is not currently supported.

                 For vanilla and vm universe jobs only, a file may be specified by giving a URL,  instead  of  a
                 file  name.  The  implementation  for  URL  transfers requires both configuration and available
                 plug-in.

                 If you have a plugin which handles  https://  URLs  (and  HTCondor  ships  with  one  enabled),
                 HTCondor  supports  pre-signing  S3 URLs.  This allows you to specify S3 URLs for this command,
                 for transfer_output_remaps, and for output_destination.  By pre-signing the URLs on the  submit
                 node,  HTCondor  avoids transferring your S3 credentials to the execute node.  You must specify
                 aws_access_key_id_file  and  aws_secret_access_key_file;  you  may   specify   aws_region,   if
                 necessary;  see  below.   To  use  the S3 service provided by AWS, use S3 URLs of the following
                 forms:

                     # For older buckets that aren't region-specific.
                     s3://<bucket>/<key>

                     # For newer, region-specific buckets.
                     s3://<bucket>.s3.<region>.amazonaws.com/<key>

                 To use other S3 services, where <host> must contain a .:

                     s3://<host>/<key>

                     # If necessary
                     aws_region = <region>

                 You  may  specify  the   corresponding   access   key   ID   and   secret   access   key   with
                 s3_access_key_id_file  and s3_secret_access_key_file if you prefer (which may reduce confusion,
                 if you're not using AWS).

                 If you must access S3 using temporary credentials, you may specify  the  temporary  credentials
                 using  aws_access_key_id_file  and  aws_secret_access_key_file  for  the  files  containing the
                 corresponding temporary token, and +EC2SessionToken for the file containing the session token.

                 Temporary credentials have a limited lifetime.  If you are using  S3  only  to  download  input
                 files,  the job must start before the credentials expire.  If you are using S3 to upload output
                 files, the job must finish before the credentials expire.  HTCondor  does  not  know  when  the
                 credentials will expire; if they do so before they are needed, file transfer will fail.

                 HTCondor does not presently support transferring entire buckets or directories from S3.

                 HTCondor  supports  Google  Cloud Storage URLs -- gs:// -- via Google's "interoperability" API.
                 You may specify gs:// URLs as if they were s3:// URLs, and they work the  same  way.   You  may
                 specify  the  corresponding  access key ID and secret access key with gs_access_key_id_file and
                 gs_secret_access_key_file if you prefer (which may reduce confusion).

                 Note that (at present), you may not provide more than one set of credentials for s3:// or gs://
                 file transfer; this implies that all such URLs download from or upload to the same service.

          public_input_files = <file, file2>
                 A list of files on the AP that HTCondor should use a pre-configured HTTP server on  the  AP  to
                 transfer.  These  files  will  not be encrypted, and will be publically fetchable by anyone who
                 knows their name.

          transfer_output_files = < file1,file2,file... >
                 This command forms an explicit list of output files and directories to be transferred back from
                 the temporary working directory on the execute machine  to  the  access  point.  If  there  are
                 multiple  files, they must be delimited with commas. Setting transfer_output_files to the empty
                 string ("") means that no files are to be transferred.

                 For HTCondor-C jobs and all other non-grid  universe  jobs,  if  transfer_output_files  is  not
                 specified,  HTCondor  will automatically transfer back all files in the job's temporary working
                 directory which have been modified or created by the job.  Subdirectories are not  scanned  for
                 output,  so  if  output  from  subdirectories  is  desired,  the output list must be explicitly
                 specified. For grid universe jobs other than HTCondor-C, desired  output  files  must  also  be
                 explicitly  listed.  Another  reason  to explicitly list output files is for a job that creates
                 many files, and the user wants only a subset transferred back.

                 For grid universe jobs other than with grid type condor, to  have  files  other  than  standard
                 output and standard error transferred from the execute machine back to the access point, do use
                 transfer_output_files,  listing  all  files  to  be  transferred.  These files are found on the
                 execute machine in the working directory of the job.

                 When a path to an output file or directory is specified, it specifies the path to the  file  on
                 the  execute side. As a destination on the submit side, the file is placed in the job's initial
                 working directory, and it is named using the base name of  the  original  path.   For  example,
                 path/to/output_file  becomes  output_file  in the job's initial working directory. The name and
                 path  of  the  file  that  is  written  on  the  submit  side  may   be   modified   by   using
                 transfer_output_remaps.   Note  that  this  remap  function  only works with files but not with
                 directories.

                 When a directory is specified, the behavior  depends  on  whether  there  is  a  trailing  path
                 separator character.  When a directory is specified with a trailing path separator, it is as if
                 each  of  the  items  within  the  directory  were listed in the transfer list.  Therefore, the
                 contents are transferred, but the directory itself is not.  When  there  is  no  trailing  path
                 separator,  the  directory  itself  is  transferred  with  all  of  its contents inside it.  On
                 platforms such as Windows where the path separator is not  a  forward  slash  (/),  a  trailing
                 forward  slash  is  treated as equivalent to a trailing path separator.  An example of an input
                 directory specified with a trailing forward slash is input_data/.

                 For grid universe jobs other than HTCondor-C, the transfer  of  directories  is  not  currently
                 supported.

                 Symbolic links to files are transferred as the files they point to.  Transfer of symbolic links
                 to directories is not currently supported.

          transfer_checkpoint_files = < file1,file2,file3... >
                 If  present,  this  command  defines  the list of files and/or directories which constitute the
                 job's checkpoint.  When the job successfully checkpoints -- see checkpoint_exit_code  --  these
                 files will be transferred to the submit node's spool.

                 If this command is absent, the output is transferred instead.

                 If  no  files or directories are specified, nothing will be transferred.  This is generally not
                 useful.

                 The list is interpreted like  transfer_output_files,  but  there  is  no  corresponding  remaps
                 command.

          checkpoint_destination = <destination-URL>
                 When present, defines a URL that specifies both a plug-in and a destination for the transfer of
                 the  entire checkpoint of a job.  The plug-in does the transfer of files, and no files are sent
                 back to the access point.

          preserve_relative_paths = < True | False >
                 For vanilla and Docker -universe jobs (and others that use the shadow), this  command  modifies
                 the  behavior  of  the  file transfer commands.  When set to true, the destination for an entry
                 that is a relative path in a file transfer list becomes its relative path,  not  its  basename.
                 For  example,  input_data/b  (and  its  contents,  if it is a directory) will be transferred to
                 input_data/b, not b.  This applies to the input, output, and checkpoint lists.

                 Trailing slashes are ignored when preserve_relative_paths is set.

          transfer_output_remaps = < " name = newname ; name2 = newname2 ... ">
                 This specifies the name (and optionally path) to use when downloading  output  files  from  the
                 completed  job.  Normally,  output  files are transferred back to the initial working directory
                 with the same name they had in the execution directory. This gives you the option to save  them
                 with  a different path or name. If you specify a relative path, the final path will be relative
                 to the job's initial working directory, and HTCondor will create directories  as  necessary  to
                 transfer the file.

                 name describes an output file name produced by your job, and newname describes the file name it
                 should be downloaded to.  Multiple remaps can be specified by separating each with a semicolon.
                 If  you  wish  to  remap  file  names  that  contain  equals signs or semicolons, these special
                 characters may be escaped with a backslash. You cannot specify directories to be remapped.

                 Note that whether an output file is transferred is controlled by transfer_output_files. Listing
                 a file in transfer_output_remaps is not sufficient to cause it to be transferred.

          transfer_plugins = < tag=plugin ; tag2,tag3=plugin2 ... >
                 Specifies the file transfer plugins (see Third Party/Delegated file, credential and  checkpoint
                 transfer) that should be transferred along with the input files prior to invoking file transfer
                 plugins for files specified in transfer_input_files. tag should be a URL prefix that is used in
                 transfer_input_files,  and  plugin  is the path to a file transfer plugin that will handle that
                 type of URL transfer.

          when_to_transfer_output = < ON_EXIT | ON_EXIT_OR_EVICT | ON_SUCCESS >
                 Setting when_to_transfer_output to ON_EXIT will cause HTCondor to  transfer  the  job's  output
                 files  back  to  the submitting machine when the job completes (exits on its own).  If a job is
                 evicted and started again, the subsequent execution will start with  only  the  executable  and
                 input  files  in  the scratch directory sandbox.  If transfer_output_files is not set, HTCondor
                 considers all new files in the sandbox's top-level directory to be the  output;  subdirectories
                 and their contents will not be transferred.

                 Setting  when_to_transfer_output  to ON_EXIT_OR_EVICT will cause HTCondor to transfer the job's
                 output files when the job completes (exits on its own) and when the job is evicted.   When  the
                 job  is evicted, HTCondor will transfer the output files to a temporary directory on the submit
                 node (determined by the SPOOL configuration variable).  When the job restarts, these files will
                 be transferred instead of the input files.   If  transfer_output_files  is  not  set,  HTCondor
                 considers  all  files in the sandbox's top-level directory to be the output; subdirectories and
                 their contents will not be transferred.

                 Setting when_to_transfer_output to ON_SUCCESS will cause HTCondor to transfer the job's  output
                 files  when  the  job  completes  successfully.   Success  is  defined by the success_exit_code
                 command,  which  must  be  set,  even  if  the  successful  value  is  the   default   0.    If
                 transfer_output_files  is  not set, HTCondor considers all new files in the sandbox's top-level
                 directory to be the output; subdirectories and their contents will not be transferred.

                 In all three cases, the job will go on hold if transfer_output_files  specifies  a  file  which
                 does not exist at transfer time.

          aws_access_key_id_file, s3_access_key_id_file
                 One of these commands is required if you specify an s3:// URL; they specify the file containing
                 the access key ID (and only the access key ID) used to pre-sign the URLs.  Use only one.

          aws_secret_access_key_file, s3_secret_access_key_file
                 One of these commands is required if you specify an s3:// URL; they specify the file containing
                 the  secret  access  key  (and only the secret access key) used to pre-sign the URLs.  Use only
                 one.

          aws_region
                 Optional if you specify an S3 URL (and ignored otherwise), this command specifies the region to
                 use if one is not specified in the URL.

          gs_access_key_id_file
                 Required if you specify a gs:// URLs, this command specifies the file containing the access key
                 ID (and only the access key ID) used to pre-sign the URLs.

          gs_secret_access_key_file
                 Required if you specify a gs:// URLs, this command specifies the  file  containing  the  secret
                 access key (and only the secret access key) used to pre-sign the URLs.

       POLICY COMMANDS

          allowed_execute_duration = <integer>
                 The  longest time for which a job may be executing.  Jobs which exceed this duration will go on
                 hold.  This time does not include file-transfer time.  Jobs  which  self-checkpoint  have  this
                 long to write out each checkpoint.

                 This  attribute  is intended to help minimize the time wasted by jobs which may erroneously run
                 forever.

          allowed_job_duration = <integer>
                 The longest time for which a job may continuously be in the running state.  Jobs  which  exceed
                 this  duration will go on hold.  Exiting the running state resets the job duration used by this
                 command.

                 This command is intended to help minimize the time wasted by jobs  which  may  erroneously  run
                 forever.

          max_retries = <integer>
                 The  maximum  number  of  retries allowed for this job (must be non-negative). If the job fails
                 (does not exit with the success_exit_code exit code) it will be retried up to max_retries times
                 (unless retries are ceased because of the retry_until command). If max_retries is not  defined,
                 and  either  retry_until  or success_exit_code is, the value of DEFAULT_JOB_MAX_RETRIES will be
                 used for the maximum number of retries.

                 The combination of the max_retries,  retry_until,  and  success_exit_code  commands  causes  an
                 appropriate  OnExitRemove  expression  to  be  automatically generated. If retry command(s) and
                 on_exit_remove are both defined, the OnExitRemove expression will be generated  by  OR'ing  the
                 expression specified in OnExitRemove and the expression generated by the retry commands.

          retry_until = <Integer | ClassAd Boolean Expression>
                 An integer value or boolean expression that prevents further retries from taking place, even if
                 max_retries  have  not been exhausted.  If retry_until is an integer, the job exiting with that
                 exit code will cause retries to cease. If retry_until is a ClassAd expression,  the  expression
                 evaluating  to  True  will cause retries to cease.  For example, if you only want to retry exit
                 codes 17, 34, and 81:

                     max_retries = 5
                     retry_until = !member( ExitCode, {17, 34, 81} )

          success_exit_code = <integer>
                 The exit code that is considered successful for this job. Defaults to 0 if not defined.

                 Note: non-zero values of success_exit_code should generally not be  used  for  DAG  node  jobs,
                 unless  when_to_transfer_output  is  set  to  ON_SUCCESS in order to avoid failed jobs going on
                 hold.

                 At the present time, condor_dagman does not take into account the value  of  success_exit_code.
                 This  means  that, if success_exit_code is set to a non-zero value, condor_dagman will consider
                 the job failed when it actually succeeds. For single-proc DAG node jobs, this can  be  overcome
                 by using a POST script that takes into account the value of success_exit_code (although this is
                 not  recommended).  For  multi-proc  DAG  node jobs, there is currently no way to overcome this
                 limitation.

          checkpoint_exit_code = <integer>
                 The exit code which indicates that the  executable  has  exited  after  successfully  taking  a
                 checkpoint.    The   checkpoint   will   transferred   and   the   executable  restarted.   See
                 users-manual/self-checkpointing-applications:Self-Checkpointing Applications for details.

          hold = <True | False>
                 If hold is set to True, then the submitted job will be placed into the Hold state. Jobs in  the
                 Hold state will not run until released by condor_release. Defaults to False.

          keep_claim_idle = <integer>
                 An integer number of seconds that a job requests the condor_schedd to wait before releasing its
                 claim after the job exits or after the job is removed.

                 The  process  by  which the condor_schedd claims a condor_startd is somewhat time-consuming. To
                 amortize this cost, the condor_schedd tries to reuse claims to run subsequent jobs, after a job
                 using a claim is done. However, it can only do this if there is an idle job in the queue at the
                 moment the previous job completes. Sometimes, and especially  for  the  node  jobs  when  using
                 DAGMan,  there  is  a  subsequent  job about to be submitted, but it has not yet arrived in the
                 queue when the previous job completes. As a result, the condor_schedd releases the  claim,  and
                 the  next  job  must  wait  an  entire  negotiation cycle to start. When this submit command is
                 defined with a non-negative integer, when the job exits, the condor_schedd tries  as  usual  to
                 reuse  the  claim.  If  it  cannot, instead of releasing the claim, the condor_schedd keeps the
                 claim until either the number of seconds given as a parameter, or a new job which matches  that
                 claim  arrives,  whichever  comes  first.  The  condor_startd  in  question  will remain in the
                 Claimed/Idle state, and the original job will be "charged" (in terms of priority) for the  time
                 in this state.

          leave_in_queue = <ClassAd Boolean Expression>
                 When  the  ClassAd  Expression  evaluates  to  True, the job is not removed from the queue upon
                 completion. This allows the user of a remotely spooled job to retrieve output  files  in  cases
                 where  HTCondor  would have removed them as part of the cleanup associated with completion. The
                 job will only exit the queue once it has been marked for removal (via condor_rm,  for  example)
                 and the leave_in_queue, expression has become False.  leave_in_queue defaults to False.

                 As   an   example,   if   the  job  is  to  be  removed  once  the  output  is  retrieved  with
                 condor_transfer_data, then use

                     leave_in_queue = (JobStatus == 4) && ((StageOutFinish =?= UNDEFINED) ||\
                                      (StageOutFinish == 0))

          next_job_start_delay = <ClassAd Boolean Expression>
                 This expression specifies the number of seconds to delay after starting up this job before  the
                 next  job  is  started.  The  maximum  allowed delay is specified by the HTCondor configuration
                 variable MAX_NEXT_JOB_START_DELAY, which defaults to 10 minutes. This command does not apply to
                 scheduler or local universe jobs.

                 This command has been historically used to implement a form of job start  throttling  from  the
                 job submitter's perspective. It was effective for the case of multiple job submission where the
                 transfer  of  extremely large input data sets to the execute machine caused machine performance
                 to suffer. This command is no longer useful,  as  throttling  should  be  accomplished  through
                 configuration of the condor_schedd daemon.

          on_exit_hold = <ClassAd Boolean Expression>
                 The ClassAd expression is checked when the job exits, and if True, places the job into the Hold
                 state.  If  False  (the  default  value  when  not  defined),  then  nothing  happens  and  the
                 on_exit_remove expression is checked to determine if that needs to be applied.

                 For example: Suppose a job is known to run for a minimum of an hour.  If the  job  exits  after
                 less than an hour, the job should be placed on hold and an e-mail notification sent, instead of
                 being allowed to leave the queue.

                     on_exit_hold = (time() - JobStartDate) < (60 * $(MINUTE))

                 This  expression  places the job on hold if it exits for any reason before running for an hour.
                 An e-mail will be sent to the user explaining that the job was  placed  on  hold  because  this
                 expression became True.

                 periodic_*  expressions take precedence over on_exit_* expressions, and *_hold expressions take
                 precedence over a *_remove expressions.

                 Only job ClassAd attributes will be defined for use by this ClassAd expression. This expression
                 is available for the vanilla, java, parallel, grid, local and scheduler universes.

          on_exit_hold_reason = <ClassAd String Expression>
                 When the job is placed  on  hold  due  to  the  on_exit_hold  expression  becoming  True,  this
                 expression  is  evaluated to set the value of HoldReason in the job ClassAd. If this expression
                 is UNDEFINED or produces an empty or invalid string, a default description is used.

          on_exit_hold_subcode = <ClassAd Integer Expression>
                 When the job is placed  on  hold  due  to  the  on_exit_hold  expression  becoming  True,  this
                 expression  is  evaluated to set the value of HoldReasonSubCode in the job ClassAd. The default
                 subcode is 0. The HoldReasonCode will be set to 3, which indicates that the job  went  on  hold
                 due to a job policy expression.

          on_exit_remove = <ClassAd Boolean Expression>
                 The  ClassAd  expression  is  checked  when  the job exits, and if True (the default value when
                 undefined), then it allows the job to leave the queue normally.  If  False,  then  the  job  is
                 placed  back into the Idle state. If the user job runs under the vanilla universe, then the job
                 restarts from the beginning.

                 For example, suppose a job occasionally segfaults, but chances are that  the  job  will  finish
                 successfully  if  the  job  is  run again with the same data. The on_exit_remove expression can
                 cause the job to run again with the following command. Assume that the  signal  identifier  for
                 the segmentation fault is 11 on the platform where the job will be running.

                     on_exit_remove = (ExitBySignal == False) || (ExitSignal != 11)

                 This expression lets the job leave the queue if the job was not killed by a signal or if it was
                 killed  by  a signal other than 11, representing segmentation fault in this example. So, if the
                 exited due to signal 11, it will stay in the job queue. In any other case of the  job  exiting,
                 the job will leave the queue as it normally would have done.

                 As  another example, if the job should only leave the queue if it exited on its own with status
                 0, this on_exit_remove expression works well:

                     on_exit_remove = (ExitBySignal == False) && (ExitCode == 0)

                 If the job was killed by a signal or exited with a non-zero exit status, HTCondor  would  leave
                 the job in the queue to run again.

                 periodic_*  expressions take precedence over on_exit_* expressions, and *_hold expressions take
                 precedence over a *_remove expressions.

                 Only job ClassAd attributes will be defined for use by this ClassAd expression.

          periodic_hold = <ClassAd Boolean Expression>
                 This expression is checked periodically when the job is not in the Held state.  If  it  becomes
                 True, the job will be placed on hold.  If unspecified, the default value is False.

                 periodic_*  expressions take precedence over on_exit_* expressions, and *_hold expressions take
                 precedence over a *_remove expressions.

                 Only job ClassAd attributes will be defined for use by this ClassAd expression. Note  that,  by
                 default, this expression is only checked once every 60 seconds. The period of these evaluations
                 can   be  adjusted  by  setting  the  PERIODIC_EXPR_INTERVAL,  MAX_PERIODIC_EXPR_INTERVAL,  and
                 PERIODIC_EXPR_TIMESLICE configuration macros.

          periodic_hold_reason = <ClassAd String Expression>
                 When the job is placed on  hold  due  to  the  periodic_hold  expression  becoming  True,  this
                 expression  is  evaluated to set the value of HoldReason in the job ClassAd. If this expression
                 is UNDEFINED or produces an empty or invalid string, a default description is used.

          periodic_hold_subcode = <ClassAd Integer Expression>
                 When the job is placed on  hold  due  to  the  periodic_hold  expression  becoming  true,  this
                 expression  is  evaluated to set the value of HoldReasonSubCode in the job ClassAd. The default
                 subcode is 0. The HoldReasonCode will be set to 3, which indicates that the job  went  on  hold
                 due to a job policy expression.

          periodic_release = <ClassAd Boolean Expression>
                 This  expression  is  checked periodically when the job is in the Held state. If the expression
                 becomes True,  the  job  will  be  released.   If  the  job  was  held  via  condor_hold  (i.e.
                 HoldReasonCode is 1), then this expression is ignored.

                 Only  job  ClassAd attributes will be defined for use by this ClassAd expression. Note that, by
                 default, this expression is only checked once every 60 seconds. The period of these evaluations
                 can  be  adjusted  by  setting  the  PERIODIC_EXPR_INTERVAL,  MAX_PERIODIC_EXPR_INTERVAL,   and
                 PERIODIC_EXPR_TIMESLICE configuration macros.

          periodic_remove = <ClassAd Boolean Expression>
                 This expression is checked periodically. If it becomes True, the job is removed from the queue.
                 If unspecified, the default value is False.

                 See the Examples section of this manual page for an example of a periodic_remove expression.

                 periodic_*  expressions take precedence over on_exit_* expressions, and *_hold expressions take
                 precedence over a *_remove expressions. So, the periodic_remove expression takes precedent over
                 the on_exit_remove expression, if the two describe conflicting actions.

                 Only job ClassAd attributes will be defined for use by this ClassAd expression. Note  that,  by
                 default, this expression is only checked once every 60 seconds. The period of these evaluations
                 can   be  adjusted  by  setting  the  PERIODIC_EXPR_INTERVAL,  MAX_PERIODIC_EXPR_INTERVAL,  and
                 PERIODIC_EXPR_TIMESLICE configuration macros.

          periodic_vacate = <ClassAd Boolean Expression>
                 This expression is checked periodically for running jobs. If it becomes True,  job  is  evicted
                 from  the  machine it is running on, and return to the queue, in an Idle state. If unspecified,
                 the default value is False.

                 Only job ClassAd attributes will be defined for use by this ClassAd expression. Note  that,  by
                 default, this expression is only checked once every 60 seconds. The period of these evaluations
                 can   be  adjusted  by  setting  the  PERIODIC_EXPR_INTERVAL,  MAX_PERIODIC_EXPR_INTERVAL,  and
                 PERIODIC_EXPR_TIMESLICE configuration macros.

       COMMANDS FOR THE GRID

          arc_application = <XML-string>
                 For grid universe jobs of type arc, provides additional XML attributes under the  <Application>
                 section of the ARC ADL job description which are not covered by regular submit description file
                 parameters.

          arc_resources = <XML-string>
                 For  grid  universe  jobs of type arc, provides additional XML attributes under the <Resources>
                 section of the ARC ADL job description which are not covered by regular submit description file
                 parameters.

          arc_rte = < rte1 option,rte2 >
                 For grid universe jobs of type arc, provides a list of Runtime Environment names that  the  job
                 requires  on  the  ARC  system.   The  list  is  comma-delimited. If a Runtime Environment name
                 supports options, those  can  be  provided  after  the  name,  separated  by  spaces.   Runtime
                 Environment names are defined by the ARC server.

          azure_admin_key = <pathname>
                 For  grid  type  azure  jobs,  specifies  the path and file name of a file that contains an SSH
                 public key. This key can be used to log into the administrator account of the instance via SSH.

          azure_admin_username = <account name>
                 For grid type azure jobs, specifies the name of an administrator account to be created  in  the
                 instance. This account can be logged into via SSH.

          azure_auth_file = <pathname>
                 For  grid type azure jobs, specifies a path and file name of the authorization file that grants
                 permission for HTCondor to use the Azure account. If  it's  not  defined,  then  HTCondor  will
                 attempt to use the default credentials of the Azure CLI tools.

          azure_image = <image id>
                 For  grid  type  azure  jobs,  identifies  the  disk  image to be used for the boot disk of the
                 instance. This image must already be registered within Azure.

          azure_location = <image id>
                 For grid type azure jobs, identifies the location within Azure where  the  instance  should  be
                 run. As an example, one current location is centralus.

          azure_size = <machine type>
                 For  grid  type  azure jobs, the hardware configuration that the virtual machine instance is to
                 run on.

          batch_extra_submit_args = <command-line arguments>
                 Used for batch grid universe jobs.  Specifies additional command-line arguments to be given  to
                 the target batch system's job submission command.

          batch_project = <projectname>
                 Used  for  batch  grid  universe  jobs.   Specifies  the name of the PBS/LSF/SGE/SLURM project,
                 account, or allocation that should be charged for the resources used by the job.

          batch_queue = <queuename>
                 Used for batch grid universe jobs.  Specifies the name of the PBS/LSF/SGE/SLURM job queue  into
                 which  the  job  should  be  submitted.  If  not  specified,  the default queue is used.  For a
                 multi-cluster SLURM configuration, which cluster to use can be specified by supplying the  name
                 after  an  @ symbol.  For example, to submit a job to the debug queue on cluster foo, you would
                 use the value debug@foo.

          batch_runtime = <seconds>
                 Used for batch grid universe jobs.  Specifies a limit in seconds on the execution time  of  the
                 job.  This limit is enforced by the PBS/LSF/SGE/SLURM scheduler.

          cloud_label_names = <name0,name1,name...>
                 For  grid  type  gce  jobs,  specifies  the  case of tag names that will be associated with the
                 running instance. This is only necessary if a tag name case matters. By default the  list  will
                 be automatically generated.

          cloud_label_<name> = <value>
                 For grid type gce jobs, specifies a label and value to be associated with the running instance.
                 The label name will be lower-cased; use cloud_label_names to change the case.

          delegate_job_GSI_credentials_lifetime = <seconds>
                 Specifies  the  maximum  number  of  seconds  for  which delegated proxies should be valid. The
                 default behavior when this command is not specified is determined by the configuration variable
                 DELEGATE_JOB_GSI_CREDENTIALS_LIFETIME, which defaults to one day. A value of 0  indicates  that
                 the delegated proxy should be valid for as long as allowed by the credential used to create the
                 proxy.  This  setting  currently  only  applies  to proxies delegated for non-grid jobs and for
                 HTCondor-C   jobs.    This   variable   has   no   effect   if   the   configuration   variable
                 DELEGATE_JOB_GSI_CREDENTIALS is False, because in that case the job proxy is copied rather than
                 delegated.

          ec2_access_key_id = <pathname>
                 For grid type ec2 jobs, identifies the file containing the access key.

          ec2_ami_id = <EC2 xMI ID>
                 For  grid  type  ec2 jobs, identifies the machine image. Services compatible with the EC2 Query
                 API may refer to these with abbreviations  other  than  AMI,  for  example  EMI  is  valid  for
                 Eucalyptus.

          ec2_availability_zone = <zone name>
                 For  grid  type  ec2  jobs, specifies the Availability Zone that the instance should be run in.
                 This command is optional, unless ec2_ebs_volumes is set. As an example,  one  current  zone  is
                 us-east-1b.

          ec2_block_device_mapping = <block-device>:<kernel-device>,<block-device>:<kernel-device>, ...
                 For  grid  type  ec2 jobs, specifies the block device to kernel device mapping. This command is
                 optional.

          ec2_ebs_volumes = <ebs name>:<device name>,<ebs name>:<device name>,...
                 For grid type ec2 jobs, optionally specifies a list of Elastic Block Store (EBS) volumes to  be
                 made available to the instance and the device names they should have in the instance.

          ec2_elastic_ip = <elastic IP address>
                 For  grid  type  ec2  jobs,  and optional specification of an Elastic IP address that should be
                 assigned to this instance.

          ec2_iam_profile_arn = <IAM profile ARN>
                 For grid type ec2 jobs, an Amazon Resource Name (ARN) identifying  which  Identity  and  Access
                 Management (IAM) (instance) profile to associate with the instance.

          ec2_iam_profile_name = <IAM profile name>
                 For  grid  type  ec2  jobs,  a  name  identifying  which  Identity  and Access Management (IAM)
                 (instance) profile to associate with the instance.

          ec2_instance_type = <instance type>
                 For grid type ec2 jobs, identifies the instance type. Different services  may  offer  different
                 instance types, so no default value is set.

          ec2_keypair = <ssh key-pair name>
                 For  grid  type ec2 jobs, specifies the name of an SSH key-pair that is already registered with
                 the EC2 service. The associated private key can be used to ssh into the virtual machine once it
                 is running.

          ec2_keypair_file = <pathname>
                 For grid type ec2 jobs, specifies the complete path and file name of a file into which HTCondor
                 will write an SSH key for use with ec2 jobs. The key can  be  used  to  ssh  into  the  virtual
                 machine once it is running. If ec2_keypair is specified for a job, ec2_keypair_file is ignored.

          ec2_parameter_names = ParameterName1, ParameterName2, ...
                 For  grid  type ec2 jobs, a space or comma separated list of the names of additional parameters
                 to pass when instantiating an instance.

          ec2_parameter_<name> = <value>
                 For  grid  type  ec2  jobs,  specifies  the  value  for  the  correspondingly  named  (instance
                 instantiation)  parameter.  <name>  is  the  parameter  name  specified  in  the submit command
                 ec2_parameter_names, but with any periods replaced by underscores.

          ec2_secret_access_key = <pathname>
                 For grid type ec2 jobs, specifies the path and file name containing the secret access key.

          ec2_security_groups = group1, group2, ...
                 For grid type ec2 jobs, defines the list of EC2 security groups which should be associated with
                 the job.

          ec2_security_ids = id1, id2, ...
                 For grid type ec2 jobs, defines the list of EC2 security group IDs which should  be  associated
                 with the job.

          ec2_spot_price = <bid>
                 For  grid  type  ec2  jobs,  specifies  the  spot  instance bid, which is the most that the job
                 submitter is willing to pay per hour to run this job.

          ec2_tag_names = <name0,name1,name...>
                 For grid type ec2 jobs, specifies the case of tag  names  that  will  be  associated  with  the
                 running  instance.  This is only necessary if a tag name case matters. By default the list will
                 be automatically generated.

          ec2_tag_<name> = <value>
                 For grid type ec2 jobs, specifies a tag to be associated with the  running  instance.  The  tag
                 name will be lower-cased; use ec2_tag_names to change the case.

          WantNameTag = <True | False>
                 For  grid type ec2 jobs, a job may request that its 'name' tag be (not) set by HTCondor. If the
                 job does not otherwise specify any tags, not setting its name tag will eliminate a call by  the
                 EC2 GAHP, improving performance.

          ec2_user_data = <data>
                 For  grid  type ec2 jobs, provides a block of data that can be accessed by the virtual machine.
                 If both ec2_user_data and ec2_user_data_file are specified for a job, the two  blocks  of  data
                 are concatenated, with the data from this ec2_user_data submit command occurring first.

          ec2_user_data_file = <pathname>
                 For  grid  type  ec2 jobs, specifies a path and file name whose contents can be accessed by the
                 virtual machine. If both ec2_user_data and ec2_user_data_file are specified for a job, the  two
                 blocks of data are concatenated, with the data from that ec2_user_data submit command occurring
                 first.

          ec2_vpc_ip = <a.b.c.d>
                 For  grid  type  ec2  jobs,  that  are  part  of  a  Virtual  Private  Cloud (VPC), an optional
                 specification of the IP address that this instance should have within the VPC.

          ec2_vpc_subnet = <subnet specification string>
                 For grid type ec2 jobs, an optional specification of the Virtual Private Cloud (VPC) that  this
                 instance should be a part of.

          gce_account = <account name>
                 For  grid  type  gce  jobs,  specifies the Google cloud services account to use. If this submit
                 command  isn't  specified,  then  a  random  account  from  the  authorization  file  given  by
                 gce_auth_file will be used.

          gce_auth_file = <pathname>
                 For  grid  type  gce jobs, specifies a path and file name of the authorization file that grants
                 permission for HTCondor to use the Google account. If this command is not specified,  then  the
                 default file of the Google command-line tools will be used.

          gce_image = <image id>
                 For  grid  type gce jobs, the identifier of the virtual machine image representing the HTCondor
                 job to be run. This virtual machine image must already be  register  with  GCE  and  reside  in
                 Google's Cloud Storage service.

          gce_json_file = <pathname>
                 For  grid type gce jobs, specifies the path and file name of a file that contains JSON elements
                 that should be added to the instance description submitted to the GCE service.

          gce_machine_type = <machine type>
                 For grid type gce jobs, the long form of the URL that describes the machine configuration  that
                 the virtual machine instance is to run on.

          gce_metadata = <name=value,...,name=value>
                 For grid type gce jobs, a comma separated list of name and value pairs that define metadata for
                 a virtual machine instance that is an HTCondor job.

          gce_metadata_file = <pathname>
                 For grid type gce jobs, specifies a path and file name of the file that contains metadata for a
                 virtual  machine instance that is an HTCondor job. Within the file, each name and value pair is
                 on its own line; so, the pairs are separated by the newline character.

          gce_preemptible = <True | False>
                 For grid type gce jobs, specifies whether the virtual machine instance should  be  preemptible.
                 The default is for the instance to not be preemptible.

          grid_resource = <grid-type-string> <grid-specific-parameter-list>
                 For  each  grid-type-string  value, there are further type-specific values that must specified.
                 This submit description file command allows  each  to  be  given  in  a  space-separated  list.
                 Allowable  grid-type-string  values  are arc, azure, batch, condor, ec2, and gce.  The HTCondor
                 manual chapter on Grid Computing details the variety of grid types.

                 For a grid-type-string of batch, the single parameter is the name of the  local  batch  system,
                 and will be one of pbs, lsf, slurm, or sge.

                 For  a  grid-type-string of condor, the first parameter is the name of the remote condor_schedd
                 daemon. The second parameter is the name of the pool to which the remote  condor_schedd  daemon
                 belongs.

                 For a grid-type-string of ec2, one additional parameter specifies the EC2 URL.

                 For a grid-type-string of arc, the single parameter is the name of the ARC resource to be used.

          transfer_error = <True | False>
                 For jobs submitted to the grid universe only. If True, then the error output (from stderr) from
                 the  job  is transferred from the remote machine back to the access point. The name of the file
                 after transfer is given by the error command. If False,  no  transfer  takes  place  (from  the
                 remote  machine  to  access point), and the name of the file is given by the error command. The
                 default value is True.

          transfer_input = <True | False>
                 For jobs submitted to the  grid  universe  only.  If  True,  then  the  job  input  (stdin)  is
                 transferred from the machine where the job was submitted to the remote machine. The name of the
                 file that is transferred is given by the input command. If False, then the job's input is taken
                 from  a  pre-staged  file on the remote machine, and the name of the file is given by the input
                 command. The default value is True.

                 For transferring files other than stdin, see transfer_input_files.

          transfer_output = <True | False>
                 For jobs submitted to the grid universe only. If True, then the output (from stdout)  from  the
                 job is transferred from the remote machine back to the access point. The name of the file after
                 transfer  is  given  by  the output command. If False, no transfer takes place (from the remote
                 machine to access point), and the name of the file is given by the output command. The  default
                 value is True.

                 For transferring files other than stdout, see transfer_output_files,

          use_x509userproxy = <True | False>
                 Set  this  command  to  True  to  indicate  that  the  job  requires  an  X.509  user proxy. If
                 x509userproxy is set, then that file is used for the proxy. Otherwise, the proxy is looked  for
                 in the standard locations. If x509userproxy is set or if the job is a grid universe job of grid
                 type arc, then the value of use_x509userproxy is forced to True. Defaults to False.

          x509userproxy = <full-pathname>
                 Used  to  override the default path name for X.509 user certificates.  The default location for
                 X.509 proxies is the /tmp directory, which is generally a local file system. Setting this value
                 would allow HTCondor to access the proxy in a shared file system (for example,  AFS).  HTCondor
                 will  use  the proxy specified in the submit description file first. If nothing is specified in
                 the submit description file, it will use the  environment  variable  X509_USER_PROXY.  If  that
                 variable  is  not  present,  it will search in the default location. Note that proxies are only
                 valid for a limited time. Condor_submit will not submit a job with an expired  proxy,  it  will
                 return  an error. Also, if the configuration parameter CRED_MIN_TIME_LEFT is set to some number
                 of seconds, and if the proxy will expire before that  many  seconds,  condor_submit  will  also
                 refuse  to  submit  the  job.  That  is, if CRED_MIN_TIME_LEFT is set to 60, condor_submit will
                 refuse to submit a job whose proxy will expire 60 seconds from the time of submission.

                 x509userproxy is relevant when the universe is vanilla, or when the universe is  grid  and  the
                 type of grid system is one of condor, or arc. Defining a value causes the proxy to be delegated
                 to  the  execute machine.  Further, VOMS attributes defined in the proxy will appear in the job
                 ClassAd.

          use_scitokens = <True | False | Auto>
                 Set this command to True to indicate that the job requires a scitoken.   If  scitokens_file  is
                 set,  then that file is used for the scitoken filename. Otherwise, the the scitoken filename is
                 looked for in the BEARER_TOKEN_FILE environment variable. If scitokens_file  is  set  then  the
                 value of use_scitokens defaults to True.  If the filename is not defined in on one of these two
                 places,  then  condor_submit  will  fail  with  an  error message.  Set this command to Auto to
                 indicate that  the  job  will  use  a  scitoken  if  scitokens_file  or  the  BEARER_TOKEN_FILE
                 environment variable is set, but it will not be an error if no file is specified.

          scitokens_file = <full-pathname>
                 Used to set the path to the file containing the scitoken that the job needs, or to override the
                 path to the scitoken contained in the BEARER_TOKEN_FILE environment variable.

       COMMANDS FOR PARALLEL, JAVA, and SCHEDULER UNIVERSES

          hold_kill_sig = <signal-number>
                 For  the scheduler universe only, signal-number is the signal delivered to the job when the job
                 is put on hold with condor_hold.  signal-number may be either  the  platform-specific  name  or
                 value of the signal. If this command is not present, the value of kill_sig is used.

          jar_files = <file_list>
                 Specifies  a  list  of  additional JAR files to include when using the Java universe. JAR files
                 will be transferred along with the executable and automatically added to the classpath.

          java_vm_args = <argument_list>
                 Specifies a list of additional arguments to the Java VM itself, When  HTCondor  runs  the  Java
                 program,  these  are  the  arguments  that  go  before  the class name. This can be used to set
                 VM-specific arguments like stack size, garbage-collector arguments and initial property values.

          machine_count = <max>
                 For the parallel universe, a single value (max)  is  required.  It  is  neither  a  maximum  or
                 minimum, but the number of machines to be dedicated toward running the job.

          remove_kill_sig = <signal-number>
                 For  the scheduler universe only, signal-number is the signal delivered to the job when the job
                 is removed with condor_rm.  signal-number may be either the platform-specific name or value  of
                 the signal.  This example shows it both ways for a Linux signal:

                     remove_kill_sig = SIGUSR1
                     remove_kill_sig = 10

                 If this command is not present, the value of kill_sig is used.

       COMMANDS FOR THE VM UNIVERSE

          vm_disk = file1:device1:permission1, file2:device2:permission2:format2, ...
                 A  list of comma separated disk files. Each disk file is specified by 4 colon separated fields.
                 The first field is the path and file name of the disk file.  The  second  field  specifies  the
                 device.  The  third  field  specifies  permissions, and the optional fourth field specifies the
                 image format. If a disk file will be transferred by HTCondor, then the first field should  just
                 be the simple file name (no path information).

                 An example that specifies two disk files:

                     vm_disk = /myxen/diskfile.img:sda1:w,/myxen/swap.img:sda2:w

          vm_checkpoint = <True | False>
                 A  boolean  value  specifying whether or not to take checkpoints. If not specified, the default
                 value is False. In the current implementation, setting both vm_checkpoint and vm_networking  to
                 True  does  not  yet  work  in all cases. Networking cannot be used if a vm universe job uses a
                 checkpoint in order to continue execution after migration to another machine.

          vm_macaddr = <MACAddr>
                 Defines that MAC address that the virtual machine's  network  interface  should  have,  in  the
                 standard format of six groups of two hexadecimal digits separated by colons.

          vm_memory = <MBytes-of-memory>
                 The amount of memory in MBytes that a vm universe job requires.

          vm_networking = <True | False>
                 Specifies  whether  to  use  networking  or  not.  In  the current implementation, setting both
                 vm_checkpoint and vm_networking to True does not yet work in all cases.  Networking  cannot  be
                 used  if  a vm universe job uses a checkpoint in order to continue execution after migration to
                 another machine.

          vm_networking_type = <nat | bridge >
                 When vm_networking is True, this definition augments  the  job's  requirements  to  match  only
                 machines with the specified networking. If not specified, then either networking type matches.

          vm_no_output_vm = <True | False>
                 When  True, prevents HTCondor from transferring output files back to the machine from which the
                 vm universe job was submitted. If not specified, the default value is False.

          vm_type = <xen | kvm>
                 Specifies the underlying virtual machine software that this job expects.

          xen_initrd = <image-file>
                 When xen_kernel gives a file name for the kernel  image  to  use,  this  optional  command  may
                 specify  a  path  to  a  ramdisk  (initrd) image file. If the image file will be transferred by
                 HTCondor, then the value should just be the simple file name (no path information).

          xen_kernel = <included | path-to-kernel>
                 A value of included specifies that the kernel is included in the disk file. If not one of these
                 values, then the value is a path and file name of the kernel to be used. If a kernel file  will
                 be  transferred  by  HTCondor,  then  the  value  should  just be the simple file name (no path
                 information).

          xen_kernel_params = <string>
                 A string that is appended to the Xen kernel command line.

          xen_root = <string>
                 A string that is appended to the Xen kernel command line  to  specify  the  root  device.  This
                 string  is  required  when xen_kernel gives a path to a kernel. Omission for this required case
                 results in an error message during submission.

       COMMANDS FOR THE DOCKER UNIVERSE

          docker_image = < image-name >
                 Defines the name of the Docker image that is the basis for the docker container.

          docker_network_type = < host | none | custom_admin_defined_value>
                 If docker_network_type is set to the string host, then the job is run using the host's network.
                 If docker_network_type is set to the string none, then the job is run with no network. If  this
                 is  not  set,  each  job gets a private network interface.  Some administrators may define site
                 specific docker networks on a given worker node.  When this is the case, additional values  may
                 be valid here.

          docker_pull_policy = < always >
                 if docker_pull_policy is set to always, when a docker universe job starts on a worker node, the
                 option  "--pull  always"  will  be  passed to the docker run command.  This only impacts worker
                 nodes which already have a locally cached version of the image.  With this option, docker  will
                 always  check  with  the  repo to see if the cached version is out of date.  This requires more
                 network connectivity, and may cause docker  hub  to  throttle  future  pull  requests.   It  is
                 generally recommened to never mutate docker image tag name, and avoid needing this option.

          container_service_names = <service-name>[, <service-name>]*
                 A  string-  or  comma-  separated  list  of  service  names.   Each  service-name  must  have a
                 corresponding <service-name>_container_port command specifying a port number (an integer from 0
                 to 65535).  HTCondor will ask Docker to forward from a host port to the specified  port  inside
                 the  container.  When Docker has done so, HTCondor will add an attribute to the job ad for each
                 service, <service-name>HostPort, which contains the port number on the host forwarding  to  the
                 corresponding service.

          <service-name>_container_port = port_number
                 See above.

          <service-name>_HostPort = port_number
                 See above.

          docker_override_entrypoint = <True | False>
                 If  docker_override_entrypoint is set to True and executable is not empty, the image entrypoint
                 is replaced with the executable.  The default value (False)  follows  the  same  logic  as  the
                 docker engine uses with images (see docker run):

                     • Without entrypoint, executable runs as main PID

                     • With entrypoint, it is launched with the excutable as first argument

                 Any additional arguments will follow the executable.

       COMMANDS FOR THE CONTAINER UNIVERSE

          container_image = < image-name >
                 Defines the name of the container image. Can be a singularity .sif file, a singularity exploded
                 directory, or a path to an image in a docker style repository

          container_target_dir = < path-to-directory-inside-container >
                 Defines  the  working directory of the job inside the container.  Will be mapped to the scratch
                 directory on the worker node.

       ADVANCED COMMANDS

          accounting_group = <accounting-group-name>
                 Causes jobs to negotiate under the given accounting group. This value is advertised in the  job
                 ClassAd  as  AcctGroup.  The  HTCondor  Administrator's  manual contains more information about
                 accounting groups.

          accounting_group_user = <accounting-group-user-name>
                 Sets the name associated with this job to be used for resource usage accounting purposes,  such
                 as  computation of fair-share priority and reporting via condor_userprio.  If not set, defaults
                 to the value of the job ClassAd attribute User. This value is advertised in the job ClassAd  as
                 AcctGroupUser.

          concurrency_limits = <string-list>
                 A  list of resources that this job needs. The resources are presumed to have concurrency limits
                 placed upon them, thereby limiting the number of concurrent jobs in execution  which  need  the
                 named  resource.  Commas  and space characters delimit the items in the list.  Each item in the
                 list is a string that identifies the limit, or it is a ClassAd expression that evaluates  to  a
                 string, and it is evaluated in the context of machine ClassAd being considered as a match. Each
                 item in the list also may specify a numerical value identifying the integer number of resources
                 required  for  the  job.  The syntax follows the resource name by a colon character (:) and the
                 numerical value. Details on concurrency limits are in the HTCondor Administrator's manual.

          concurrency_limits_expr = <ClassAd String Expression>
                 A ClassAd expression  that  represents  the  list  of  resources  that  this  job  needs  after
                 evaluation.  The  ClassAd  expression may specify machine ClassAd attributes that are evaluated
                 against a matched machine. After evaluation, the list sets concurrency_limits.

          copy_to_spool = <True | False>
                 If copy_to_spool is True, then condor_submit copies the executable to the local spool directory
                 before running it on a remote host. As copying can be quite time consuming and unnecessary, the
                 default value is False for all job universes.  When False,  condor_submit  does  not  copy  the
                 executable to a local spool directory.

          coresize = <size>
                 Should the user's program abort and produce a core file, coresize specifies the maximum size in
                 bytes  of  the  core  file  which  the user wishes to keep. If coresize is not specified in the
                 command file, this is set to 0 (meaning no core will be generated).

          cron_day_of_month = <Cron-evaluated Day>
                 The set of days of the month for which a deferral time  applies.  The  HTCondor  User's  manual
                 section on Time Scheduling for Job Execution has further details.

          cron_day_of_week = <Cron-evaluated Day>
                 The  set  of  days  of  the  week for which a deferral time applies. The HTCondor User's manual
                 section on Time Scheduling for Job Execution has further details.

          cron_hour = <Cron-evaluated Hour>
                 The set of hours of the day for which a deferral  time  applies.  The  HTCondor  User's  manual
                 section on Time Scheduling for Job Execution has further details.

          cron_minute = <Cron-evaluated Minute>
                 The  set  of  minutes  within  an  hour for which a deferral time applies.  The HTCondor User's
                 manual section on Time Scheduling for Job Execution has further details.

          cron_month = <Cron-evaluated Month>
                 The set of months within a year for which a deferral time applies.  The HTCondor User's  manual
                 section on Time Scheduling for Job Execution has further details.

          cron_prep_time = <ClassAd Integer Expression>
                 Analogous to deferral_prep_time.  The number of seconds prior to a job's deferral time that the
                 job may be matched and sent to an execution machine.

          cron_window = <ClassAd Integer Expression>
                 Analogous  to the submit command deferral_window.  It allows cron jobs that miss their deferral
                 time to begin execution.

                 The HTCondor User's manual section on Time Scheduling for Job Execution has further details.

          dagman_log = <pathname>
                 DAGMan inserts this command to specify an event log that it watches to maintain  the  state  of
                 the DAG. If the log command is not specified in the submit file, DAGMan uses the log command to
                 specify the event log.

          deferral_prep_time = <ClassAd Integer Expression>
                 The number of seconds prior to a job's deferral time that the job may be matched and sent to an
                 execution machine.

                 The HTCondor User's manual section on Time Scheduling for Job Execution has further details.

          deferral_time = <ClassAd Integer Expression>
                 Allows  a  job  to  specify  the  time at which its execution is to begin, instead of beginning
                 execution as soon as it arrives at the execution machine. The deferral time  is  an  expression
                 that  evaluates  to  a  Unix  Epoch  timestamp (the number of seconds elapsed since 00:00:00 on
                 January 1, 1970, Coordinated Universal Time). Deferral time is evaluated with  respect  to  the
                 execution machine. This option delays the start of execution, but not the matching and claiming
                 of  a  machine  for  the  job.  If the job is not available and ready to begin execution at the
                 deferral time, it has missed its deferral time. A job that misses its deferral time will be put
                 on hold in the queue.

                 The HTCondor User's manual section on Time Scheduling for Job Execution has further details.

                 Due to implementation details, a deferral time may not be used for scheduler universe jobs.

          deferral_window = <ClassAd Integer Expression>
                 The deferral window is used in conjunction with the deferral_time command to  allow  jobs  that
                 miss their deferral time to begin execution.

                 The HTCondor User's manual section on Time Scheduling for Job Execution has further details.

          description = <string>
                 A string that sets the value of the job ClassAd attribute JobDescription. When set, tools which
                 display the executable such as condor_q will instead use this string.

          email_attributes = <list-of-job-ad-attributes>
                 A  comma-separated  list  of attributes from the job ClassAd. These attributes and their values
                 will be included in the e-mail notification of job completion.

          image_size = <size>
                 Advice to HTCondor specifying the maximum virtual image size to which the job will grow  during
                 its execution. HTCondor will then execute the job only on machines which have enough resources,
                 (such  as  virtual  memory),  to  support  executing  the  job. If not specified, HTCondor will
                 automatically make a (reasonably accurate) estimate  about  the  job's  size  and  adjust  this
                 estimate  as  the  program runs.  If specified and underestimated, the job may crash due to the
                 inability to acquire more address space; for example, if malloc() fails. If the image  size  is
                 overestimated, HTCondor may have difficulty finding machines which have the required resources.
                 size is specified in KiB. For example, for an image size of 8 MiB, size should be 8000.

          initialdir = <directory-path>
                 Used to give jobs a directory with respect to file input and output.  Also provides a directory
                 (on the machine from which the job is submitted) for the job event log, when a full path is not
                 specified.

                 For  vanilla  universe  jobs  where  there  is  a shared file system, it is the current working
                 directory on the machine where the job is executed.

                 For vanilla or grid universe jobs where file transfer mechanisms are utilized (there is  not  a
                 shared  file  system), it is the directory on the machine from which the job is submitted where
                 the input files come from, and where the job's output files go to.

                 For scheduler universe jobs, it is the directory on the machine from which the job is submitted
                 where the job runs; the current working directory for file input and  output  with  respect  to
                 relative path names.

                 Note that the path to the executable is not relative to initialdir if it is a relative path, it
                 is relative to the directory in which the condor_submit command is run.

          job_ad_information_attrs = <attribute-list>
                 A  comma-separated  list  of job ClassAd attribute names. The named attributes and their values
                 are written to the job event log  whenever  any  event  is  being  written  to  the  log.  This
                 implements  the  same  thing as the configuration variable EVENT_LOG_INFORMATION_ATTRS (see the
                 admin-manual/configuration-macros:daemon logging  configuration  file  entries  page),  but  it
                 applies to the job event log, instead of the system event log.

          job_lease_duration = <number-of-seconds>
                 For vanilla, parallel, VM, and java universe jobs only, the duration in seconds of a job lease.
                 The  default  value is 2,400, or forty minutes. If a job lease is not desired, the value can be
                 explicitly set to 0 to disable the job lease  semantics.  The  value  can  also  be  a  ClassAd
                 expression  that  evaluates  to  an  integer.  The  HTCondor  User's  manual section on Special
                 Environment Considerations has further details.

          job_machine_attrs = <attr1, attr2, ...>
                 A comma and/or space separated list of machine attribute names that should be recorded  in  the
                 job   ClassAd  in  addition  to  the  ones  specified  by  the  condor_schedd  daemon's  system
                 configuration variable SYSTEM_JOB_MACHINE_ATTRS. When there are multiple run attempts,  history
                 of  machine  attributes  from  previous run attempts may be kept. The number of run attempts to
                 store may be extended beyond the system-specified history  length  by  using  the  submit  file
                 command

          job_machine_attrs_history_length
                 A  machine  attribute  named  X  will  be  inserted  into the job ClassAd as an attribute named
                 MachineAttrX0. The previous value of this attribute will be named MachineAttrX1,  the  previous
                 to  that  will be named MachineAttrX2, and so on, up to the specified history length. A history
                 of length 1 means that only MachineAttrX0 will be recorded.  The  value  recorded  in  the  job
                 ClassAd  is  the evaluation of the machine attribute in the context of the job ClassAd when the
                 condor_schedd daemon initiates the start up of  the  job.  If  the  evaluation  results  in  an
                 Undefined  or  Error  result,  the  value  recorded  in  the job ad will be Undefined or Error,
                 respectively.

          want_graceful_removal = <boolean expression>
                 If true, this job will be given a chance to shut down cleanly when removed.  The  job  will  be
                 given as much time as the administrator of the execute resource allows, which may be none.  The
                 default is false.  For details, see the configuration setting GRACEFULLY_REMOVE_JOBS.

          kill_sig = <signal-number>
                 When  HTCondor  needs to kick a job off of a machine, it will send the job the signal specified
                 by signal-number, which needs to be an integer which represents a valid signal on the execution
                 machine.  The default value is SIGTERM, which is the standard way to  terminate  a  program  in
                 Unix.

          kill_sig_timeout = <seconds>
                 This   submit   command   should   no   longer   be   used.   job_max_vacate_time  instead.  If
                 job_max_vacate_time is not defined, this defines the number of  seconds  that  HTCondor  should
                 wait following the sending of the kill signal defined by kill_sig and forcibly killing the job.
                 The  actual  amount  of  time  between  sending  the signal and forcibly killing the job is the
                 smallest of this value and the configuration variable KILLING_TIMEOUT
                  , as defined on the execute machine.

          load_profile = <True | False>
                 When True, loads the account profile of the dedicated run account for Windows jobs. May not  be
                 used with run_as_owner.

          log_xml = <True | False>
                 If  log_xml  is  True,  then  the  job  event  log  file will be written in ClassAd XML. If not
                 specified, XML is not used. Note that the file is an XML  fragment;  it  is  missing  the  file
                 header and footer. Do not mix XML and non-XML within a single file. If multiple jobs write to a
                 single job event log file, ensure that all of the jobs specify this option in the same way.

          match_list_length = <integer value>
                 Defaults to the value zero (0). When match_list_length is defined with an integer value greater
                 than  zero  (0), attributes are inserted into the job ClassAd. The maximum number of attributes
                 defined is given by the integer value. The job ClassAds introduced are given as

                     LastMatchName0 = "most-recent-Name"
                     LastMatchName1 = "next-most-recent-Name"

                 The value for each introduced ClassAd is given by the value of  the  Name  attribute  from  the
                 machine ClassAd of a previous execution (match). As a job is matched, the definitions for these
                 attributes  will  roll,  with  LastMatchName1  becoming LastMatchName2, LastMatchName0 becoming
                 LastMatchName1, and LastMatchName0 being set by the most recent value of the Name attribute.

                 An intended use of these job attributes is in the requirements expression. The requirements can
                 allow a job to prefer a match with either the same or a  different  resource  than  a  previous
                 match.

          job_max_vacate_time = <integer expression>
                 An  integer-valued  expression  (in  seconds)  that  may be used to adjust the time given to an
                 evicted job for gracefully shutting down. If the job's setting is less than the machine's,  the
                 job's is used. If the job's setting is larger than the machine's, the result depends on whether
                 the  job  has  any  excess  retirement  time. If the job has more retirement time left than the
                 machine's max vacate time setting, then retirement time will be converted into  vacating  time,
                 up to the amount requested by the job.

                 Setting  this  expression does not affect the job's resource requirements or preferences. For a
                 job to only run on a machine with a minimum MachineMaxVacateTime, or to preferentially  run  on
                 such machines, explicitly specify this in the requirements and/or rank expressions.

          manifest = <True | False>
                 For  vanilla  and Docker -universe jobs (and others that use the shadow), specifies if HTCondor
                 (the starter) should produce a "manifest", which is directory containing three files: the  list
                 of  files and directories at the top level of the sandbox when file transfer in completes (in),
                 the same when file transfer out begins (out), and a dump of the environment set for the job  (‐
                 Environment).

                 This feature is not presently available for Windows.

          manifest_dir = <directory name>
                 For vanilla and Docker -universe jobs (and others that use the shadow), specifies the directory
                 in  which  to  record  the  manifest.   Specifying this enables the creation of a manifest.  By
                 default, the manifest directory is named <cluster>_<proc>_manifest, to avoid conflicts.

                 This feature is not presently available for Windows.

          max_job_retirement_time = <integer expression>
                 An integer-valued expression (in seconds) that does nothing unless the machine  that  runs  the
                 job  has been configured to provide retirement time. Retirement time is a grace period given to
                 a job to finish when a resource claim is about to be preempted. The default  behavior  in  many
                 cases  is  to  take  as much retirement time as the machine offers, so this command will rarely
                 appear in a submit description file.

                 When a resource claim is to be preempted, this expression in  the  submit  file  specifies  the
                 maximum run time of the job (in seconds, since the job started). This expression has no effect,
                 if  it  is  greater  than  the  maximum  retirement time provided by the machine policy. If the
                 resource claim is not  preempted,  this  expression  and  the  machine  retirement  policy  are
                 irrelevant.  If  the  resource  claim  is  preempted  the  job will be allowed to run until the
                 retirement time expires, at which point it is hard-killed. The job will be soft-killed when  it
                 is getting close to the end of retirement in order to give it time to gracefully shut down. The
                 amount  of lead-time for soft-killing is determined by the maximum vacating time granted to the
                 job.

                 Any jobs running with nice_user priority have a default max_job_retirement_time  of  0,  so  no
                 retirement  time  is  utilized by default. In all other cases, no default value is provided, so
                 the maximum amount of retirement time is utilized by default.

                 Setting this expression does not affect the job's resource requirements or preferences.  For  a
                 job  to  only run on a machine with a minimum MaxJobRetirementTime, or to preferentially run on
                 such machines, explicitly specify this in the requirements and/or rank expressions.

          nice_user = <True | False>
                 Normally, when a machine becomes available to HTCondor, HTCondor decides which job to run based
                 upon user and job priorities. Setting nice_user equal to True tells HTCondor not  to  use  your
                 regular  user  priority,  but  that  this job should have last priority among all users and all
                 jobs. So jobs submitted in this fashion run only on machines which no other  non-nice_user  job
                 wants  - a true bottom-feeder job! This is very handy if a user has some jobs they wish to run,
                 but do not wish to use resources that could instead be used  to  run  other  people's  HTCondor
                 jobs.  Jobs  submitted  in  this  fashion  have  an  accounting group.  The accounting group is
                 configurable by setting NICE_USER_ACCOUNTING_GROUP_NAME which defaults to nice-user The default
                 value is False.

          noop_job = <ClassAd Boolean Expression>
                 When this boolean expression is True, the job  is  immediately  removed  from  the  queue,  and
                 HTCondor  makes  no  attempt  at  running  the  job.  The  log file for the job will show a job
                 submitted event and a job terminated event, along with an exit  code  of  0,  unless  the  user
                 specifies a different signal or exit code.

          noop_job_exit_code = <return value>
                 When  noop_job is in the submit description file and evaluates to True, this command allows the
                 job to specify the return value as shown in the job's log file job  terminated  event.  If  not
                 specified,  the  job  will  show  as  having terminated with status 0. This overrides any value
                 specified with noop_job_exit_signal.

          noop_job_exit_signal = <signal number>
                 When noop_job is in the submit description file and evaluates to True, this command allows  the
                 job  to  specify the signal number that the job's log event will show the job having terminated
                 with.

          remote_initialdir = <directory-path>
                 The path specifies the directory in which the job is to be executed on the remote machine.

          rendezvousdir = <directory-path>
                 Used to specify the shared file system directory to be used for file system authentication when
                 submitting to a remote scheduler. Should be a path to a preexisting directory.

          run_as_owner = <True | False>
                 A boolean value that causes the job to be run under the login of the submitter, if supported by
                 the joint configuration of the submit and execute machines. On Unix platforms, this defaults to
                 True, and on Windows platforms, it defaults to False. May not be used with  load_profile.   See
                 the  HTCondor  manual  Platform-Specific  Information  chapter  for  administrative  details on
                 configuring Windows to support this option.

          stack_size = <size in bytes>
                 This command applies only to Linux platforms.  An integer number  of  bytes,  representing  the
                 amount  of  stack space to be allocated for the job. This value replaces the default allocation
                 of stack space, which is unlimited in size.

          starter_debug = <log levels>
                 This command causes the condor_starter to write a separate copy of its daemon log  in  the  job
                 scratch  directory.   If  the  value  is  True,  then  the  logging  level is the same that the
                 condor_starter is configured to use for its  normal  daemon  log.   Any  other  value  will  be
                 interpreted the same way as <SUBSYS>_DEBUG to set the logging level.

          starter_log = <pathname>
                 When  the condor_starter is writing a job-specific daemon log (see starter_debug), this command
                 causes the log file to be transferred to the Access Point along with the job's output  sandbox.
                 The  log  is written to the given pathname.  If starter_debug isn't set, then it will be set to
                 True.

          submit_event_notes = <note>
                 A string that is appended to the submit event in the job's log  file.   For  DAGMan  jobs,  the
                 string  DAG  Node: and the node's name is automatically defined for submit_event_notes, causing
                 the logged submit event to identify the DAG node job submitted.

          ulog_execute_attrs = <attribute-list>
                 A comma-separated list of machine ClassAd attribute  names.  The  named  attributes  and  their
                 values are written as part of the execution event in the job event log.

          use_oauth_services = <list of credential service names>
                 A  comma-separated  list  of  credential-providing  service  names  for which the job should be
                 provided credentials for the job execution environment. The credential service  providers  must
                 be configured by the pool admin.

          <credential_service_name>_oauth_permissions [_<handle>] = <scope>
                 A   string  containing  the  scope(s)  that  should  be  requested  for  the  credential  named
                 <credential_service_name>[_<handle>], where <handle> is optionally  provided  to  differentiate
                 between multiple credentials from the same credential service provider.

          <credential_service_name>_oauth_resource [_<handle>] = <resource>
                 A  string  containing  the resource (or "audience") that should be requested for the credential
                 named  <credential_service_name>[_<handle>],  where  <handle>   is   optionally   provided   to
                 differentiate between multiple credentials from the same credential service provider.

          MY.<attribute> = <value> or +<attribute> = <value>
                 A  macro  that  begins  with  MY.  or  a line that begins with a '+' (plus) character instructs
                 condor_submit to insert the given attribute (without + or MY.) into the job  ClassAd  with  the
                 given value. The macro can be referenced in other submit statements by using $(MY.<attribute>).
                 A +<attribute> is converted to MY.<attribute> when the file is read.

                 Note  that  setting  an  job  attribute  in  this way should not be used in place of one of the
                 specific commands listed above. Often, the command name does  not  directly  correspond  to  an
                 attribute  name;  furthermore,  many submit commands result in actions more complex than simply
                 setting an attribute or attributes. See Job ClassAd Attributes  for  a  list  of  HTCondor  job
                 attributes.

       MACROS AND COMMENTS

       In addition to commands, the submit description file can contain macros and comments.

          Macros Parameterless macros in the form of $(macro_name:default initial value) may be used anywhere in
                 HTCondor submit description files to provide textual substitution at submit time. Macros can be
                 defined by lines in the form of

                     <macro_name> = <string>

                 Several  pre-defined  macros are supplied by the submit description file parser. The $(Cluster)
                 or $(ClusterId) macro supplies the value of the

                  ClusterId job ClassAd attribute, and the $(Process) or $(ProcId) macro supplies the  value  of
                 the  ProcId  job  ClassAd  attribute.   The  $(JobId)  macro  supplies  the  full job id. It is
                 equivalent to $(ClusterId).$(ProcId).  These macros are intended to aid in the specification of
                 input/output files, arguments, etc., for clusters with lots of jobs, and/or could  be  used  to
                 supply an HTCondor process with its own cluster and process numbers on the command line.

                 The  $(Node)  macro  is  defined for parallel universe jobs, and is especially relevant for MPI
                 applications. It is a unique value assigned for  the  duration  of  the  job  that  essentially
                 identifies  the  machine (slot) on which a program is executing. Values assigned start at 0 and
                 increase monotonically. The values are assigned as the parallel job is about to start.

                 Recursive definition of macros is permitted. An example of a construction  that  works  is  the
                 following:

                     foo = bar
                     foo =  snap $(foo)

                 As a result, foo = snap bar.

                 Note that both left- and right- recursion works, so

                     foo = bar
                     foo =  $(foo) snap

                 has as its result foo = bar snap.

                 The construction

                     foo = $(foo) bar

                 by  itself  will  not  work,  as  it  does  not  have an initial base case.  Mutually recursive
                 constructions such as:

                     B = bar
                     C = $(B)
                     B = $(C) boo

                 will not work, and will fill memory with expansions.

                 A default value may be specified, for use if the macro has no definition. Consider the example

                     D = $(E:24)

                 Where E is not defined within the submit description  file,  the  default  value  24  is  used,
                 resulting in

                     D = 24

                 This  is  useful  for creating submit templates where values can be passed on the condor_submit
                 command line, but that have a default value as well.  In the above example, if you give a value
                 for E on the command line like this

                     condor_submit E=99 <submit-file>

                 The value of 99 is used for E, resulting in

                     D = 99

                 To use the dollar sign character ($) as a literal, without macro expansion, use

                     $(DOLLAR)

                 In addition to the normal macro, there is also a special kind of macro  called  a  substitution
                 macro
                  that  allows  the  substitution  of  a machine ClassAd attribute value defined on the resource
                 machine itself (gotten after a match to the machine  has  been  made)  into  specific  commands
                 within the submit description file. The substitution macro is of the form:

                     $$(attribute)

                 As  this  form  of  the  substitution macro is only evaluated within the context of the machine
                 ClassAd, use of a scope resolution prefix TARGET. or MY. is not allowed.

                 A common use of this form of the substitution macro is for the heterogeneous submission  of  an
                 executable:

                     executable = povray.$$(OpSys).$$(Arch)

                 Values  for the OpSys and Arch attributes are substituted at match time for any given resource.
                 This example allows HTCondor to automatically choose the correct  executable  for  the  matched
                 machine.

                 An  extension  to the syntax of the substitution macro provides an alternative string to use if
                 the machine attribute within the substitution macro is undefined. The syntax appears as:

                     $$(attribute:string_if_attribute_undefined)

                 An example using this extended syntax provides a path name to a required input file. Since  the
                 file  can be placed in different locations on different machines, the file's path name is given
                 as an argument to the program.

                     arguments = $$(input_file_path:/usr/foo)

                 On the machine, if the attribute input_file_path is not defined, then the path /usr/foo is used
                 instead.

                 As a special case that only works within  the  submit  file  environment  command,  the  string
                 $$(CondorScratchDir)  is  expanded  to the value of the job's scratch directory.  This does not
                 work for scheduler universe or grid universe jobs.

                 For example, to set PYTHONPATH to a subdirectory of the job scratch dir, one could set

                     environment = PYTHONPATH=$$(CondorScratchDir)/some/directory

                 A further extension to the syntax of the substitution macro allows the evaluation of a  ClassAd
                 expression to define the value. In this form, the expression may refer to machine attributes by
                 prefacing them with the TARGET. scope resolution prefix. To place a ClassAd expression into the
                 substitution macro, square brackets are added to delimit the expression. The syntax appears as:

                     $$([ClassAd expression])

                 An  example of a job that uses this syntax may be one that wants to know how much memory it can
                 use. The application cannot detect this itself, as it would potentially use all of  the  memory
                 on  a  multi-slot  machine.  So  the  job determines the memory per slot, reducing it by 10% to
                 account for miscellaneous overhead,  and  passes  this  as  a  command  line  argument  to  the
                 application. In the submit description file will be

                     arguments = --memory $$([TARGET.Memory * 0.9])

                 To insert two dollar sign characters ($$) as literals into a ClassAd string, use

                     $$(DOLLARDOLLAR)

                 The  environment  macro,  $ENV,  allows the evaluation of an environment variable to be used in
                 setting a submit description file command.  The syntax used is

                     $ENV(variable)

                 An example  submit  description  file  command  that  uses  this  functionality  evaluates  the
                 submitter's home directory in order to set the path and file name of a log file:

                     log = $ENV(HOME)/jobs/logfile

                 The environment variable is evaluated when the submit description file is processed.

                 The  $RANDOM_CHOICE  macro allows a random choice to be made from a given list of parameters at
                 submission time. For an expression, if some randomness needs to be  generated,  the  macro  may
                 appear as

                     $RANDOM_CHOICE(0,1,2,3,4,5,6)

                 When evaluated, one of the parameters values will be chosen.

          Comments
                 Blank  lines  and  lines  beginning with a pound sign ('#') character are ignored by the submit
                 description file parser.

SUBMIT VARIABLES

       While processing the queue command in a submit file or from the command line, condor_submit will set  the
       values  of several automatic submit variables so that they can be referred to by statements in the submit
       file. With the exception of Cluster and Process, if these variables are set by the submit file, they will
       not be modified during queue processing.

          ClusterId
                 Set to the integer value that the ClusterId attribute that the job ClassAd will have  when  the
                 job  is  submitted.  All  jobs  in  a  single  submit will normally have the same value for the
                 ClusterId. If the -dry-run argument is specified, The value will be 1.

          Cluster
                 Alternate name for the ClusterId submit variable. Before HTCondor version 8.4 this was the only
                 name.

          ProcId Set to the integer value that the ProcId attribute of the job ClassAd will have when the job is
                 submitted. The value will start at 0 and increment by 1 for each job submitted.

          Process
                 Alternate name for the ProcId submit variable. Before HTCondor version 8.4 this  was  the  only
                 name.

          JobId  Set to $(ClusterId).$(ProcId) so that it will expand to the full id of the job.

          Node   For  parallel universes, set to the value #pArAlLeLnOdE# or #MpInOdE# depending on the parallel
                 universe type For other universes it is set to nothing.

          Step   Set to the step value as it varies from 0 to N-1 where N is the number provided  on  the  queue
                 argument.  This variable changes at the same rate as ProcId when it changes at all.  For submit
                 files that don't make use of the queue number option, Step will always be 0. For  submit  files
                 that don't make use of any of the foreach options, Step and ProcId will always be the same.

          ItemIndex
                 Set to the index within the item list being processed by the various queue foreach options. For
                 submit  files  that  don't  make  use of any queue foreach list, ItemIndex will always be 0 For
                 submit files that make use of a slice to select only some items in a  foreach  list,  ItemIndex
                 will only be set to selected values.

          Row    Alternate name for ItemIndex.

          Item   when a queue foreach option is used and no variable list is supplied, this variable will be set
                 to the value of the current item.

       The automatic variables below are set before parsing the submit file, and will not vary during processing
       unless the submit file itself sets them.

          ARCH   Set  to  the CPU architecture of the machine running condor_submit.  The value will be the same
                 as the automatic configuration variable of the same name.

          OPSYS  Set to the name of the operating system on the machine running condor_submit. The value will be
                 the same as the automatic configuration variable of the same name.

          OPSYSANDVER
                 Set to the name and major version of the operating system on the machine running condor_submit.
                 The value will be the same as the automatic configuration variable of the same name.

          OPSYSMAJORVER
                 Set to the major version of the operating system on  the  machine  running  condor_submit.  The
                 value will be the same as the automatic configuration variable of the same name.

          OPSYSVER
                 Set to the version of the operating system on the machine running condor_submit. The value will
                 be the same as the automatic configuration variable of the same name.

          SPOOL  Set  to  the  full  path  of  the  HTCondor  spool directory. The value will be the same as the
                 automatic configuration variable of the same name.

          IsLinux
                 Set to true if the operating system of the machine running condor_submit is  a  Linux  variant.
                 Set to false otherwise.

          IsWindows
                 Set to true if the operating system of the machine running condor_submit is a Microsoft Windows
                 variant. Set to false otherwise.

          SUBMIT_FILE
                 Set  to  the  full  pathname  of  the  submit  file being processed by condor_submit. If submit
                 statements are read from standard input, it is set to nothing.

          SUBMIT_TIME
                 Set to the unix timestamp of the current time when the job is submitted.

          YEAR   Set to the 4 digit year when the job is submitted.

          MONTH  Set to the 2 digit month when the job is submitted.

          DAY    Set to the 2 digit day when the job is submitted.

EXIT STATUS

       condor_submit will exit with a status value of 0 (zero) upon success, and a non-zero value upon failure.

EXAMPLES

       • Submit Description File Example 1: This submit description file example queues 150 runs of program  foo
         which  must have been compiled and linked for an Intel x86 processor running RHEL 3.  HTCondor will not
         attempt to run the processes on machines which have less than 32 Megabytes of physical memory,  and  it
         will  run  them  on  machines  which have at least 64 Megabytes, if such machines are available. Stdin,
         stdout, and stderr will refer to in.0, out.0, and err.0 for the first run of this program (process  0).
         Stdin,  stdout, and stderr will refer to in.1, out.1, and err.1 for process 1, and so forth. A log file
         containing entries about where and when HTCondor runs, transfers file, if it's  evicted,  and  when  it
         terminates,  among  other  things,  the  various  processes  in  this cluster will be written into file
         foo.log.

            ####################
            #
            # Example 2: Show off some fancy features including
            # use of pre-defined macros and logging.
            #
            ####################

            Executable     = foo
            Universe       = vanilla
            Requirements   = OpSys == "LINUX" && Arch =="INTEL"
            Rank           = Memory >= 64
            Request_Memory = 32 Mb
            Image_Size     = 28 Mb

            Error   = err.$(Process)
            Input   = in.$(Process)
            Output  = out.$(Process)
            Log = foo.log
            Queue 150

       • Submit Description File Example 2: This example targets  the  /bin/sleep  program  to  run  only  on  a
         platform  running  a  RHEL  6  operating  system.  The example presumes that the pool contains machines
         running more than one version of Linux, and this job needs  the  particular  operating  system  to  run
         correctly.

            ####################
            #
            # Example 3: Run on a RedHat 6 machine
            #
            ####################
            Universe     = vanilla
            Executable   = /bin/sleep
            Arguments    = 30
            Requirements = (OpSysAndVer == "RedHat6")

            Error   = err.$(Process)
            Input   = in.$(Process)
            Output  = out.$(Process)
            Log     = sleep.log
            Queue

       • Command  Line  example:  The  following  command uses the -append option to add two commands before the
         job(s) is queued. A log file and an error log file  are  specified.  The  submit  description  file  is
         unchanged.

            $ condor_submit -a "log = out.log" -a "error = error.log" mysubmitfile

         Note that each of the added commands is contained within quote marks because there are space characters
         within the command.

       • periodic_remove  example:  A  job should be removed from the queue, if the total suspension time of the
         job is more than half of the run time of the job.

         Including the command

            periodic_remove = CumulativeSuspensionTime >
                              ((RemoteWallClockTime - CumulativeSuspensionTime) / 2.0)

         in the submit description file causes this to happen.

GENERAL REMARKS

       • For security reasons, HTCondor will refuse to run any jobs submitted by user root (UID =  0)  or  by  a
         user whose default group is group wheel (GID = 0). Jobs submitted by user root or a user with a default
         group of wheel will appear to sit forever in the queue in an idle state.

       • All path names specified in the submit description file must be less than 256 characters in length, and
         command  line  arguments  must be less than 4096 characters in length; otherwise, condor_submit gives a
         warning message but the jobs will not execute properly.

       • Somewhat understandably, behavior gets bizarre if the user makes the  mistake  of  requesting  multiple
         HTCondor  jobs  to write to the same file, and/or if the user alters any files that need to be accessed
         by an HTCondor job which is still in the queue. For example, the compressing of data  or  output  files
         before an HTCondor job has completed is a common mistake.

SEE ALSO

       HTCondor User Manual

AUTHOR

       HTCondor Team

COPYRIGHT

       1990-2024,  Center  for High Throughput Computing, Computer Sciences Department, University of Wisconsin-
       Madison, Madison, WI, US. Licensed under the Apache License, Version 2.0.

                                                  Jan 04, 2025                                  CONDOR_SUBMIT(1)