Belle II Software development
LSF Class Reference
Inheritance diagram for LSF:
Batch Backend

Classes

class  LSFResult
 

Public Member Functions

 __init__ (self, *, backend_args=None)
 
 can_submit (self, njobs=1)
 
 bjobs (cls, output_fields=None, job_id="", username="", queue="")
 
 bqueues (cls, output_fields=None, queues=None)
 
 submit (self, job, check_can_submit=True, jobs_per_check=100)
 
 get_batch_submit_script_path (self, job)
 
 get_submit_script_path (self, job)
 

Public Attributes

int global_job_limit = self.default_global_job_limit
 The active job limit.
 
int sleep_between_submission_checks = self.default_sleep_between_submission_checks
 Seconds we wait before checking if we can submit a list of jobs.
 
dict backend_args = {**self.default_backend_args, **backend_args}
 The backend args that will be applied to jobs unless the job specifies them itself.
 

Static Public Attributes

str cmd_wkdir = "#BSUB -cwd"
 Working directory directive.
 
str cmd_stdout = "#BSUB -o"
 stdout file directive
 
str cmd_stderr = "#BSUB -e"
 stderr file directive
 
str cmd_queue = "#BSUB -q"
 Queue directive.
 
str cmd_name = "#BSUB -J"
 Job name directive.
 
list submission_cmds = []
 Shell command to submit a script, should be implemented in the derived class.
 
int default_global_job_limit = 1000
 Default global limit on the total number of submitted/running jobs that the user can have.
 
int default_sleep_between_submission_checks = 30
 Default time betweeon re-checking if the active jobs is below the global job limit.
 
str submit_script = "submit.sh"
 Default submission script name.
 
str exit_code_file = "__BACKEND_CMD_EXIT_STATUS__"
 Default exit code file name.
 
dict default_backend_args = {}
 Default backend_args.
 

Protected Member Functions

 _add_batch_directives (self, job, batch_file)
 
 _create_cmd (self, script_path)
 
 _submit_to_batch (cls, cmd)
 
 _create_parent_job_result (cls, parent)
 
 _create_job_result (cls, job, batch_output)
 
 _make_submit_file (self, job, submit_file_path)
 
 _ (self, job, check_can_submit=True, jobs_per_check=100)
 
 _ (self, job, check_can_submit=True, jobs_per_check=100)
 
 _ (self, jobs, check_can_submit=True, jobs_per_check=100)
 
 _add_wrapper_script_setup (self, job, batch_file)
 
 _add_wrapper_script_teardown (self, job, batch_file)
 

Static Protected Member Functions

 _add_setup (job, batch_file)
 

Detailed Description

Backend for submitting calibration processes to a qsub batch system.

Definition at line 1620 of file backends.py.

Constructor & Destructor Documentation

◆ __init__()

__init__ ( self,
* ,
backend_args = None )
 

Definition at line 1641 of file backends.py.

1641 def __init__(self, *, backend_args=None):
1642 """
1643 """
1644 super().__init__(backend_args=backend_args)
1645

Member Function Documentation

◆ _() [1/3]

_ ( self,
job,
check_can_submit = True,
jobs_per_check = 100 )
protectedinherited
Submit method of Batch backend for a `SubJob`. Should take `SubJob` object, create needed directories,
create batch script, and send it off with the batch submission command.
It should apply the correct options (default and user requested).

Should set a Result object as an attribute of the job.

Definition at line 1211 of file backends.py.

1211 def _(self, job, check_can_submit=True, jobs_per_check=100):
1212 """
1213 Submit method of Batch backend for a `SubJob`. Should take `SubJob` object, create needed directories,
1214 create batch script, and send it off with the batch submission command.
1215 It should apply the correct options (default and user requested).
1216
1217 Should set a Result object as an attribute of the job.
1218 """
1219 # Make sure the output directory of the job is created, commented out due to permission issues
1220 # job.output_dir.mkdir(parents=True, exist_ok=True)
1221 # Make sure the working directory of the job is created
1222 job.working_dir.mkdir(parents=True, exist_ok=True)
1223 job.copy_input_sandbox_files_to_working_dir()
1224 job.dump_input_data()
1225 # Make submission file if needed
1226 batch_submit_script_path = self.get_batch_submit_script_path(job)
1227 self._make_submit_file(job, batch_submit_script_path)
1228 # Get the bash file we will actually run, might be the same file
1229 script_path = self.get_submit_script_path(job)
1230 # Construct the batch submission script (with directives if that is supported)
1231 with open(script_path, mode="w") as batch_file:
1232 self._add_batch_directives(job, batch_file)
1233 self._add_wrapper_script_setup(job, batch_file)
1234 self._add_setup(job, batch_file)
1235 print(job.full_command, file=batch_file)
1236 self._add_wrapper_script_teardown(job, batch_file)
1237 os.chmod(script_path, 0o755)
1238 B2INFO(f"Submitting {job}")
1239 # Do the actual batch submission
1240 cmd = self._create_cmd(batch_submit_script_path)
1241 output = self._submit_to_batch(cmd)
1242 self._create_job_result(job, output)
1243 job.status = "submitted"
1244 B2INFO(f"{job} submitted")
1245

◆ _() [2/3]

_ ( self,
job,
check_can_submit = True,
jobs_per_check = 100 )
protectedinherited
Submit method of Batch backend. Should take job object, create needed directories, create batch script,
and send it off with the batch submission command, applying the correct options (default and user requested.)

Should set a Result object as an attribute of the job.

Definition at line 1247 of file backends.py.

1247 def _(self, job, check_can_submit=True, jobs_per_check=100):
1248 """
1249 Submit method of Batch backend. Should take job object, create needed directories, create batch script,
1250 and send it off with the batch submission command, applying the correct options (default and user requested.)
1251
1252 Should set a Result object as an attribute of the job.
1253 """
1254 # Make sure the output directory of the job is created, commented out due to permissions issue
1255 # job.output_dir.mkdir(parents=True, exist_ok=True)
1256 # Make sure the working directory of the job is created
1257 job.working_dir.mkdir(parents=True, exist_ok=True)
1258 # Check if we have any valid input files
1259 job.check_input_data_files()
1260 # Add any required backend args that are missing (I'm a bit hesitant to actually merge with job.backend_args)
1261 # just in case you want to resubmit the same job with different backend settings later.
1262 # job_backend_args = {**self.backend_args, **job.backend_args}
1263
1264 # If there's no splitter then we just submit the Job with no SubJobs
1265 if not job.splitter:
1266 # Get all of the requested files for the input sandbox and copy them to the working directory
1267 job.copy_input_sandbox_files_to_working_dir()
1268 job.dump_input_data()
1269 # Make submission file if needed
1270 batch_submit_script_path = self.get_batch_submit_script_path(job)
1271 self._make_submit_file(job, batch_submit_script_path)
1272 # Get the bash file we will actually run
1273 script_path = self.get_submit_script_path(job)
1274 # Construct the batch submission script (with directives if that is supported)
1275 with open(script_path, mode="w") as batch_file:
1276 self._add_batch_directives(job, batch_file)
1277 self._add_wrapper_script_setup(job, batch_file)
1278 self._add_setup(job, batch_file)
1279 print(job.full_command, file=batch_file)
1280 self._add_wrapper_script_teardown(job, batch_file)
1281 os.chmod(script_path, 0o755)
1282 B2INFO(f"Submitting {job}")
1283 # Do the actual batch submission
1284 cmd = self._create_cmd(batch_submit_script_path)
1285 output = self._submit_to_batch(cmd)
1286 self._create_job_result(job, output)
1287 job.status = "submitted"
1288 B2INFO(f"{job} submitted")
1289 else:
1290 # Create subjobs according to the splitter's logic
1291 job.splitter.create_subjobs(job)
1292 # Submit the subjobs
1293 self.submit(list(job.subjobs.values()))
1294 # After submitting subjobs, make a Job.result for the parent Job object, used to call ready() on
1295 self._create_parent_job_result(job)
1296

◆ _() [3/3]

_ ( self,
jobs,
check_can_submit = True,
jobs_per_check = 100 )
protectedinherited
Submit method of Batch Backend that takes a list of jobs instead of just one and submits each one.

Definition at line 1298 of file backends.py.

1298 def _(self, jobs, check_can_submit=True, jobs_per_check=100):
1299 """
1300 Submit method of Batch Backend that takes a list of jobs instead of just one and submits each one.
1301 """
1302 B2INFO(f"Submitting a list of {len(jobs)} jobs to a Batch backend")
1303 # Technically this could be a list of Jobs or SubJobs. And if it is a list of Jobs then it might not
1304 # be necessary to check if we can submit right now. We could do it later during the submission of the
1305 # SubJob list. However in the interest of simpler code we just do the check here, and re-check again
1306 # if a SubJob list comes through this function. Slightly inefficient, but much simpler logic.
1307
1308 # The first thing to do is make sure that we are iterating through the jobs list in chunks that are
1309 # equal to or smaller than the global limit. Otherwise nothing will ever submit.
1310
1311 if jobs_per_check > self.global_job_limit:
1312 B2INFO(f"jobs_per_check (={jobs_per_check}) but this is higher than the global job "
1313 f"limit for this backend (={self.global_job_limit}). Will instead use the "
1314 " value of the global job limit.")
1315 jobs_per_check = self.global_job_limit
1316
1317 # We group the jobs list into chunks of length jobs_per_check
1318 for jobs_to_submit in grouper(jobs_per_check, jobs):
1319 # Wait until we are allowed to submit
1320 while not self.can_submit(njobs=len(jobs_to_submit)):
1321 B2INFO("Too many jobs are currently in the batch system globally. Waiting until submission can continue...")
1322 time.sleep(self.sleep_between_submission_checks)
1323 else:
1324 # We loop here since we have already checked if the number of jobs is low enough, we don't want to hit this
1325 # function again unless one of the jobs has subjobs.
1326 B2INFO(f"Submitting the next {len(jobs_to_submit)} jobs...")
1327 for job in jobs_to_submit:
1328 self.submit(job, check_can_submit, jobs_per_check)
1329 B2INFO(f"All {len(jobs)} requested jobs submitted")
1330

◆ _add_batch_directives()

_add_batch_directives ( self,
job,
batch_file )
protected
Adds LSF BSUB directives for the job to a script.

Reimplemented from Batch.

Definition at line 1646 of file backends.py.

1646 def _add_batch_directives(self, job, batch_file):
1647 """
1648 Adds LSF BSUB directives for the job to a script.
1649 """
1650 job_backend_args = {**self.backend_args, **job.backend_args} # Merge the two dictionaries, with the job having priority
1651 batch_queue = job_backend_args["queue"]
1652 print("#!/bin/bash", file=batch_file)
1653 print("# --- Start LSF ---", file=batch_file)
1654 print(" ".join([LSF.cmd_queue, batch_queue]), file=batch_file)
1655 print(" ".join([LSF.cmd_name, job.name]), file=batch_file)
1656 print(" ".join([LSF.cmd_wkdir, str(job.working_dir)]), file=batch_file)
1657 print(" ".join([LSF.cmd_stdout, Path(job.working_dir, _STDOUT_FILE).as_posix()]), file=batch_file)
1658 print(" ".join([LSF.cmd_stderr, Path(job.working_dir, _STDERR_FILE).as_posix()]), file=batch_file)
1659 print("# --- End LSF ---", file=batch_file)
1660

◆ _add_setup()

_add_setup ( job,
batch_file )
staticprotectedinherited
Adds setup lines to the shell script file.

Definition at line 806 of file backends.py.

806 def _add_setup(job, batch_file):
807 """
808 Adds setup lines to the shell script file.
809 """
810 for line in job.setup_cmds:
811 print(line, file=batch_file)
812

◆ _add_wrapper_script_setup()

_add_wrapper_script_setup ( self,
job,
batch_file )
protectedinherited
Adds lines to the submitted script that help with job monitoring/setup. Mostly here so that we can insert
`trap` statements for Ctrl-C situations.

Definition at line 813 of file backends.py.

813 def _add_wrapper_script_setup(self, job, batch_file):
814 """
815 Adds lines to the submitted script that help with job monitoring/setup. Mostly here so that we can insert
816 `trap` statements for Ctrl-C situations.
817 """
818 start_wrapper = f"""# ---
819# trap ctrl-c and call ctrl_c()
820trap '(ctrl_c 130)' SIGINT
821trap '(ctrl_c 143)' SIGTERM
822
823function write_exit_code() {{
824 echo "Writing $1 to exit status file"
825 echo "$1" > {self.exit_code_file}
826 exit $1
827}}
828
829function ctrl_c() {{
830 trap '' SIGINT SIGTERM
831 echo "** Trapped Ctrl-C **"
832 echo "$1" > {self.exit_code_file}
833 exit $1
834}}
835# ---"""
836 print(start_wrapper, file=batch_file)
837

◆ _add_wrapper_script_teardown()

_add_wrapper_script_teardown ( self,
job,
batch_file )
protectedinherited
Adds lines to the submitted script that help with job monitoring/teardown. Mostly here so that we can insert
an exit code of the job cmd being written out to a file. Which means that we can know if the command was
successful or not even if the backend server/monitoring database purges the data about our job i.e. If PBS
removes job information too quickly we may never know if a job succeeded or failed without some kind of exit
file.

Definition at line 838 of file backends.py.

838 def _add_wrapper_script_teardown(self, job, batch_file):
839 """
840 Adds lines to the submitted script that help with job monitoring/teardown. Mostly here so that we can insert
841 an exit code of the job cmd being written out to a file. Which means that we can know if the command was
842 successful or not even if the backend server/monitoring database purges the data about our job i.e. If PBS
843 removes job information too quickly we may never know if a job succeeded or failed without some kind of exit
844 file.
845 """
846 end_wrapper = """# ---
847write_exit_code $?"""
848 print(end_wrapper, file=batch_file)
849

◆ _create_cmd()

_create_cmd ( self,
script_path )
protected
 

Reimplemented from Batch.

Definition at line 1661 of file backends.py.

1661 def _create_cmd(self, script_path):
1662 """
1663 """
1664 submission_cmd = self.submission_cmds[:]
1665 submission_cmd.append(script_path.as_posix())
1666 submission_cmd = " ".join(submission_cmd)
1667 return [submission_cmd]
1668

◆ _create_job_result()

_create_job_result ( cls,
job,
batch_output )
protected
 

Reimplemented from Batch.

Definition at line 1777 of file backends.py.

1777 def _create_job_result(cls, job, batch_output):
1778 """
1779 """
1780 m = re.search(r"Job <(\d+)>", str(batch_output))
1781 if m:
1782 job_id = m.group(1)
1783 else:
1784 raise BackendError(f"Failed to get the batch job ID of {job}. LSF output was:\n{batch_output}")
1785
1786 B2INFO(f"Job ID of {job} recorded as: {job_id}")
1787 job.result = cls.LSFResult(job, job_id)
1788

◆ _create_parent_job_result()

_create_parent_job_result ( cls,
parent )
protected
We want to be able to call `ready()` on the top level `Job.result`. So this method needs to exist
so that a Job.result object actually exists. It will be mostly empty and simply updates subjob
statuses and allows the use of ready().

Reimplemented from Backend.

Definition at line 1773 of file backends.py.

1773 def _create_parent_job_result(cls, parent):
1774 parent.result = cls.LSFResult(parent, None)
1775

◆ _make_submit_file()

_make_submit_file ( self,
job,
submit_file_path )
protectedinherited
Useful for the HTCondor backend where a submit is needed instead of batch
directives pasted directly into the submission script. It should be overwritten
if needed.

Reimplemented in HTCondor.

Definition at line 1179 of file backends.py.

1179 def _make_submit_file(self, job, submit_file_path):
1180 """
1181 Useful for the HTCondor backend where a submit is needed instead of batch
1182 directives pasted directly into the submission script. It should be overwritten
1183 if needed.
1184 """
1185

◆ _submit_to_batch()

_submit_to_batch ( cls,
cmd )
protected
Do the actual batch submission command and collect the output to find out the job id for later monitoring.

Reimplemented from Batch.

Definition at line 1670 of file backends.py.

1670 def _submit_to_batch(cls, cmd):
1671 """
1672 Do the actual batch submission command and collect the output to find out the job id for later monitoring.
1673 """
1674 sub_out = subprocess.check_output(cmd, stderr=subprocess.STDOUT, universal_newlines=True, shell=True)
1675 return sub_out
1676

◆ bjobs()

bjobs ( cls,
output_fields = None,
job_id = "",
username = "",
queue = "" )
Simplistic interface to the `bjobs` command. lets you request information about all jobs matching the filters
'job_id', 'username', and 'queue'. The result is the JSON dictionary returned by output of the ``-json`` bjobs option.

Parameters:
    output_fields (list[str]): A list of bjobs -o fields that you would like information about e.g. ['stat', 'name', 'id']
    job_id (str): String representation of the Job ID given by bsub during submission If this argument is given then
        the output of this function will be only information about this job. If this argument is not given, then all jobs
        matching the other filters will be returned.
    username (str): By default bjobs (and this function) return information about only the current user's jobs. By giving
        a username you can access the job information of a specific user's jobs. By giving ``username='all'`` you will
        receive job information from all known user jobs matching the other filters.
    queue (str): Set this argument to receive job information about jobs that are in the given queue and no other.

Returns:
    dict: JSON dictionary of the form:

    .. code-block:: python

      {
        "NJOBS":<njobs returned by command>,
        "JOBS":[
                {
                  <output field: value>, ...
                }, ...
               ]
      }

Definition at line 1813 of file backends.py.

1813 def bjobs(cls, output_fields=None, job_id="", username="", queue=""):
1814 """
1815 Simplistic interface to the `bjobs` command. lets you request information about all jobs matching the filters
1816 'job_id', 'username', and 'queue'. The result is the JSON dictionary returned by output of the ``-json`` bjobs option.
1817
1818 Parameters:
1819 output_fields (list[str]): A list of bjobs -o fields that you would like information about e.g. ['stat', 'name', 'id']
1820 job_id (str): String representation of the Job ID given by bsub during submission If this argument is given then
1821 the output of this function will be only information about this job. If this argument is not given, then all jobs
1822 matching the other filters will be returned.
1823 username (str): By default bjobs (and this function) return information about only the current user's jobs. By giving
1824 a username you can access the job information of a specific user's jobs. By giving ``username='all'`` you will
1825 receive job information from all known user jobs matching the other filters.
1826 queue (str): Set this argument to receive job information about jobs that are in the given queue and no other.
1827
1828 Returns:
1829 dict: JSON dictionary of the form:
1830
1831 .. code-block:: python
1832
1833 {
1834 "NJOBS":<njobs returned by command>,
1835 "JOBS":[
1836 {
1837 <output field: value>, ...
1838 }, ...
1839 ]
1840 }
1841 """
1842 B2DEBUG(29, f"Calling LSF.bjobs(output_fields={output_fields}, job_id={job_id}, username={username}, queue={queue})")
1843 # We must always return at least one output field when using JSON and -o options. So we choose the job id
1844 if not output_fields:
1845 output_fields = ["id"]
1846 # Output fields should be space separated but in a string.
1847 field_list_cmd = "\""
1848 field_list_cmd += " ".join(output_fields)
1849 field_list_cmd += "\""
1850 cmd_list = ["bjobs", "-o", field_list_cmd]
1851 # If the queue name is set then we add to the command options
1852 if queue:
1853 cmd_list.extend(["-q", queue])
1854 # If the username is set then we add to the command options
1855 if username:
1856 cmd_list.extend(["-u", username])
1857 # Can now add the json option before the final positional argument (if used)
1858 cmd_list.append("-json")
1859 # If the job id is set then we add to the end of the command
1860 if job_id:
1861 cmd_list.append(job_id)
1862 # We get a JSON serialisable summary from bjobs. Requires the shell argument.
1863 cmd = " ".join(cmd_list)
1864 B2DEBUG(29, f"Calling subprocess with command = '{cmd}'")
1865 output = decode_json_string(subprocess.check_output(cmd, stderr=subprocess.STDOUT, universal_newlines=True, shell=True))
1866 output["NJOBS"] = output["JOBS"]
1867 output["JOBS"] = output["RECORDS"]
1868 del output["RECORDS"]
1869 del output["COMMAND"]
1870 return output
1871

◆ bqueues()

bqueues ( cls,
output_fields = None,
queues = None )
Simplistic interface to the `bqueues` command. lets you request information about all queues matching the filters.
The result is the JSON dictionary returned by output of the ``-json`` bqueues option.

Parameters:
    output_fields (list[str]): A list of bqueues -o fields that you would like information about
        e.g. the default is ['queue_name' 'status' 'max' 'njobs' 'pend' 'run']
    queues (list[str]): Set this argument to receive information about only the queues that are requested and no others.
        By default you will receive information about all queues.

Returns:
    dict: JSON dictionary of the form:

    .. code-block:: python

      {
        "COMMAND":"bqueues",
        "QUEUES":46,
        "RECORDS":[
          {
            "QUEUE_NAME":"b2_beast",
            "STATUS":"Open:Active",
            "MAX":"200",
            "NJOBS":"0",
            "PEND":"0",
            "RUN":"0"
          }, ...
      }

Definition at line 1873 of file backends.py.

1873 def bqueues(cls, output_fields=None, queues=None):
1874 """
1875 Simplistic interface to the `bqueues` command. lets you request information about all queues matching the filters.
1876 The result is the JSON dictionary returned by output of the ``-json`` bqueues option.
1877
1878 Parameters:
1879 output_fields (list[str]): A list of bqueues -o fields that you would like information about
1880 e.g. the default is ['queue_name' 'status' 'max' 'njobs' 'pend' 'run']
1881 queues (list[str]): Set this argument to receive information about only the queues that are requested and no others.
1882 By default you will receive information about all queues.
1883
1884 Returns:
1885 dict: JSON dictionary of the form:
1886
1887 .. code-block:: python
1888
1889 {
1890 "COMMAND":"bqueues",
1891 "QUEUES":46,
1892 "RECORDS":[
1893 {
1894 "QUEUE_NAME":"b2_beast",
1895 "STATUS":"Open:Active",
1896 "MAX":"200",
1897 "NJOBS":"0",
1898 "PEND":"0",
1899 "RUN":"0"
1900 }, ...
1901 }
1902 """
1903 B2DEBUG(29, f"Calling LSF.bqueues(output_fields={output_fields}, queues={queues})")
1904 # We must always return at least one output field when using JSON and -o options. So we choose the job id
1905 if not output_fields:
1906 output_fields = ["queue_name", "status", "max", "njobs", "pend", "run"]
1907 # Output fields should be space separated but in a string.
1908 field_list_cmd = "\""
1909 field_list_cmd += " ".join(output_fields)
1910 field_list_cmd += "\""
1911 cmd_list = ["bqueues", "-o", field_list_cmd]
1912 # Can now add the json option before the final positional argument (if used)
1913 cmd_list.append("-json")
1914 # If the queue name is set then we add to the end of the command
1915 if queues:
1916 cmd_list.extend(queues)
1917 # We get a JSON serialisable summary from bjobs. Requires the shell argument.
1918 cmd = " ".join(cmd_list)
1919 B2DEBUG(29, f"Calling subprocess with command = '{cmd}'")
1920 output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, universal_newlines=True, shell=True)
1921 return decode_json_string(output)
1922
1923

◆ can_submit()

can_submit ( self,
njobs = 1 )
Checks the global number of jobs in LSF right now (submitted or running) for this user.
Returns True if the number is lower that the limit, False if it is higher.

Parameters:
    njobs (int): The number of jobs that we want to submit before checking again. Lets us check if we
        are sufficiently below the limit in order to (somewhat) safely submit. It is slightly dangerous to
        assume that it is safe to submit too many jobs since there might be other processes also submitting jobs.
        So njobs really shouldn't be abused when you might be getting close to the limit i.e. keep it <=250
        and check again before submitting more.

Reimplemented from Batch.

Definition at line 1789 of file backends.py.

1789 def can_submit(self, njobs=1):
1790 """
1791 Checks the global number of jobs in LSF right now (submitted or running) for this user.
1792 Returns True if the number is lower that the limit, False if it is higher.
1793
1794 Parameters:
1795 njobs (int): The number of jobs that we want to submit before checking again. Lets us check if we
1796 are sufficiently below the limit in order to (somewhat) safely submit. It is slightly dangerous to
1797 assume that it is safe to submit too many jobs since there might be other processes also submitting jobs.
1798 So njobs really shouldn't be abused when you might be getting close to the limit i.e. keep it <=250
1799 and check again before submitting more.
1800 """
1801 B2DEBUG(29, "Calling LSF().can_submit()")
1802 job_info = self.bjobs(output_fields=["stat"])
1803 total_jobs = job_info["NJOBS"]
1804 B2INFO(f"Total jobs active in the LSF system is currently {total_jobs}")
1805 if (total_jobs + njobs) > self.global_job_limit:
1806 B2INFO(f"Since the global limit is {self.global_job_limit} we cannot submit {njobs} jobs until some complete.")
1807 return False
1808 else:
1809 B2INFO("There is enough space to submit more jobs.")
1810 return True
1811

◆ get_batch_submit_script_path()

get_batch_submit_script_path ( self,
job )
inherited
Construct the Path object of the script file that we will submit using the batch command.
For most batch backends this is the same script as the bash script we submit.
But for some they require a separate submission file that describes the job.
To implement that you can implement this function in the Backend class.

Reimplemented in HTCondor.

Definition at line 1331 of file backends.py.

1331 def get_batch_submit_script_path(self, job):
1332 """
1333 Construct the Path object of the script file that we will submit using the batch command.
1334 For most batch backends this is the same script as the bash script we submit.
1335 But for some they require a separate submission file that describes the job.
1336 To implement that you can implement this function in the Backend class.
1337 """
1338 return Path(job.working_dir, self.submit_script)
1339

◆ get_submit_script_path()

get_submit_script_path ( self,
job )
inherited
Construct the Path object of the bash script file that we will submit. It will contain
the actual job command, wrapper commands, setup commands, and any batch directives

Definition at line 859 of file backends.py.

859 def get_submit_script_path(self, job):
860 """
861 Construct the Path object of the bash script file that we will submit. It will contain
862 the actual job command, wrapper commands, setup commands, and any batch directives
863 """
864 return Path(job.working_dir, self.submit_script)
865
866

◆ submit()

submit ( self,
job,
check_can_submit = True,
jobs_per_check = 100 )
inherited
 

Reimplemented from Backend.

Definition at line 1204 of file backends.py.

1204 def submit(self, job, check_can_submit=True, jobs_per_check=100):
1205 """
1206 """
1207 raise NotImplementedError("This is an abstract submit(job) method that shouldn't have been called. "
1208 "Did you submit a (Sub)Job?")
1209

Member Data Documentation

◆ backend_args

dict backend_args = {**self.default_backend_args, **backend_args}
inherited

The backend args that will be applied to jobs unless the job specifies them itself.

Definition at line 796 of file backends.py.

◆ cmd_name

str cmd_name = "#BSUB -J"
static

Job name directive.

Definition at line 1633 of file backends.py.

◆ cmd_queue

str cmd_queue = "#BSUB -q"
static

Queue directive.

Definition at line 1631 of file backends.py.

◆ cmd_stderr

str cmd_stderr = "#BSUB -e"
static

stderr file directive

Definition at line 1629 of file backends.py.

◆ cmd_stdout

str cmd_stdout = "#BSUB -o"
static

stdout file directive

Definition at line 1627 of file backends.py.

◆ cmd_wkdir

str cmd_wkdir = "#BSUB -cwd"
static

Working directory directive.

Definition at line 1625 of file backends.py.

◆ default_backend_args

dict default_backend_args = {}
staticinherited

Default backend_args.

Definition at line 788 of file backends.py.

◆ default_global_job_limit

int default_global_job_limit = 1000
staticinherited

Default global limit on the total number of submitted/running jobs that the user can have.

This limit will not affect the total number of jobs that are eventually submitted. But the jobs won't actually be submitted until this limit can be respected i.e. until the number of total jobs in the Batch system goes down. Since we actually submit in chunks of N jobs, before checking this limit value again, this value needs to be a little lower than the real batch system limit. Otherwise you could accidentally go over during the N job submission if other processes are checking and submitting concurrently. This is quite common for the first submission of jobs from parallel calibrations.

Note that if there are other jobs already submitted for your account, then these will count towards this limit.

Definition at line 1155 of file backends.py.

◆ default_sleep_between_submission_checks

int default_sleep_between_submission_checks = 30
staticinherited

Default time betweeon re-checking if the active jobs is below the global job limit.

Definition at line 1157 of file backends.py.

◆ exit_code_file

str exit_code_file = "__BACKEND_CMD_EXIT_STATUS__"
staticinherited

Default exit code file name.

Definition at line 786 of file backends.py.

◆ global_job_limit

int global_job_limit = self.default_global_job_limit
inherited

The active job limit.

This is 'global' because we want to prevent us accidentally submitting too many jobs from all current and previous submission scripts.

Definition at line 1166 of file backends.py.

◆ sleep_between_submission_checks

sleep_between_submission_checks = self.default_sleep_between_submission_checks
inherited

Seconds we wait before checking if we can submit a list of jobs.

Only relevant once we hit the global limit of active jobs, which is a lot usually.

Definition at line 1169 of file backends.py.

◆ submission_cmds

list submission_cmds = []
staticinherited

Shell command to submit a script, should be implemented in the derived class.

Definition at line 1142 of file backends.py.

◆ submit_script

submit_script = "submit.sh"
staticinherited

Default submission script name.

Definition at line 784 of file backends.py.


The documentation for this class was generated from the following file: