Hprc banner tamu.png

Difference between revisions of "Ada:Batch Queues"

From TAMU HPRC
Jump to: navigation, search
(Queues)
(Queues)
Line 19: Line 19:
 
medium    4 / 4 / 4000          5 hr / 24 hr            64 GB and 256 GB nodes
 
medium    4 / 4 / 4000          5 hr / 24 hr            64 GB and 256 GB nodes
 
long      4 / 4 / 2000          24 hr / 7 days          64 GB and 256 GB nodes              Maximum of 6000 cores for all running jobs in this queue
 
long      4 / 4 / 2000          24 hr / 7 days          64 GB and 256 GB nodes              Maximum of 6000 cores for all running jobs in this queue
xlarge    1 / 1/ 280             1 hr / 200 hr            1 TB nodes (11), 2 TB nodes (4)
+
xlarge    1 / 1 / 280           1 hr / 200 hr            1 TB nodes (11), 2 TB nodes (4)
 
vnc        1 / 1 / 20            1 hr / 6 hr              All 30 nodes with GPUs
 
vnc        1 / 1 / 20            1 hr / 6 hr              All 30 nodes with GPUs
 
</pre>
 
</pre>

Revision as of 16:48, 11 May 2015

Queues

LSF, upon job submission, sends your jobs to appropriate batch queues. These are (software) service stations configured to control the scheduling and dispatch of jobs that have arrived in them. Batch queues are characterised by all sorts of parameters. Some of the most important are: (1) the total number of jobs that can be concurrently running (number of run slots); (2) the wall-clock time limit per job; (3) the type and number of nodes it can distpatch jobs to; (4) which users or user groups can use that queue; etc. These settings control whether a job will lie idle in the queue or be dispatched quickly for execution.

The current (May 2015) queue structure. It is in flux. It should consolidate into something more permanent by, say, the end of March 2015.

Queue      Min/Default/Max Cpus   Default/Max Walltime     Compute Node Types                  Notes
special         None              1 hr / 36 hr             All
devel      1 / 1 / 320            10 min / 1 hr            All
small      1 / 1 / 3              1 hr / 120 hr            64 GB and 256 GB nodes
short      4 / 4 / 8000           1 hr / 5 hr              64 GB and 256 GB nodes
medium     4 / 4 / 4000           5 hr / 24 hr             64 GB and 256 GB nodes
long       4 / 4 / 2000           24 hr / 7 days           64 GB and 256 GB nodes              Maximum of 6000 cores for all running jobs in this queue
xlarge     1 / 1 / 280            1 hr / 200 hr            1 TB nodes (11), 2 TB nodes (4)
vnc        1 / 1 / 20             1 hr / 6 hr              All 30 nodes with GPUs

LSF determines which queue will receive a job for processing. The selection is determined mainly by the resources (e.g., number of cpus, wall-clock limit) specified, explicitly or by default. There are exceptions: (1) the the xlarge queue that is associated with nodes that have 1TB or 2TB of main memory; and, (2) the special queue which gives one access to all of the compute nodes. To use either one you must specify the -q xlarge option in your job file as well as have special permission for such access. Output from the bjobs command contains the name of the queue associated with a given job.

Fair-Share Policy

... pending ...

Public and Private/Group Queues

... pending ...


The Interactive Queue

... pending ...