Hprc banner tamu.png

Difference between revisions of "Terra:Batch Job Files"

From TAMU HPRC
Jump to: navigation, search
Line 7: Line 7:
 
* Add your commands and/or scripting
 
* Add your commands and/or scripting
 
* Submit the job to the batch system
 
* Submit the job to the batch system
 +
 +
Several of the most important options are described below. These options are typically all that is needed to run a job on Terra.
 +
 +
{| class="wikitable" style="text-align: center;"
 +
| System Name:
 +
| Terra
 +
|-
 +
| Host Name:
 +
| terra.tamu.edu
 +
|-
 +
| Operating System:
 +
| Linux (CentOS 7)
 +
|-
 +
| Total Compute Cores/Nodes:
 +
| 8,512 cores<br>304 nodes
 +
|-
 +
| Compute Nodes:
 +
| 256 compute nodes, each with 64GB RAM <br> 48 GPU nodes, each with a single Tesla K80 GPU and 128GB of RAM
 +
|-
 +
| Interconnect:
 +
| Intel OmniPath100 Series switches.
 +
|-
 +
| Peak Performance:
 +
| ~X TFLOPs (TBD)
 +
|-
 +
| Global Disk:
 +
| 1.5PB (raw) via IBM's GSS26 appliance for general use <br>1.5PB (raw) via IBM's GSS256 purchased by a dedicated for GeoPhysics
 +
|-
 +
| File System:
 +
| General Parallel File System (GPFS)
 +
|-
 +
| Batch Facility:
 +
| [http://slurm.schedmd.com/ Slurm by SchedMD]
 +
|-
 +
| Location:
 +
| Teague Data Center
 +
|-
 +
| Production Date:
 +
| Fall 2016 (tentative)
 +
|}
  
 
Several examples of Slurm job files for Terra are listed below. For translating Ada/LSF job files, the [[HPRC:Batch_Translation | Batch Job Translation Guide]] provides some reference.  
 
Several examples of Slurm job files for Terra are listed below. For translating Ada/LSF job files, the [[HPRC:Batch_Translation | Batch Job Translation Guide]] provides some reference.  
  
 
Documentation for advanced options can be found under [[Terra:Batch:Advanced Documentation | Advanced Documentation]].
 
Documentation for advanced options can be found under [[Terra:Batch:Advanced Documentation | Advanced Documentation]].

Revision as of 13:25, 25 October 2016

Job Files

While not the only method of submitted a job, job files fulfill the needs of most users.

The general idea behind job files follows:

  • Make resource requests
  • Add your commands and/or scripting
  • Submit the job to the batch system

Several of the most important options are described below. These options are typically all that is needed to run a job on Terra.

System Name: Terra
Host Name: terra.tamu.edu
Operating System: Linux (CentOS 7)
Total Compute Cores/Nodes: 8,512 cores
304 nodes
Compute Nodes: 256 compute nodes, each with 64GB RAM
48 GPU nodes, each with a single Tesla K80 GPU and 128GB of RAM
Interconnect: Intel OmniPath100 Series switches.
Peak Performance: ~X TFLOPs (TBD)
Global Disk: 1.5PB (raw) via IBM's GSS26 appliance for general use
1.5PB (raw) via IBM's GSS256 purchased by a dedicated for GeoPhysics
File System: General Parallel File System (GPFS)
Batch Facility: Slurm by SchedMD
Location: Teague Data Center
Production Date: Fall 2016 (tentative)

Several examples of Slurm job files for Terra are listed below. For translating Ada/LSF job files, the Batch Job Translation Guide provides some reference.

Documentation for advanced options can be found under Advanced Documentation.