Ada Quick Start Guide
- 1 Ada Quick Start Guide
- 1.1 Ada Usage Policies
- 1.2 Accessing Ada
- 1.3 Navigating Ada & Storage Quotas
- 1.4 The Batch System
- 1.5 Finding Software
- 1.6 Running A Program / Preparing a Job File
- 1.7 Submitting and Monitoring Jobs
- 1.8 tamubatch
- 1.9 Additional Topics
Ada Usage Policies
Access to Ada is granted with the condition that you will understand and adhere to all TAMU HPRC and Ada-specific policies.
General policies can be found on the HPRC Policies page.
Ada-specific policies can be found on the Ada Policies page.
Most access to Ada is done via a secure shell session. In addition, two-factor authentication is required to login to any cluster.
Users on Windows computers use either PuTTY or MobaXterm. If MobaXterm works on your computer, it is usually easier to use. When starting an ssh session in PuTTY, choose the connection type 'SSH', select port 22, and then type the hostname 'ada.tamu.edu'. For MobaXterm, select 'Session', 'SSH', and then remote host 'ada.tamu.edu'. Check the box to specify username and type your NetID. After selecting 'Ok', you will be prompted for Duo Two Factor Authentication. For more detailed instructions, visit the Two Factor Authentication page.
Users on Mac and Linux/Unix should use whatever SSH-capable terminal is available on their system. The command to connect to Ada is as follows. Be sure to replace [NetID] with your TAMU NetID.
[user1@localhost ~]$ ssh [NetID]@ada.tamu.edu
Note: In this example [user1@localhost ~]$ represents the command prompt on your local machine.
Your login password is the same that used on Howdy. You will not see your password as your type it into the login prompt.
Off Campus Access
Please visit this page to find information on accessing Ada remotely.
For more detailed instructions on how to access our systems, please see the HPRC Access page.
When users first access Ada, they will be within their home directory. This directory has smaller storage quotas and should not be used for general purpose.
Users can navigate to their home directory with the following command:
[NetID@ada1 ~]$ cd /home/NetID
A User's scratch directory has more storage space than their home directory and is recommended for general purpose use. Users can navigate to their scratch directory with the following command:
[NetID@ada1 ~]$ cd /scratch/user/NetID
Users can navigate to scratch or home easily by using their respective environment variables.
Navigate to scratch with the following command:
[NetID@ada1 ~]$ cd $SCRATCH
Navigate to home with the following command:
[NetID@ada1 ~]$ cd $HOME
A user's scratch directory is restricted to 1TB/250,000 files of storage. This storage quota is expandable upon request. A user's scratch directory is NOT backed up.
A user's home directory is restricted to 10GB/10,000 files of storage. This storage quota is not expandable. A user's home directory is backed up on a nightly basis.
Users can see the current status of their storage quotas with:
[NetID@ada1 ~]$ showquota
If you need a storage quota increase, please contact us with justification and the expected length of time that you will need the quota increase.
The Batch System
The batch system is a load distribution implementation that ensures convenient and fair use of a shared resource. Submitting jobs to a batch system allows a user to reserve specific resources with minimal interference to other users. All users are required to submit resource-intensive processing to the compute nodes through the batch system - attempting to circumvent the batch system is not allowed.
On Ada, LSF is the batch system that provides job management. More information on LSF can be found in the Ada Batch page.
Software on Ada is loaded using modules.
A list of the most popular software on our systems is available on the HPRC Available Software page.
To find most available software on Ada, use the following command:
[NetID@ada1 ~]$ module avail
To search for particular software by keyword, use:
[NetID@ada1 ~]$ module spider keyword
To load a module, use:
[NetID@ada1 ~]$ module load moduleName
To list all currently loaded modules, use:
[NetID@ada1 ~]$ module list
To remove all currently loaded modules, use:
[NetID@ada1 ~]$ module purge
If you need new software or an update, please contact us with your request.
There are restrictions on what software we can install. There is also regularly a queue of requested software installations.
Please account for delays in your installation request timeline.
Running A Program / Preparing a Job File
In order to properly run a program on Ada, users will need to create a job file and submit a job.
The simple example job file below requests 1 core on 1 node with 2.5GB of RAM for 1.5 hour. Note that most nodes (>800) on Ada have 20 cores with 54GB of usable memory. Some nodes (~26) have 20 cores with 245GB of usable memory. A few special nodes have more cores and memory than this. Users are encouraged to ensure that their job requirements will fit within these restrictions.
Any modules that need to be loaded or executable commands will replace the "#First Executable Line" in this example.
##NECESSARY JOB SPECIFICATIONS #BSUB -J JobExample1 #Set the job name to "JobExample1" #BSUB -L /bin/bash #Uses the bash login shell to initialize the job's execution environment. #BSUB -W 1:30 #Set the wall clock limit to 1hr and 30min #BSUB -n 1 #Request 1 task #BSUB -R "span[ptile=1]" #Request 1 task/core per node #BSUB -R "rusage[mem=2560]" #Request 2560MB (2.5GB) per node #BSUB -M 2560 #Set the per process enforceable memory limit to 2560MB #BSUB -o Example1Out.%J #Send stdout/err to "Example1Out.[jobID]" #First Executable Line
Note: If your job file has been written on an older Mac or DOS workstation, you will need to use "dos2unix" to remove certain characters that interfere with parsing the script.
[NetID@ada1 ~]$ dos2unix MyJob.lsf
Submitting and Monitoring Jobs
Once a job file is ready, it is time to submit the job. Users can submit their jobs to LSF with the following command:
[NetID@ada1 ~]$ bsub < MyJob.lsf Verifying job submission parameters... Verifying project account... Account to charge: 000000000000 Balance (SUs): 5000.0000 SUs to charge: 1.5000 Job <0000000> is submitted to default queue <sn_short>.
After the job has been submitted, users are able to monitor their own jobs with several methods. To see the status of all of your jobs, use the following command:
[NetID@ada1 ~]$ bjobs JOBID STAT USER QUEUE JOB_NAME NEXEC_HOST SLOTS RUN_TIME TIME_LEFT 0000000 RUN NetID sn_short JobExample1 1 1 0 second(s) 1:30 L 0000001 PEND NetID sn_short JobExample1 1 1 0 second(s) 1:30 L
To see the status of one job, use the following command, where XXXXXXX is the JobID:
[NetID@ada1 ~]$ bjobs XXXXXXX JOBID STAT USER QUEUE JOB_NAME NEXEC_HOST SLOTS RUN_TIME TIME_LEFT XXXXXXX RUN NetID sn_short JobExample1 1 1 0 second(s) 1:30 L
To cancel a job, use the following command, where XXXXXXX is the JobID:
[NetID@ada1 ~]$ bkill XXXXXXX Job <XXXXXXX> is being terminated
tamubatch is an automatic batch job script that submits jobs for the user without the need of writing a batch script on the Ada and Terra clusters. The user just needs to provide the executable commands in a text file and tamubatch will automatically submit the job to the cluster. There are flags that the user may specify which allows control over the parameters for the job submitted.
tamubatch is still in beta and has not been fully developed. Although there are still bugs and testing issues that are currently being worked on, tamubatch can already submit jobs to both the Ada and Terra clusters if given a file of executable commands.
For more information, visit this page.
Translating Ada/LSF <--> Terra/Slurm
The HPRC Batch Translation page contains information on converting between LSF, PBS, and Slurm.
Our staff has also written some example jobs for specific software. These software-specific examples can be seen on the Individual Software Pages where available.
Files can be transferred to Ada using the scp command or a file transfer program.
Our users most commonly utilize:
Advice: while GUIs are acceptable for file transfers, the cp and scp commands are much quicker and may significantly benefit workflow.
Reliably Transferring Large Files
For files larger than several GB, users will want to consider the use of a more fault-tolerant utility such as rsync.
[NetID@ada1 ~]$ rsync -av [-z] localdir/ userid@remotesystem:/path/to/remotedir/
An rsync example can be seen on the Ada Fast Transfer page.
Graphic User Interfaces (Visualization)
The Open on Demand Portal is the preferred way to access visual applications on Ada. From the OOD portal, you can browse and manage the files in your home and scratch directories. This allows user's to easily view media files (.png .jpeg .mp4 etc) through their web browser of choice.
Here's a quick summary of the functions of the OOD Portal:
- View and submit visualization jobs
- View and submit interactive applications such as Matlab & ANSYS
- View your active jobs on Ada
- Access the shell of Ada, Terra, and Curie
- View and manage the files in your Ada home and scratch directories
Shared spaces are directories that can be accessed by multiple authorized users. In order to request a shared space, users must submit their request to email@example.com with the following information:
- Number of files to be stored in the shared directory (approximate)
- Total amount of data to be stored in the shared directory (approximate)
- Primary contact for authorizing access to the shared directory
- Initial users to have access to the shared directory after creation
The location of the shared directory will be in