Hprc banner tamu.png

Difference between revisions of "SW:ANSYS:ICEM"

From TAMU HPRC
Jump to: navigation, search
(Ada Example)
(Terra Example)
 
(17 intermediate revisions by 4 users not shown)
Line 1: Line 1:
==ICEM CFD==
+
=ICEM CFD=
 +
__TOC__
 +
==Description==
 
ANSYS ICEM CFD software supports a wide range of direct CAD interfaces and geometry readers, including faceted geometry and mesh readers. Components from different formats can easily be combined within one meshing session. The flexible mesh generation tools offer the capability to parametrically create volume or surface meshes from geometry or mesh in multi-block structured, unstructured hexahedral, Cartesian, tetrahedral, tetra/prism hybrid, hexa hybrid and unstructured quad/tri shell formats. - Homepage: [http://ansys.com/ http://ansys.com/]
 
ANSYS ICEM CFD software supports a wide range of direct CAD interfaces and geometry readers, including faceted geometry and mesh readers. Components from different formats can easily be combined within one meshing session. The flexible mesh generation tools offer the capability to parametrically create volume or surface meshes from geometry or mesh in multi-block structured, unstructured hexahedral, Cartesian, tetrahedral, tetra/prism hybrid, hexa hybrid and unstructured quad/tri shell formats. - Homepage: [http://ansys.com/ http://ansys.com/]
  
==Access==
+
{{:SW:ANSYS:Help}}
ANSYS ICEM CFD is open to all HPRC users when used within the terms of our ANSYS license agreement.
 
  
<font color=red>''IMPORTANT NOTE''</font> ''REGARDING THE ANSYS LICENSE:''
+
{{:SW:ANSYS:Access}}
    Terms of the ANSYS Academic license require that all users be within 50 miles
 
        of the TAMU main campus in College Station, TX. Use of this software outside
 
        the designated area represents a breach of the license and any users caught
 
        doing so may be subject to account suspension and/or other action.
 
 
 
If you have particular concerns about whether specific usage falls within the TAMU HPRC license, please send an email to the HPRC Helpdesk.
 
Usage of ANSYS ICEM CFD is restricted by the number of available tokens. To see the number of available tokens, use the [[SW:License_Checker | License Checker Tool]].
 
  
 
===Loading the Module===
 
===Loading the Module===
 
To see all versions of ANSYS available on Ada:
 
To see all versions of ANSYS available on Ada:
  module spider ANSYS
+
[NetID@cluster ~]$ '''module spider ANSYS'''
  
To load the default ANSYS module on Ada:
+
To load a particular version of ANSYS on Ada (Example: 19.3):
  module load ANSYS
+
[NetID@cluster ~]$ '''module load ANSYS/19.3'''
  
To load a particular version of ANSYS on Ada (Example: 17.1):
+
===Known Issues===
  module load ANSYS/17.1
+
There is a known issue when using the Ansys GUI for any version of Ansys. The Geometry Editor in the Ansys Workbench will not run properly without also loading an NVIDIA module.<br>
 +
To bypass this bug, load the '''OpenGL/NVIDIA''' module before running the Ansys Workbench:
 +
[ netID@cluster ~]$ '''module load OpenGL/NVIDIA'''
  
 +
There is a known issue with all Ansys versions on '''Terra''' that requires unsetting a Slurm environment variable. To prevent this error, please make sure to have the following line in your job script '''before''' the launch command.
 +
unset SLURM_GTIDS
 +
 +
The same error occurs when using the Ansys GUI in a VNC job on '''Terra'''. To prevent this error, please use the following command in your VNC job:
 +
[ netID@terra ~]$ '''unset SLURM_GTIDS'''
 
{{:SW:Login_Node_Warning}}
 
{{:SW:Login_Node_Warning}}
  
Line 29: Line 30:
  
 
===Ada Example===
 
===Ada Example===
<pre>
+
'''Example 1:''' A serial (single core) ICEM CFD Job example:
#BSUB -J ICEMJob1   # sets the job name to ICEMJob1.
+
#BSUB -J ICEMJob1     # sets the job name to ICEMJob1.
#BSUB -L /bin/bash       # uses the bash login shell to initialize the job's execution environment.
+
#BSUB -L /bin/bash           # uses the bash login shell to initialize the job's execution environment.
#BSUB -W 5:00             # sets to 5 hours the job's runtime wall-clock limit.
+
#BSUB -W 5:00               # sets to 5 hours the job's runtime wall-clock limit.
#BSUB -n 1               # assigns 1 core for execution.
+
#BSUB -n 1                   # assigns 1 core for execution.
#BSUB -R "span[ptile=1]" # assigns 1 core per node.
+
#BSUB -R "span[ptile=1]"     # assigns 1 core per node.
#BSUB -R "rusage[mem=5000]"  # reserves 5000MB per process/CPU for the job (5GB * 2 Cores = 10GB per node)
+
#BSUB -R "rusage[mem=5000]"  # reserves ~5GB per process/CPU for the job  
#BSUB -M 5000   # sets to 5,000MB (~5GB) the per process enforceable memory limit.
+
#BSUB -M 5000     # sets to ~5GB the per process enforceable memory limit.
#BSUB -o stdout1.%J       # directs the job's standard output to stdout1.jobid
+
#BSUB -o stdout1.%J         # directs the job's standard output to stdout1.jobid  
 +
 +
 +
## Load the necessary modules
 +
module purge
 +
module load ANSYS/19.3
 +
 +
## Launch the solver with proper parameters
 +
icemcfd -batch ''inputfile''
  
  
##Load the necessary modules
+
To submit the batch job, run: (where ''jobscript'' is a file that looks like one of the above examples)
module load ANSYS/17.1
+
[ NetID@ada1 ~]$ '''bsub < jobscript'''
  
## Launch the solver with proper parameters
+
===Terra Example===
icemcfd -batch FileName
+
'''Example 1:''' A serial (single core) ICEM CFD Job example: (Last updated March 24, 2017)
</pre>
+
#!/bin/bash
 +
 +
##NECESSARY JOB SPECIFICATIONS
 +
#SBATCH --job-name=ICEMJob      # Sets the job name to ICEMJob
 +
#SBATCH --time=5:00:00          # Sets the runtime limit to 5 hr
 +
#SBATCH --ntasks=1              # Requests 1 core
 +
#SBATCH --ntasks-per-node=1    # Requests 1 core per node (1 node)
 +
#SBATCH --mem=5G                # Requests 5GB of memory per node
 +
#SBATCH --output=stdout1.o%J    # Sends stdout and stderr to stdout1.o[jobID]
 +
 +
## Load the necessary modules
 +
module purge
 +
module load ANSYS/19.3
 +
 +
## Known error fix
 +
unset SLURM_GTIDS
 +
 +
## Launch the solver with proper parameters
 +
icemcfd -batch ''inputfile''
  
FileName would be an ICEM CFD replay script(.rpl) or a tcl-script(.tcl).
 
  
To submit the batch job, run:
+
To submit the batch job, run: (where ''jobscript'' is a file that looks like one of the above examples)
 
+
[ NetID@terra ~]$ '''sbatch ''jobscript'''''
<pre>
 
bsub < jobscript
 
</pre>
 
 
 
where jobscript would look like the example above.
 
 
 
===Terra Example===
 
 
 
<pre>
 
TO BE ADDED LATER.
 
</pre>
 
  
 
{{:SW:VNC_Node_Warning}}
 
{{:SW:VNC_Node_Warning}}
 
 
[[Category:Software]]
 
[[Category:Software]]

Latest revision as of 13:33, 10 June 2021

ICEM CFD

Description

ANSYS ICEM CFD software supports a wide range of direct CAD interfaces and geometry readers, including faceted geometry and mesh readers. Components from different formats can easily be combined within one meshing session. The flexible mesh generation tools offer the capability to parametrically create volume or surface meshes from geometry or mesh in multi-block structured, unstructured hexahedral, Cartesian, tetrahedral, tetra/prism hybrid, hexa hybrid and unstructured quad/tri shell formats. - Homepage: http://ansys.com/

Help

  • Student Forum. Register here and discuss simulation with students worldwide
  • Access to advanced material and video - see instructions here

Documentation

The PDF documentation for ANSYS release 18.2 is available for our users and can be found at: https://hprc.tamu.edu/softwareDocs/ansys/

Access

ANSYS is open to all HPRC users when used within the terms of our ANSYS license agreement.

IMPORTANT NOTE REGARDING THE ANSYS LICENSE: (July 12, 2017)

   Use of ANSYS is only permitted for users that are affiliated with Texas A&M at 
   College Station.  Users meeting this criteria are permitted to use ANSYS on HPRC 
   systems from anywhere in the United States (including Alaska and Hawaii).  Use 
   of this software outside the designated area represents a breach of the license 
   and any users caught doing so may be subject to account suspension and/or other action.

If you have particular concerns about whether specific usage falls within the TAMU HPRC license, please send an email to the HPRC Helpdesk. Usage of ANSYS is restricted by the number of available tokens. To see the number of available tokens, use the License Checker Tool.

Loading the Module

To see all versions of ANSYS available on Ada:

[NetID@cluster ~]$ module spider ANSYS

To load a particular version of ANSYS on Ada (Example: 19.3):

[NetID@cluster ~]$ module load ANSYS/19.3

Known Issues

There is a known issue when using the Ansys GUI for any version of Ansys. The Geometry Editor in the Ansys Workbench will not run properly without also loading an NVIDIA module.
To bypass this bug, load the OpenGL/NVIDIA module before running the Ansys Workbench:

[ netID@cluster ~]$ module load OpenGL/NVIDIA

There is a known issue with all Ansys versions on Terra that requires unsetting a Slurm environment variable. To prevent this error, please make sure to have the following line in your job script before the launch command.

unset SLURM_GTIDS

The same error occurs when using the Ansys GUI in a VNC job on Terra. To prevent this error, please use the following command in your VNC job:

[ netID@terra ~]$ unset SLURM_GTIDS

Usage on the Login Nodes

Please limit interactive processing to short, non-intensive usage. Use non-interactive batch jobs for resource-intensive and/or multiple-core processing. Users are requested to be responsible and courteous to other users when using software on the login nodes.

The most important processing limits here are:

  • ONE HOUR of PROCESSING TIME per login session.
  • EIGHT CORES per login session on the same node or (cumulatively) across all login nodes.

Anyone found violating the processing limits will have their processes killed without warning. Repeated violation of these limits will result in account suspension.
Note: Your login session will disconnect after one hour of inactivity.

Usage on the Compute Nodes

Non-interactive batch jobs on the compute nodes allows for resource-demanding processing. Non-interactive jobs have higher limits on the number of cores, amount of memory, and runtime length.

For instructions on how to create and submit a batch job, please see the appropriate wiki page for each respective cluster:

Ada Example

Example 1: A serial (single core) ICEM CFD Job example:

#BSUB -J ICEMJob1	     # sets the job name to ICEMJob1.
#BSUB -L /bin/bash           # uses the bash login shell to initialize the job's execution environment.
#BSUB -W 5:00                # sets to 5 hours the job's runtime wall-clock limit.
#BSUB -n 1                   # assigns 1 core for execution.
#BSUB -R "span[ptile=1]"     # assigns 1 core per node.
#BSUB -R "rusage[mem=5000]"  # reserves ~5GB per process/CPU for the job 
#BSUB -M 5000		     # sets to ~5GB the per process enforceable memory limit.
#BSUB -o stdout1.%J          # directs the job's standard output to stdout1.jobid 


## Load the necessary modules
module purge
module load ANSYS/19.3

## Launch the solver with proper parameters
icemcfd -batch inputfile


To submit the batch job, run: (where jobscript is a file that looks like one of the above examples)

[ NetID@ada1 ~]$ bsub < jobscript

Terra Example

Example 1: A serial (single core) ICEM CFD Job example: (Last updated March 24, 2017)

#!/bin/bash

##NECESSARY JOB SPECIFICATIONS
#SBATCH --job-name=ICEMJob      # Sets the job name to ICEMJob
#SBATCH --time=5:00:00          # Sets the runtime limit to 5 hr
#SBATCH --ntasks=1              # Requests 1 core
#SBATCH --ntasks-per-node=1     # Requests 1 core per node (1 node)
#SBATCH --mem=5G                # Requests 5GB of memory per node
#SBATCH --output=stdout1.o%J    # Sends stdout and stderr to stdout1.o[jobID]

## Load the necessary modules
module purge
module load ANSYS/19.3

## Known error fix
unset SLURM_GTIDS 

## Launch the solver with proper parameters
icemcfd -batch inputfile 


To submit the batch job, run: (where jobscript is a file that looks like one of the above examples)

[ NetID@terra ~]$ sbatch jobscript

Usage on the VNC Nodes

The VNC nodes allow for usage of the a graphical user interface (GUI) without disrupting other users.

VNC jobs and GUI usage do come with restrictions. All VNC jobs are limited to a single node (Terra: 28 cores/64GB). There are fewer VNC nodes than comparable compute nodes.

For more information, including instructions, on using software on the VNC nodes, please visit our Terra Remote Visualization page.