Ansys Remote Solve Manager User's Guide
Ansys Remote Solve Manager User's Guide
Ansys Remote Solve Manager User's Guide
ANSYS, Ansys Workbench, AUTODYN, CFX, FLUENT and any and all ANSYS, Inc. brand, product, service and feature
names, logos and slogans are registered trademarks or trademarks of ANSYS, Inc. or its subsidiaries located in the
United States or other countries. ICEM CFD is a trademark used by ANSYS, Inc. under license. CFX is a trademark
of Sony Corporation in Japan. All other brand, product, service and feature names or trademarks are the property
of their respective owners. FLEXlm and FLEXnet are trademarks of Flexera Software LLC.
Disclaimer Notice
THIS ANSYS SOFTWARE PRODUCT AND PROGRAM DOCUMENTATION INCLUDE TRADE SECRETS AND ARE CONFID-
ENTIAL AND PROPRIETARY PRODUCTS OF ANSYS, INC., ITS SUBSIDIARIES, OR LICENSORS. The software products
and documentation are furnished by ANSYS, Inc., its subsidiaries, or affiliates under a software license agreement
that contains provisions concerning non-disclosure, copying, length and nature of use, compliance with exporting
laws, warranties, disclaimers, limitations of liability, and remedies, and other provisions. The software products
and documentation may be used, disclosed, transferred, or copied only in accordance with the terms and conditions
of that software license agreement.
ANSYS, Inc. and ANSYS Europe, Ltd. are UL registered ISO 9001: 2015 companies.
For U.S. Government users, except as specifically granted by the ANSYS, Inc. software license agreement, the use,
duplication, or disclosure by the United States Government is subject to restrictions stated in the ANSYS, Inc.
software license agreement and FAR 12.212 (for non-DOD licenses).
Third-Party Software
See the legal information in the product help files for the complete Legal Notice for ANSYS proprietary software
and third-party software. If you are unable to access the Legal Notice, contact ANSYS, Inc.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. iii
Remote Solve Manager (RSM)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
iv of ANSYS, Inc. and its subsidiaries and affiliates.
Remote Solve Manager (RSM)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. v
Remote Solve Manager (RSM)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
vi of ANSYS, Inc. and its subsidiaries and affiliates.
List of Figures
1.1. General RSM Workflow .......................................................................................................................... 15
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. vii
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
viii of ANSYS, Inc. and its subsidiaries and affiliates.
List of Tables
4.1. Overview of ARC Setup Options ............................................................................................................ 65
4.2. Options for 'arcsubmit' .......................................................................................................................... 92
4.3. Options for 'arcconfig node modify' ...................................................................................................... 94
4.4. Options for 'arcconfig queue modify' ..................................................................................................... 98
4.5. Operators for Migration ....................................................................................................................... 101
4.6. Options for Migration .......................................................................................................................... 101
7.1. Options for Creating an RSM Configuration ......................................................................................... 139
7.2. Arguments Used for Managing Credentials .......................................................................................... 148
7.3. Operators for Migration ....................................................................................................................... 151
7.4. Options for Migration .......................................................................................................................... 151
11.1. Key RSM Log Files .............................................................................................................................. 199
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. ix
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
x of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 1: RSM Overview
Ansys Remote Solve Manager (RSM) provides the central framework for configuring and monitoring job
submission to HPC resources. Whether jobs are submitted to a cluster or to a Cloud portal, RSM's integ-
rated environment and tools enable you to easily connect to existing IT infrastructure, providing you
with seamless access to powerful compute resources when needed.
Jobs can be submitted directly to RSM from client applications such as Ansys Workbench, or indirectly
via a Cloud portal.
• RSM Configuration. Define configurations that enable you to run Ansys applications on HPC resources.
With its wizard-like interface, the RSM Configuration application lets you easily create configurations
for cluster or portal job submission.
RSM configurations enable you to integrate RSM with a third-party job scheduler such as Microsoft
HPC or LSF, or with an Ansys RSM Cluster (ARC). You can also create a configuration for job submission
to a third-party Cloud compute service. Regardless of the resource type, all RSM configurations are
defined in a consistent way.
Configuration tasks include establishing communication protocols, specifying file handling methods,
setting up RSM queues, and caching account credentials. For more information, see RSM Configura-
tion (p. 31).
• Ansys RSM Cluster (ARC). If you are not using a third-party job scheduler such as Microsoft HPC or
LSF, you can use the built-in Ansys RSM Cluster (ARC) system that is provided with every RSM install-
ation. An ARC operates in the same way that a commercial cluster does, running Ansys applications
in local or distributed mode, but uses its own scheduling capability rather than that of a third-party
job scheduler.
An ARC that comprises a single node (whether it be either a user's local machine or a specific machine
in your network) does not require any special setup. An ARC that comprises multiple nodes requires
service configuration and node setup, but provides more powerful features and enables you to run
distributed parallel jobs in a multi-node environment. For more information, see Ansys RSM Cluster
(ARC) Configuration (p. 63).
• Job Monitoring. View the status of submitted jobs, view job logs, and troubleshoot failed jobs directly
from the Workbench, or using the RSM Job Monitoring application.
– For information on monitoring jobs in Workbench, see Monitoring and Controlling Remote Solve
Manager Jobs in Workbench in the Workbench User's Guide.
– For information about monitoring jobs using the RSM Job Monitoring application, see RSM Job
Monitoring (p. 185) in the RSM User's Guide.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 11
RSM Overview
Job
A processing task submitted from a client application such as Ansys Workbench. Examples include
solution updates, project updates, and design point updates submitted to RSM for remote processing.
Certain files in RSM are used to define and control RSM jobs. A job script is used to perform a
processing task (such as running a finite element solver). A job template can be used to further
customize a job.
Client Machine
The computer on which Ansys Workbench and Ansys applications are installed, and on which jobs
are submitted to RSM. RSM is automatically installed with Ansys Workbench products.
Client Application
The Ansys application that submits jobs to RSM. Examples include Ansys Workbench, Ansys Fluent,
and Ansys CFX.
HPC Resource
A system that supports High Performance Computing, such as a cluster, or a cloud computing en-
vironment. The system uses a queueing system to make optimal use of all available resources.
Node
A single computer. A multi-node HPC system consists of a head node, where jobs are submitted
for scheduling, and one or more execution nodes, which are used for computational work.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
12 of ANSYS, Inc. and its subsidiaries and affiliates.
RSM Roles and Terminology
Note:
• The local machine (localhost) can be an execution node if it is serving as the HPC
resource.
Queue
A list of execution nodes that are suited to run a particular class of jobs.
When you submit a job to RSM, you submit it to an RSM Queue, which maps to an HPC Queue.
The HPC Queue has one or more execution nodes assigned to it, and determines when and where
the job will run based on resource requests and current available resources. Queue definitions are
part of the configurations that are defined in RSM.
HPC Configuration
A set of properties defined in RSM which specify the following information about an HPC resource:
• HPC type (ARC, Windows HPC, LSF, PBS Pro, SLURM, UGE (SGE), Custom)
• A set of HPC queues that will be used for running Ansys applications
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 13
RSM Overview
When you submit a job to RSM, you select an RSM queue in the solve properties. The RSM queue is
associated with a configuration that is defined in RSM. The RSM configuration specifies how the client
machine will communicate with the HPC resource, and identifies HPC queues.
Every RSM installation contains a default Ansys RSM Cluster (ARC) that can be used on the local machine
or configured on a remote machine. For more information see Ansys RSM Cluster (ARC) Configura-
tion (p. 63). If your organization uses a commercial cluster, such as Windows HPC, LSF or PBS Pro, you
can configure RSM to submit jobs to that cluster as well. For more information see RSM Integration
with a Cluster or Cloud Portal (p. 121).
Working with Ansys Support, it is also possible to create a custom configuration that enables users to
submit jobs to a third-party Cloud.
If jobs will be submitted to a remote cluster, a Launcher Service must be installed on the cluster submit
host if the client cannot communicate directly with the remote resource. This service is used to launch
a User Proxy process, which authenticates the account prior to job submission.
Note:
The Launcher Service is not required if SSH is being used as a communication protocol, or
if the remote resource is a Cloud. In these cases, authentication is already taken care of.
The Cluster API manages cluster operations such as handling inputs and outputs, carrying out job
commands, and retrieving cluster status information.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
14 of ANSYS, Inc. and its subsidiaries and affiliates.
File Handling
Below is more detailed information about how files are handled in RSM.
When you are setting up a solve job in a client application (for example, Ansys Workbench), job
input files are placed in a working directory on the client machine. The location of this directory is
controlled by the client application. Refer to the client application documentation to determine
where input files are placed when submitting jobs to RSM.
The HPC staging directory is a shared directory that all HPC nodes can access.
If the client working directory is always set under the HPC staging directory in the client application,
then no file transfer is needed, because the client files will already be in the HPC staging directory.
If files need to be transferred, you can specify the desired transfer method in the RSM configuration.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 15
RSM Overview
HPC jobs can be run in the HPC staging directory, or in a scratch directory local to the execution
node. You specify this in the File Management properties when defining a configuration in RSM.
RSM provides the following options for transferring files from the client to the HPC resource:
You can also customize the file transfer process using your own mechanism.
In the case of a Cloud portal, client files will be transferred to HPC resources as needed using the
portal mechanism. File management operations related to job execution are configured on the HPC
side.
The file transfer method and HPC staging directory are specified in the File Management properties
of an RSM configuration. For more information, see Specifying RSM Configuration Settings (p. 35).
When you submit a job to RSM in a client application, you can specify a number of solve settings to be
used for the job. For example, the Mechanical application contains a Max number of used cores setting
that enables you to limit the number of CPUs/nodes allocated for the HPC job. This information is passed
along on the solver command line. The command line is parsed in the job script, and this information
is passed on to the HPC resource.
For information about how RSM integrates with Ansys applications, refer to the following:
1.4.1. RSM-Supported Applications and Solvers
1.4.2. RSM Integration with Workbench
• CFX
• Fluent
• Forte
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
16 of ANSYS, Inc. and its subsidiaries and affiliates.
RSM Supported Third-Party Job Schedulers/Commercial Batch-Queuing Systems
• Icepak
• Mechanical APDL
• Polyflow
• Explicit Dynamics
Note:
Rigid Body Dynamics jobs can be submitted only to an Ansys RSM Cluster (ARC).
Many Ansys Workbench applications enable you to use RSM; however, the following considerations
may apply:
• When a client application is restricted to the local machine, RSM may enable the client application
to run in the background.
• When a client application can send jobs to a remote machine, the job may be run completely on
one node, or may be broken into pieces so that each piece can run in parallel on multiple nodes
(possibly including the client machine). In the case where a job is being run in parallel on multiple
machines, you need to ensure that the software that controls the parallel processing (for example,
MPI) is supported on all of the execution nodes.
• Platform LSF
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 17
RSM Overview
Operating systems: Red Hat Enterprise Linux (RHEL), SUSE Linux Enterprise (SLES)
• PBS Professional
Operating systems: Red Hat Enterprise Linux (RHEL), SUSE Linux Enterprise (SLES)
Operating systems: Red Hat Enterprise Linux (RHEL), SUSE Linux Enterprise (SLES)
Operating systems: Red Hat Enterprise Linux (RHEL), SUSE Linux Enterprise (SLES)
• Windows HPC
Operating systems: Windows Server 2012 R2 (Standard) with Microsoft HPC Pack 2012 R2, Windows
Server 2016 with HPC Pack
Some stand-alone Ansys applications support a slightly different list of third-party job schedulers. Refer
to the Job Schedulers and Queuing Systems Support document at http://www.ansys.com/solutions/solutions-
by-role/it-professionals/platform-support/previous-releases.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
18 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 2: RSM Installation and Startup
For successful job submission to an HPC resource via RSM, the RSM application must be installed on
the same machine where the client application (for example, Workbench) is installed.
If jobs will be submitted to a remote cluster, then the RSM launcher service must also be installed and
running on the cluster submit host. This does not apply if SSH is being used for client-to-HPC commu-
nication, or if the remote HPC resource is a Cloud.
Important:
If after installing RSM and starting the RSM launcher service you wish to install another Ansys
product using the Ansys unified installer, make sure that you stop the RSM launcher service
before proceeding with the installation, and then restart it after the installation. See Installing
and Configuring the RSM Launcher Service (p. 20).
In this chapter:
2.1. RSM Software Installation
2.2. Installing and Configuring the RSM Launcher Service
2.3. Uninstalling RSM
2.4. Uninstalling the RSM Launcher Service
You can also use the standard Ansys product installation to install RSM by itself (see Installing a Stan-
dalone RSM Package (p. 19)).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 19
RSM Installation and Startup
a. Under Ansys Additional Tools, select the Remote Solve Manager Standalone Services
check box.
Note:
When you install a standalone RSM package, this does not mean that the RSM launcher
service is installed at the same time. You still need to install the RSM launcher service on
the cluster submit host. For instructions, see Installing and Configuring the RSM Launcher
Service (p. 20).
You must install the RSM launcher service on the cluster submit host.
Refer to the instructions below that pertain to the machine on which the RSM launcher service is being
installed:
2.2.1. Supported Platform Combinations
2.2.2. Installing and Configuring the RSM Launcher Service for Windows
2.2.3. Installing and Configuring the RSM Launcher Service for Linux
2.2.4. Configuring a Network Installation of RSM
• Windows to Windows
• Linux to Linux
• Windows to Linux
Submitting jobs from a Linux RSM client to a Windows submit host is not supported.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
20 of ANSYS, Inc. and its subsidiaries and affiliates.
Installing and Configuring the RSM Launcher Service
To use this method, you must install the RSM launcher service on the Linux submit host. This
eliminates the need for an external communication protocol such as SSH, and enables RSM to
communicate directly with the submit host.
There are two options for starting the RSM launcher service on the Linux submit host:
For more information refer to Installing and Configuring the RSM Launcher Service for Linux (p. 22).
For instructions, refer to Configuring RSM to Use SSH for Job Submission and/or File Transfers to a
Remote Linux Cluster (p. 121).
2.2.2. Installing and Configuring the RSM Launcher Service for Windows
When RSM is installed on a Windows machine, you can configure and install the RSM launcher process
as a Windows service so that it can be started automatically when the Windows system starts up. You
can also uninstall and restart the service using RSM-provided tools.
Note:
• The RSM launcher service cannot be started from a network installation. It is recommen-
ded that you install RSM on a local machine.
• For GPU requirements when Windows is installed as a service, see Requirements for the
GPU Accelerator in Mechanical APDL in the Ansys, Inc. Installation Guides.
2. Ensure that Ans.Rsm.* processes are not running in the Windows Task Manager.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 21
RSM Installation and Startup
AnsConfigRSM.exe -launcher
Note:
If the RSM launcher service has been removed, you can also use the above sequence of steps to re-
configure the service.
Important:
• If you change any system environment variables, you must restart the RSM launcher
service in order for the changes to take effect. If you change your user environment
variables, make sure that you end your Ans.Rsm.UPHost.exe processes (if any) on
the affected machine before trying to run jobs again.
2.2.3. Installing and Configuring the RSM Launcher Service for Linux
When installing the RSM launcher service on Linux, you can start the service manually via startup
scripts, or install it as a daemon that will start the service automatically when the machine is booted.
The following RSM configuration topics for Linux are discussed in this section:
2.2.3.1. Adding Common Job Environment Variables for Jobs
2.2.3.2. Installing the RSM Launcher Service for Linux
# The following examples show loading environment settings specific to batch system (for example, LSF, SGE/UGE).
# If defined, RSM service and jobs should then inherit this environment when a job is run.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
22 of ANSYS, Inc. and its subsidiaries and affiliates.
Installing and Configuring the RSM Launcher Service
# . /home/batch/lsf7.0/conf/profile.lsf
# . /home/batch/SGE6.2u2/default/common/settings.sh
Note:
• This profile only works on Linux. Windows users should modify their environment via
the environment interface in Windows.
• This profile only works when jobs are submitted to a remote HPC resource in non-SSH
mode. It does not work when running jobs locally (for example, on localhost using
the local queue), or if SSH is being used.
When the RSM launcher service is started manually, it runs as a process for the user who initiated
the service. A manually started RSM launcher service is stopped each time the machine is rebooted;
after a reboot, before you submit any jobs to RSM you must first restart the RSM launcher service
by running the startup script. For security reasons, it is recommended that you do not start and
run the RSM launcher service process manually as the "root" user.
If you would prefer to start the RSM launcher service automatically when the machine is booted,
you can configure a daemon as described in Starting the RSM Launcher Service Automatically at
Boot Time for Linux (p. 24).
Note:
When installing RSM to a multi-user Linux machine, Ansys strongly recommends that
you set up RSM as a daemon (see Starting the RSM Launcher Service Automatically at
Boot Time for Linux (p. 24)). Running RSM as a daemon allows you to maintain consistent
settings. If RSM is not run as daemon, the settings vary depending on which user first
starts RSM processes.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 23
RSM Installation and Startup
This script is generated as part of the RSM installation process and is located in the [RSMIn-
stall]/RSM/Config/tools/linux directory. If for some reason this script was not generated
during installation or is for other reasons not available, you can generate it yourself. For instruc-
tions, see Generating the RSM Service Startup Script for Linux (p. 200) in the RSM Troubleshoot-
ing (p. 199) section.
2.2.3.2.1.1. Manually Running the RSM Launcher Service Script for Linux
You can run the RSM launcher service script to manually start, stop, check the status of, and
restart the RSM launcher service.
You can start the RSM launcher service manually by running the service script with the
command line option start, as shown below:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher start
Note:
You can stop the RSM launcher service manually by running the service script with the
command line option stop, as shown below:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher stop
You can check the status of the RSM launcher service manually by running the service script
with the command line option status, as shown below:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher status
You can restart the RSM launcher service manually by running the service script with the
command line option restart, as shown below:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher restart
2.2.3.2.2. Starting the RSM Launcher Service Automatically at Boot Time for Linux
You can configure the RSM launcher service to start automatically when the machine is booted
by configuring it as a “daemon” service (if the service is not configured to start automatically,
then it must be started manually, as described in Starting the RSM Launcher Service Manually
for Linux (p. 23)). Daemon services are scripts or programs that run persistently in the background
of the machine, and which are usually executed at startup by the defined runlevel.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
24 of ANSYS, Inc. and its subsidiaries and affiliates.
Installing and Configuring the RSM Launcher Service
2.2.3.2.2.2. Working with the RSM Automatic Startup (Daemon) Service for Linux
2.2.3.2.2.1. Installing the RSM Automatic Startup (Daemon) Service for Linux
To install the RSM launcher service as a daemon, you must have system administrative permis-
sions (that is, you must be logged in and installing as a “root” user or “sudoer”).
For security reasons, it is recommended that you do not run the RSM launcher service as the
root user. Many Linux versions allow only root users to listen to specific ports, so the ports that
are required by RSM may be blocked by system administration. For these reasons, the RSM
daemon service installation will create a non-root user account with no logon called rsmadmin;
the account is a member of the rsmadmins user group, and has a home directory of
/home/rsmadmin. The RSM daemon service will then be run by the rsmadmin user.
Note:
• The RSM daemon service installation will only create the rsmadmin user account
if the account does not already exist. The same is true for the rsmadmins user
group if the group name does not exist. The account/group will be created locally
on the computer on which the RSM launcher service will be run. If you want the
account/group to be managed in the master server by Network Information Service
(NIS), you need to ask your IT department to create an rsmadmin user account
and rsmadmins group from NIS before running the RSM daemon service script.
• When an RSM package is installed under a directory, make sure that all its parent
directories (not the files in the directory) have both read and execution permissions
so that the RSM launcher service executable can be started by a non-root user.
There are two ways to install the RSM launcher service as a daemon: by running the rsmconfig
script, or by running the install_daemon script. The difference between the two methods
is that whereas the rsmconfig script always generates a fresh service script before starting
the service installation, the install_daemon script assumes that the service script is always
available in the WBInstallDir/RSM/Config/tools/linux directory and uses the existing
script for service installation, allowing the system administrator to perform advanced script
customizations before the service is installed.)
Both scripts are located in the RSM/Config/tools/linux directory and have the same
command line option.
tools/linux#> ./rsmconfig -help
Options:
-launcher: Install RSM Launcher service.
tools/linux# ./install_daemon
Usage: ./install_daemon [-launcher]
Options:
-launcher: Install RSM Launcher service.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 25
RSM Installation and Startup
To install the RSM launcher service as a daemon service, run either the rsmconfig script or
the install_daemon script, as follows:
4. Enter the script into the terminal window, adding the command line option -launcher.
For example:
tools/linux#> ./rsmconfig -launcher
Once the daemon service is installed, the RSM launcher service will be started automatically
without rebooting. The next time the machine is rebooted, the installed RSM launcher service
will be started automatically.
To verify that the automatic boot procedure is working correctly, reboot the system and check
to see that the service is running by typing the appropriate ps command and looking for
Ans.Rsm in the resulting display:
ps aux | grep Ans.Rsm
2.2.3.2.2.2. Working with the RSM Automatic Startup (Daemon) Service for Linux
Once the RSM daemon service is configured, any user can check the status of the service. System
administrators can also start or restart the service.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
26 of ANSYS, Inc. and its subsidiaries and affiliates.
Installing and Configuring the RSM Launcher Service
./etc/init.d/rsmlauncher221 restart
Important:
If you change any system environment variables, you must restart the RSM launcher
service in order for the changes to take effect. If you change your user environment
variables, make sure that you end your Ans.Rsm.UPHost.exe processes (if any)
on the affected machine before trying to run jobs again.
Perform the following steps on each machine where the same RSM package is used:
1. Set the system environment variable ANSYS_RSM_APPSETTINGS_DIR to point to the new RSM
application settings configuration directory. Create the directory if it does not exist. Then, add a
sub-directory for the RSM version (for example, v221). Copy the Ans.Rsm.AppSettings.config
file into the version sub-directory (for example, C:\rsmconfigs\v221). This file is located in the
RSM\Config folder where Ansys products are installed on a network drive.
Note:
On Linux:
• If you are installing the RSM launcher service as a daemon, and/or installing Ansys
RSM Cluster (ARC) services as daemons, you must set the ANSYS_RSM_APPSET-
TINGS_DIR variable in the sudo environment.
To ensure that the variable is passed correctly when installing these services as
daemons, run the applicable command below:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 27
RSM Installation and Startup
For information about setting environment variables in Linux, see Adding Common Job Environment
Variables for Jobs (p. 22).
3. If the RSM launcher service was installed on the HPC submit host, restart the launcher service on
that machine:
• For Windows: On your Administrative Tools or Administrative Services page, open the Services
dialog box. Right-click the desired service and select Restart.
• For Linux: Log into a Linux account with administrative privileges and ensure that Ans.Rsm.*
processes are not running. In a terminal window, run the following command:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher restart
If only the launcher service is installed on a machine, follow the steps below to uninstall it.
2.4.1. Uninstalling the RSM Launcher Service for Windows
2.4.2. Manually Uninstalling the RSM Launcher Service for Linux
2.4.3. Uninstalling the RSM Automatic Startup (Daemon) Service for Linux
2. Ensure that Ans.Rsm.* processes are not running in the Windows Task Manager.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
28 of ANSYS, Inc. and its subsidiaries and affiliates.
Uninstalling the RSM Launcher Service
Note:
• If you are using a Windows 7 operating system, you may need to select the Run as
administrator option from the right-click context menu.
• The uninstaller can stop the service only if it was started by and is owned by the
user performing the uninstall.
6. After the service has been uninstalled, delete the RSM installation directory.
4. Enter the rsmunconfig script into the command line, as shown below:
tools/linux#> ./rsmunconfig -launcher
Note:
• The uninstaller can only stop the service if it was started by and is owned by the user
performing the uninstall.
• If the service was running as a normal user account, this account may not have enough
permission to stop and kill other users' processes running on the same machine. In this
case, you may need root permission to kill those processes.
2.4.3. Uninstalling the RSM Automatic Startup (Daemon) Service for Linux
As with RSM daemon service installation, only a system administrator can uninstall the RSM daemon
service. Also, the uninstaller can only stop the service if it was started by and is owned by the user
performing the uninstall.
You can uninstall the RSM daemon service in one of two ways:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 29
RSM Installation and Startup
tools/linux#> ./rsmunconfig
By default the rsmunconfig script does not remove the rsmadmin user account and rsmadmins
user group that were created earlier when service was configured. This allows the same account
and user group to be reused for the next service installation and configuration, and also prevents
the accidental deletion of important files from the rsmadmin home directory (/home/rsmadmin).
If you later decide that you do not want to keep the user account and user group, you can remove
them manually (p. 30) if needed.
This script enables you to uninstall the RSM daemon service as well as the rsmadmin user account
and rsmadmins user group. For example:
tools/linux#> ./uninstall_daemon -launcher -rmadmin
If you decide to remove the account and user group, you can do so by manually by adding the
-rmadmin command line option to the uninstall_daemon script located in the WBIn-
stallDir/RSM/Config/tools/linux directory. For example:
tools/linux#> ./uninstall_daemon -rmadmin
Important:
The service account and group cannot be deleted if the RSM launcher service is still being
run by that user account and service group name. You will be prompted to answer “Yes”
or “No” from the above command when there is no service is being run by these accounts
and RSM is trying to delete them.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
30 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 3: RSM Configuration
RSM serves as a gateway for users wanting to submit jobs to a local or remote compute resource from
applications such as Workbench.
The following are the main types of HPC resources to which RSM can be configured to submit jobs:
• Commercial Cluster. Windows HPC, LSF, PBS Pro, SLURM, UGE (SGE), and custom clusters are suppor-
ted. In this scenario, a third-party application takes care of job scheduling.
• Ansys RSM Cluster (ARC). If you are not using a third-party job scheduler, you can use an Ansys RSM
Cluster (ARC) as your scheduling system. This system is installed along with RSM. An ARC can be
configured on the local machine, or on a remote one.
• Third-Party Cloud. Provides access to Cloud-hosted compute services. Refer to RSM Custom Integra-
tion (p. 155) and contact Ansys customer support for assistance with integrating RSM with a Cloud.
Job submission is established through configurations that are defined using the RSM Configuration ap-
plication. An RSM configuration for a cluster contains information about the cluster submit host, the
system used for job scheduling, desired communication protocols, and the location of job directories.
An RSM configuration for a Cloud portal identifies the portal URL. When defining an RSM configuration
you also define RSM queues that will appear in client applications when users choose to submit jobs
to RSM. Each RSM queue maps to a particular HPC queue. For more information about the components
of an RSM configuration, refer to RSM Roles and Terminology (p. 12), or the Glossary (p. 209).
• Transfer job files from the client working directory to the HPC staging directory (if necessary)
• Submit jobs to a cluster's job scheduler or a portal 's server application for scheduling and distribution
Every RSM installation has one predefined configuration named localhost. This configuration uses
a basic Ansys RSM Cluster (ARC) to submit jobs to the local machine. This enables users to run certain
types of local jobs or Mechanical background jobs right out of the box, without any special setup.
You can create as many RSM configurations as you need. Each RSM configuration that you define is
stored in an .rsmcc file, and RSM queue definitions are stored in a single queues.rsmq file. These
files are stored in a specified RSM configuration directory. Users who want to submit jobs to an HPC
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 31
RSM Configuration
resource must have access to these files from their local machines. This can be accomplished by making
the RSM configuration directory a shared directory, or having users copy the files to their local machines.
Important:
• If you will be defining RSM configurations that multiple people will use for remote job
submission, and want to make them accessible to users through a shared directory, we
recommend that you first change the location of the RSM configuration directory (the
directory in which new configurations are saved). For more information see Setting the
RSM Configuration Directory (p. 32) and Sharing and Accessing RSM Configurations (p. 54).
In this section:
3.1. Setting the RSM Configuration Directory
3.2. Launching the RSM Configuration Application
3.3. Defining RSM Configurations
3.4. Sharing and Accessing RSM Configurations
3.5. Setting Up Job Directories and File Transfers
Windows: %APPDATA%\ANSYS\v221\RSM
Linux: ~/.ansys/v221/RSM
On Linux, ~ is the home directory of the account under which RSM is being run.
To verify the location of your RSM configuration directory, see Querying the Location of the RSM Con-
figuration Directory (p. 147).
Important:
The default directory is appropriate if you will be the only one running jobs on the local
machine, as you will have full control of this directory. However, if you want to share RSM
configurations with other users, we recommend that you do not share the default RSM
configuration directory because it is associated with a specific user account, and is therefore
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
32 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
not suitable for sharing. In this case you should change the RSM configuration directory before
creating RSM configurations, as described below.
1. Create a folder in a location that is not associated with a user account (for example,
C:\some\folder).
2. If you have already created RSM configurations, and are looking to copy them to the shareable
folder, or share the folder in which they currently reside, add the following required sub-folders to
the selected folder: ANSYS\v221\RSM.
Otherwise, RSM will add these sub-folders when you start RSM. This sub-location (for example,
C:\some\folder\ANSYS\v221\RSM) is where RSM configurations must reside.
3. Use the RSM Utilities application to set the JobManagement ConfigurationDirectory setting
to the new folder. See Specifying the Location of the RSM Configuration Directory (p. 147).
Once you have set the RSM configuration directory you can begin creating RSM configurations. See
Launching the RSM Configuration Application (p. 33).
Any new RSM configurations that you create will be saved to the specified directory (in this example,
C:\some\folder\ANSYS\v221\RSM).
If you will be sharing configurations with users, see Sharing and Accessing RSM Configurations (p. 54).
• If you are using a Windows system, select Start > Ansys 2022 R1 > RSM Configuration 2022 R1.
• Details about the HPC resource to which jobs will be submitted, such as the domain name of a cluster
submit host, or the URL of a Cloud portal
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 33
RSM Configuration
The RSM Configuration application provides a friendly, wizard-like interface that enables you to define
RSM configurations quickly and easily. Defining a configuration involves a 3-part series of steps.
Initially, the list of HPC Resources will include a localhost configuration, which is a basic, single-
node Ansys RSM Cluster (ARC) installed on the local machine. By default it has one RSM queue named
Local, which maps to an HPC queue named local. This queue will run Ansys applications on the
local machine.
Note:
If you will be sharing your RSM configurations with other users, do not edit the existing
localhost configuration to create a new configuration. The localhost configuration
always refers to "the local machine", which could be the cluster submit host or a user's local
machine depending on where job submission is being initiated.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
34 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
If you want to run Ansys applications on machines other than the local machine, then you will need to
create configurations for the HPC resources that you want to access (Ansys RSM Cluster, third-party
cluster, Cloud portal).
The following topics provide a general overview of creating and managing RSM configurations:
3.3.1. Creating a New RSM Configuration
3.3.2. Specifying RSM Configuration Settings
3.3.3. Deleting an RSM Configuration
For detailed information about setting up specific types of clusters, and creating RSM configurations
for those cluster types, see Ansys RSM Cluster (ARC) Configuration (p. 63) and RSM Integration with a
Cluster or Cloud Portal (p. 121).
• To create an RSM configuration from scratch, click , or right-click in the HPC Resources list in
the left pane and select Add HPC Resource.
• To create an RSM configuration by copying an existing RSM configuration (and making changes
to the copy), select the existing configuration in the HPC Resources list and click , or right-
click and select Duplicate HPC Resource. This method is ideal when you already have an RSM
configuration defined and want to create another configuration that is similar, but has one setting
that is different, such as the file transfer method.
Note:
To access or change the directory in which RSM configuration files are generated, refer to
Specifying a Directory for RSM Configuration Files (p. 147).
As you define an RSM configuration for an HPC resource, RSM validates paths and machine names
that you enter, and presents settings that are specific to the HPC type and options that you have
chosen.
1. On the HPC Resource tab, specifying HPC information (p. 36) such as the HPC type, machine
name of the cluster submit host, or URL of the Cloud portal
2. On the File Management tab, specifying file management (p. 41) properties that determine how
job input files get transferred to the HPC staging directory
3. On the Queues tab, importing or adding HPC queues (p. 49), and mapping them to RSM queues
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 35
RSM Configuration
When you finish specifying information on a tab, click Apply to validate the information and apply
it to the configuration. Colored icons indicate the status of information on each tab:
For detailed instructions and examples of specific configuration types, refer to the following:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
36 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
HPC Configuration
Name
The name of the configuration as it appears in the HPC Resources list in the left pane. Do not
use the name of an existing configuration.
HPC type
Choose this option if you are not integrating with a third-party cluster or Cloud portal.
• Windows HPC
• LSF
• PBS Pro
• SLURM
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 37
RSM Configuration
• UGE (SGE)
• Custom
To configure RSM for job submission to a third-party Cloud, contact Ansys Customer Support
for assistance.
If UGE (SGE) is selected as the HPC type, settings become available to specify Parallel Environ-
ment (PE) names:
Shared memory parallel processing enables you to distribute solve power over multiple pro-
cessors on the same machine. Distributed parallel processing enables you to distribute solve
power across multiple cores on a single node, or across multiple nodes. For information on
configuring parallel environments, consult the documentation of the simulation product you
are using.
RSM integrates with Windows HPC, LSF, PBS Pro, SLURM, and UGE (SGE) without requiring job
script customization. For custom cluster types, customization will likely be necessary to make
job submission work. Refer to RSM Custom Integration (p. 155).
Submit host
Identify the machine that serves as the cluster submit host. This is the machine that handles
job scheduling. In other words, it is the machine on which scheduling software is installed, or,
in the case of an Ansys RSM Cluster (ARC), the machine on which the ARC Master service has
been installed.
• If jobs will be submitted to the cluster submit host from any other machine, enter the submit
host's full domain name (for example, machineName.company.com), even if the machine
on which you are currently working (the local machine) is the submit host.
• If the machine on which you are currently working (the local machine) is the cluster submit
host, and jobs will not be submitted to it from any other machine, you can enter localhost
in this field.
Important:
• Correctly identifying the submit host is a crucial step, as this is the key piece of
information that enables RSM to communicate with the cluster.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
38 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
• If the current (local) machine is the submit host, do not enter localhost in the
Submit host field if jobs will be submitted to this machine from other machines.
You must use the full domain name in this case.
• If you specify the full domain name for the Submit host and choose to use the
RSM internal socket-based file transfer method, you will see a file transfer error
such as "port number cannot be zero" when you try to use this configuration locally.
If you intend to submit only local jobs using this configuration, change the Submit
host value to localhost and use the OS file transfer method. Otherwise, ensure
that the client machine and submit host are different machines.
Scheduler-specific arguments that will be added to the job submission command line of the
job scheduler. For example, you can enter job submission arguments to specify the queue (LSF,
PBS, SGE) or the nodegroup (MS HPC) name. For valid entries, see the documentation for your
job scheduler.
Use SSH protocol for inter and intra-node communication (Linux only)
This setting is used for distributed computing with multiple nodes involved. It specifies that
RSM and solvers use SSH for communications between Linux execution nodes, and within the
nodes themselves. If left deselected, this indicates that RSH is used.
This setting will be applied to all Linux HPC nodes, allowing for solvers to run in distributed
parallel mode.
When Ansys Fluent, Ansys CFX, Ansys Mechanical, and Ansys Mechanical APDL are configured
to send solves to RSM, their solvers will use the same RSH/SSH settings as RSM.
If integrating with a custom cluster, enter the domain name of the cluster submit host (for ex-
ample, machineName.company.com), then select the platform of this machine in the adjacent
drop-down.
If integrating with a custom portal or Cloud, enter the URL of the resource to which jobs will
be submitted, then make any random selection from the platform drop-down. (Since Custom
is a general option that can be used for a variety of HPC resource types, a selection must be
made in the OS drop-down in order to proceed. In the case of a portal or Cloud, the value se-
lected is unimportant and will simply not be used.)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 39
RSM Configuration
This is a keyword of your choosing that represents the HPC resource. It is a short word or phrase
that you will append to the file names of your custom integration files. For more information,
see RSM Custom Integration (p. 155).
Note:
Communication on the HPC Resource tab refers mainly to the submission of jobs to the
cluster submit host. The transfer of job files is handled independently according to the
settings on the File Management tab. For example, if you select Use SSH or custom
communication to the submit host on the HPC Resource tab, this does not mean that
you have to use SSH for file transfers. You may instead want to use a custom mechanism
or different file transfer method altogether. See Specifying File Management Proper-
ties (p. 41).
Specifies that the RSM client can use the RSM internal communication mechanism to directly
submit jobs to the HPC resource, and monitor HPC jobs. This requires that an IT administrator
open ports and adjust firewall settings on the HPC submit host to allow communication from
the RSM client.
When the submit host is a remote machine, the RSM launcher service launches a user proxy
process on the submit host which performs operations such as job submission, monitoring, and
file transfer on the user's behalf. The RSM launcher service will use one port, while each user
proxy process will use a separate port chosen by RSM. Ports for user proxy processes are chosen
from a port range if one has been specified in the RSM application settings (see Specifying a
Port Range for User Proxy Processes (p. 146)). Otherwise, RSM will randomly select a port that
is free.
This option is only available when the submit host is a Linux machine and the RSM client is a
Windows machine.
When a job from Windows client is submitted to a remote Linux cluster, this specifies that SSH
will be used to communicate with the submit host instead of RSM's internal communication
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
40 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
mechanism. Use this option if your IT administrator does not want to open ports and adjust
firewall settings to allow communication from the RSM client, in adherence with your organiz-
ation's IT policy. In this scenario, no RSM services need to be running on the remote submit
host.
In the Account name field, specify the account name that the Windows RSM client will use to
access the remote Linux submit host.
Note:
• This account must be set up before this mode can be used. For information on
configuring SSH to allow access from a Windows machine, see Configuring PuTTY
SSH (p. 123).
• This is not an account that is specified in the Credentials section of the RSM
Configuration application. The accounts listed there are RSM client accounts, not
user proxy accounts.
When you submit a job to RSM in a client application, a client working directory is created to which
all necessary job files are written. The location of this directory is configured in the client application.
For more information, refer to Setting Up Client Working Directories to Eliminate the Need for File
Transfers (p. 58).
If the client working directory is created under a shared directory that is visible to all HPC nodes
(in other words, it is already inside the shared HPC staging directory), then it is possible for the job
to be run directly in the working directory. Otherwise, if files need to be transferred from the client
working directory to an HPC staging directory, you will need to specify this in your RSM configuration.
You will also need to specify where jobs will run on the HPC side.
File management properties are specified on the File Management tab in the editing pane.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 41
RSM Configuration
Tip:
Use the Tell me more options to view detailed information about each file transfer
method so that you can select the method that best suits your IT environment, file
storage strategy, and simulation requirements.
Important:
The HPC staging directory must a shared directory that is visible to all HPC nodes.
Operating system file transfer to existing network share (Samba, CIFS, NFS)
Use this option when the HPC staging directory is a shared location that client machines can access.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
42 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
The RSM client finds the HPC staging directory via a Windows network share or Linux mount
point that has been set up on the client machine, and copies files to it using the built-in op-
erating system copy commands.
In the Staging directory path from client field, specify the path to the shared file system
as the RSM client sees it. A Windows client will see the shared file system as a UNC path (for
example, \\machine\shareName), while a Linux client will see a mounted directory (for
example, /mounts/cluster1/staging).
In the Staging directory path on cluster field, specify a path to the HPC staging directory
which all execution nodes can see. This is a path on the HPC side (for example, \\ma-
chine\staging on a Windows machine, or /staging on a Linux machine). This is the
path to which the client-side network share or mount point is mapped. The minimum path
is the location shared to the network (for example, from the Samba configuration). The rest
of the path can include subdirectories or they will be inferred from the client path.
If jobs will be running directly on the client machine or a single-node cluster (for example,
ARC operating in basic mode), the staging area may just be a preferred local scratch area,
and may not need to be a shared path.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 43
RSM Configuration
When using this option, you must ensure that the HPC staging directory is both visible to
and writable by the client machine. For more information, see Enabling OS Copy to the HPC
Staging Directory (p. 58).
No file transfer needed. Client files will already be in an HPC staging directory.
Use this option if the client files are already located in a shared file system that is visible to
all cluster nodes.
When the client and cluster are running on the same platform, or the submit host is local-
host, further action is not required in most cases.
When the client and cluster platforms differ, it is necessary to map the client-visible path to
a cluster-visible path. The most common scenario is a user working on a Windows client, but
their work files are located in a network shared 'home' directory. For example, they work with
their files using \\homeServer\myhome, but on the Linux cluster side, this can be referred
to as $HOME.
The Network share paths on the HPC resource table is displayed if the submit host is a
Linux machine that is not localhost, and SSH is not being used. Use the table to specify
network paths that map to HPC paths:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
44 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
Mapping directory: $HOME (expands to /nfs/homes/joed). This is the path to which the client-
side network share or mount point is mapped. The minimum path is the location shared to
the network (from the Samba configuration, for example), as shown in the example here. The
rest of the path can include subdirectories or they will be inferred from the client path.
Note:
• If jobs will be submitted from Linux clients to a Linux submit host, you may not
need to enter a path in the Network share paths on the HPC resource table
if the client working directory can be used as an HPC staging directory.
In all other cases (for example, SSH is being used), you will be prompted to specify the Staging
directory path on cluster (or nothing at all):
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 45
RSM Configuration
For information on creating client working directories under a shared HPC directory, see
Setting Up Client Working Directories to Eliminate the Need for File Transfers (p. 58).
Use this option when the HPC staging directory is in a remote location that is not visible to client
machines.
RSM uses TCP sockets to stream files from the client machine to the submit host machine. In
this case you must specify the path to the directory where job files will be staged (as the
cluster sees it):
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
46 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
When transferring files to a single-node cluster, it may not be necessary for the staging dir-
ectory to be a shared path (for example, UNC path on Windows).
Use this option when the HPC staging directory is in a remote location that is not visible to client
machines, and you need to use an external mechanism such as SCP for file transfers
For a Linux cluster, you can use either SCP via SSH or a Custom mechanism for file transfers
to the HPC staging directory. For a Windows cluster, only the Custom option is available.
RSM has built-in support for SCP transfer. In using the SCP protocol for communication and
file transfer, it is not necessary to have any RSM components running on the remote submit
host.
If using a Custom mechanism, the RSM launcher service may or may not need to be running
on the submit host. This will depend on whether or not the client needs to communicate
with RSM's launcher service on the remote side to handle the file transfer.
Whether you are using SCP via SSH or a Custom mechanism, you will need to specify the
path to the HPC staging directory as the cluster sees it (for example, /staging on a Linux
machine, or \\server\staging on a Windows machine):
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 47
RSM Configuration
If using a Custom mechanism for file transfers, you will also need to specify the Custom
transfer type. This is a keyword of your choosing that represents the transfer type. It is a
short word or phrase that you will append to the file names of your custom integration files:
If using SCP via SSH, the Custom transfer type is predetermined and cannot be edited.
Finally, specify the Account name that the RSM client will use to access the remote submit
host. For a Linux cluster, if you selected Use SSH or custom communication to the submit
host on the HPC Resource tab, the account name that you specified on that tab will auto-
matically populate the Account field on the File Management tab, and cannot be edited. If
you selected Able to directly submit and monitor HPC jobs on the HPC Resource tab, the
Account field on the File Management tab can be edited. Note, however, that if you are
using a Custom mechanism, the Account name is optional.
If using SCP via SSH, you can customize the SSH-specific cluster integration files to suit your
needs. If using a Custom mechanism, you will need to create custom versions of these files.
For details refer to Configuring SSH/Custom File Transfers (p. 170).
Specify the working directory on the cluster side where job (or solver) commands will start running.
• There is a fast network connection between the cluster nodes and the HPC staging directory.
• You are using a solver that produces fewer, relatively small files as part of the solution and
does not make heavy use of local scratch space (for example, the CFX or the Fluent solver).
This option is recommended to optimize performance when one or both of the following is
true:
• There is a slower network connection between the cluster nodes and the HPC staging dir-
ectory.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
48 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
• You are using a solver that produces numerous, relatively large files as part of the solution
and makes heavy use of local scratch space (for example, Mechanical solvers).
All input files will be copied from the HPC staging directory into that local scratch directory.
Then, when the job finishes running, the requested output files generated by the job will be
copied back to the HPC staging directory.
In the Local HPC scratch directory field, enter the local path of a scratch directory on the
cluster node (for example, C:\Shares\Local_Share\ScratchDir on Windows). You
can enter the path of the scratch directory manually, or use an environment variable in the
format %VAR%.
If the cluster is running on Windows, you must create a network share path for the local
scratch directory on each node. In the Share path for local scratch field, enter the network
share path of the local scratch directory. This path starts with a non-editable [Execution-
Node] variable. When a job is submitted, the [ExecutionNode] variable will be replaced
with the actual machine name of each execution node assigned to the job.
By default, job files will be deleted from the HPC staging directory after the job has run.
Choosing Keep job files in staging directory when job is complete may be useful for
troubleshooting failed jobs. However, retained job files will consume disk space, and require
manual removal.
RSM queues are defined on the Queues tab in the editing pane:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 49
RSM Configuration
RSM provides two ways of defining RSM queues: you can either import a list of HPC queues and
define an RSM queue for each HPC queue, or you can manually add an RSM queue and assign an
HPC queue to it.
• To import a list of HPC queues, or refresh the list if you have imported HPC queues previously,
click . Then, for each HPC queue, double-click in the RSM Queue field and specify a unique
RSM queue name.
• To add an RSM queue to the list, click , then specify a unique name for the queue in the
RSM Queue field.
Double-click in the HPC Queue field and enter the name of an existing HPC queue. RSM will
check to see if the HPC queue is valid.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
50 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
To control whether or not an RSM queue is available for use, select or clear the queue's Enabled
check box.
To test an RSM queue, click in the queue's Test column, or right-click the queue in the tree
in the left pane and select Submit Test.
Note:
• You may need to click Apply on the Queues tab before being able to submit test
jobs.
Job is queued
Job is in progress
Job completed successfully
Job completed successfully and
released
Job aborted
Job aborted and released
Job failed
Job failed and released
When a job is running, the button is replaced by an button, enabling you to abort
the test job if desired.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 51
RSM Configuration
1. Right-click the queue in the tree in the left pane, then select Advanced Test.
2. In the Advanced Test dialog box, select or specify the client directory that you want to use
for the test job. You can leave it set to the default %TEMP% environment variable, or enter a
path or environment variable manually. Manually entered items will be added as drop-down
options.
3. If you want to clean up the client directory after the test job is done, enable the Cleanup Client
Directory check box.
4. Click Submit.
The status of the test is displayed in the Status column of the queue table, as described in Testing
an RSM Queue (p. 51).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
52 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining RSM Configurations
2. Accept or specify the save location, file name, and content to include.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 53
RSM Configuration
3. Click Save.
• Select the queue(s) in the queue list, then click on the queues toolbar.
• Right-click the queue in the queue list, then select Delete Selected RSM Queue(s).
• Right-click the queue in the tree in the left pane, then right-click and select Delete Queue. Note
that only enabled queues appear in the tree.
To delete an RSM configuration, select the configuration in the left pane, then click on the
toolbar, or right-click and select Delete HPC Resource.
In order for users to be able to submit jobs to an HPC resource, they must have access to the RSM
configurations that you have defined. To accomplish this, there are two approaches that you can take:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
54 of ANSYS, Inc. and its subsidiaries and affiliates.
Sharing and Accessing RSM Configurations
• If you changed the RSM configuration directory to a share-friendly folder before creating configurations
(as described in Creating a Shareable RSM Configuration Directory (p. 33)), you can go ahead and
share that folder. Make sure that the folder has read-only permission to prevent others from modifying
your RSM configurations.
• If you did not change the RSM configuration directory before creating configurations, your configur-
ations are located in the default RSM configuration directory (p. 32), which is a user-specific directory
that is not suitable for sharing.
Windows
1. Create a folder in a location that is not associated with a user account (for example,
C:\some\folder).
3. If the RSM service is currently running, stop it. As an administrator, run net stop
RSMLauncherService221.
5. Issue the following command, replacing the path with the desired value:
You can specify a local path if the directory is on the local machine, or a UNC path if the directory
is a network share.
%APPDATA%\ANSYS\v221\RSM
7. Copy the .rsmcc and .rsmq files from the default RSM configuration directory to the new dir-
ectory (for example, C:\some\folder\ANSYS\v221\RSM).
8. Set read-only permission on the folder to prevent others from modifying the configurations.
Linux
1. Create a folder in a location that is not associated with a user account (for example,
/some/folder).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 55
RSM Configuration
You can specify a local path or a mounted file system depending on where the directory resides.
~/.ansys/v221/RSM
On Linux, ~ is the home directory of the account under which RSM is being run.
6. Copy the .rsmcc and .rsmq files to the new directory (for example, /some/folder/AN-
SYS/v221/RSM).
7. Set read-only permission on the folder to prevent others from modifying the configurations.
Once the RSM configuration directory has been shared, Workbench client users should set the RSM
configuration directory on their local machines to the path of the shared RSM configuration directory.
For example, the share path might be something like \\machineName\Share\RSM for Windows
users, or /clusternodemount/share/RSM for Linux users. For details, see Specifying the Location
of the RSM Configuration Directory (p. 147).
Note:
One potential drawback of this method is that users may not be able to access the shared
RSM configurations if the host goes offline or cannot be accessed for some reason (for ex-
ample, if a user is working off-site and does not have access to the network). In this case
RSM will automatically switch the RSM configuration directory back to the default RSM con-
figuration directory (p. 32) on their local machines. This means that users will, at a minimum,
be able to submit jobs to the Ansys RSM Cluster (ARC) already installed on their local machines
using the localhost configuration that is generated in the default RSM configuration dir-
ectory when RSM is installed.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
56 of ANSYS, Inc. and its subsidiaries and affiliates.
Setting Up Job Directories and File Transfers
Method 2: Have users copy RSM configuration files to their local machines
If you are a user looking to access RSM configurations that have been defined by your RSM or system
administrator, you can do so by setting your RSM configuration directory to the shared RSM configuration
directory that was set by the administrator (see Method 1 above). Alternatively you can copy the RSM
configuration database to the appropriate directory on your machine.
1. Obtain the RSM configuration files (.rsmcc and .rsmq files) from the RSM or system administrator.
If the administrator has put the files in a shared directory that you can access, you can retrieve them
directly from there.
2. On your local machine, copy the files into your RSM configuration directory. For information about
the location of this directory, see Setting the RSM Configuration Directory (p. 32).
Note:
If any of the shared files that you are copying have the same name as files in your local RSM
configuration directory, you will need to rename your local files if you do not want them to
be overwritten. For example, you may want to rename your localhost.rsmcc file to
mylocalhost.rsmcc to distinguish it from the remote localhost.rsmcc file, as its
settings may be different.
4. Copy the RSM configurations from the network share to the new directory (for example,
C:\SharedRSMConfig\ANSYS\v221\RSM).
The files in the working directory must be made accessible to the HPC resource. You can accomplish
this in one of two ways:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 57
RSM Configuration
• In the client application, specify that working directories should be created under a shared HPC dir-
ectory.
• Make it possible for files to be transferred from the working directory to a shared HPC directory.
You must also specify the file transfer method that you want to use on the File Management tab of
an RSM configuration (see Specifying File Management Properties (p. 41)).
3.5.1. Setting Up Client Working Directories to Eliminate the Need for File
Transfers
If you set the working directory location to be under a shared HPC directory, file transfers will not be
necessary, because the files will already be in the HPC staging directory. In this case you will be able
to select the No file transfer needed option on the File Management tab of an RSM configuration.
For details about this option, refer to Specifying File Management Properties (p. 41).
The Workbench project directory or Mechanical client scratch directory determines the location of
the client working directory.
The steps for configuring the HPC staging directory for the OS Copy operation are different between
Linux and Windows.
For Windows-to-Windows file transfers, RSM uses the HPC staging network share name specified
on the File Management (p. 41) tab of the RSM configuration to locate and identify the HPC staging
directory. You must configure the directory by performing the following setup tasks:
• Share the HPC staging directory out to the RSM client machine.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
58 of ANSYS, Inc. and its subsidiaries and affiliates.
Setting Up Job Directories and File Transfers
4. In the Advanced Settings dialog box, click Share this Folder and enter the correct name for
the share. For example, if you wanted to create a network share named staging for the HPC
staging directory D:\ClusterStaging on a machine named winclusterhost, you would
enter staging for the share name. This would allow other machines to access the directory
via a UNC path: \\winclusterhost\staging.
5. Ensure that full read-write permissions are defined for the directory.
For Linux-to-Linux file transfers, RSM uses mount points to locate and identify the HPC staging
directory. You must configure the directory by performing the following setup tasks:
1. Ensure that the HPC staging directory belongs to a file system that is mounted, so that it
is visible to the RSM client machine. Use the full path for the directory.
For Windows-to-Linux transfers (using Samba or a similar Linux utility), entries in the Samba config-
uration file map the actual physical location of the Linux HPC staging directory to a predefined
Windows share name that RSM uses to locate and identify the HPC staging directory. The following
example shows how to configure a Samba share on Linux for the HPC staging directory. If you are
unable to create the share, contact your IT System Administrator for assistance with this step.
Edit the smb.conf Samba configuration file to include definitions for the Linux HPC staging dir-
ectory. The example below shows Samba’s default values for the Linux target directories.
[staging]
path = /staging
browseable = yes
writable = yes
create mode = 0664
directory mode = 0775
guest ok = no
The path should point to the actual physical location of the existing target directory.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 59
RSM Configuration
After making your changes to smb.conf, restart the Samba server by running the following
command:
/etc/init.d/smb restart
Note:
The locations of files and method of restarting the Samba service may vary for different
Linux versions.
Verify that the Samba shares are accessible by your Windows machine, indicating that they have
been properly set up. Check this by using Windows Explorer and navigating to \\linuxmachine-
name\staging, using your specific machine name in place of linuxmachinename.
2. Use the RSM Utilities application to specify the correct IP address of the HPC submit host. The
correct IP address is the address seen in the output of a “ping” program from any remote machine
to this machine using the Fully Qualified Domain Name (FQDN). Examples are 1.2.3.4 and ma-
chine.mycompany.com.
3. If using the RSM internal file transfer mechanism (p. 28), which uses TCP sockets to stream files
from the client machine to the submit host, use the RSM Utilities application to specify the correct
IP address in the SocketTransfererListenerIpAddress setting:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
60 of ANSYS, Inc. and its subsidiaries and affiliates.
Setting Up Job Directories and File Transfers
• For Windows: On your Administrative Tools or Administrative Services page, open the Services
dialog box. Restart the services by right-clicking on the service and selecting Restart.
• For Linux: Log into a Linux account with administrative privileges and ensure that Ans.Rsm.*
processes are not running. In a terminal window, run the following command:
[RSMInstall]/RSM/Config/tools/linux/rsmlauncher restart
SSH file transfer mode is actually just referencing an external PuTTY implementation and is not natively
included with RSM, but is included as an option for customers who must use this protocol based on
their specific IT security requirements. This method is also usually slower than the preferred OS File
Copy method, and thus is not recommended unless it is required.
For detailed information about using SSH, see Configuring RSM to Use SSH for Job Submission and/or
File Transfers to a Remote Linux Cluster (p. 121).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 61
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
62 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 4: Ansys RSM Cluster (ARC) Configuration
If you are not using a third-party job scheduler such as Microsoft Windows HPC or LSF, you can use the
Ansys RSM Cluster (ARC) scheduler to submit jobs to a single machine, or a group of machines. The
ARC scheduling system does not offer all of the advanced features that are available in third-party
schedulers, but does include the essential features required to schedule and run jobs that are submitted
to RSM from Ansys products. ARC enables users to get started with RSM right away, and take advantage
of HPC resources.
Every RSM installation has one predefined localhost configuration that uses the ARC scheduling
system to submit jobs to the local machine. This enables users to run certain types of local jobs or
Mechanical background jobs right out of the box, without any special setup. For details see The Default
'Localhost' Configuration (p. 68).
The ARC Configuration application provides a quick and easy way to set up a single-node or multi-node
cluster. It enables you to connect and configure cluster nodes, and start ARC services on those nodes.
Various ARC configuration commands are available to help you customize your cluster setup. For example,
you can configure individual execution nodes to have different resource allocation settings.
The command scripts used to execute ARC cluster commands are located in the
%AWP_ROOT221%\RSM\ARC\tools\winx64 directory on Windows, and the
$AWP_ROOT221/RSM/ARC/tools/linx64 on Linux.
Once you have set up the cluster, you can use the RSM Configuration application to create a cluster
configuration that enables users to submit jobs to the cluster.
Setup instructions and details about ARC command usage and options are provided in the following
sections:
4.1. Important Considerations and Requirements for the Ansys RSM Cluster (ARC)
4.2. Overview of an Ansys RSM Cluster (ARC) Configuration
4.3.The Default 'Localhost' Configuration
4.4. Creating and Deploying an Ansys RSM Cluster (ARC)
4.5. Defining an RSM Configuration for an Ansys RSM Cluster (ARC)
4.6. Ansys RSM Cluster (ARC) Command Usage and Options
4.7. Setting the ARC_ROOT Environment Variable for Ansys RSM Cluster (ARC) Job Submission
4.8. Dealing with a Firewall in a Multi-Node Ansys RSM Cluster (ARC)
4.9. Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 63
Ansys RSM Cluster (ARC) Configuration
• Ansys RSM Cluster (ARC) has basic job scheduling capabilities and provides a simple HPC workflow.
However, it is not intended to be a replacement for a commercial scheduler. If you have integrated
RSM with a third-party scheduler in the past, or require advanced functionality that the ARC system
does not offer, you should continue to work with your IT department to determine which commercial
solution will best fit your simulation and business needs. For information about configuring RSM for
use with an existing third-party scheduler, see RSM Integration with a Cluster or Cloud Portal (p. 121).
• The ARC system is not available as a standalone product. It is installed and intended to be used with
RSM, meaning that jobs can only be submitted to an ARC via RSM from an Ansys client application.
• RSM must be installed on the machine that you are designating as the submit host, as well all machines
that will accept jobs. By default, RSM is automatically installed with Ansys Workbench products when
you use the standard Ansys product installation.
• If creating a multi-node ARC, all execution nodes must be running on the same platform (Windows
or Linux).
• You cannot run two versions of an ARC (for example, 2021 R2 and 2022 R1) at the same time. To
ensure that the correct ARC version is used when jobs are submitted to an ARC via RSM, you should
set the ARC_ROOT environment variable. Refer to Setting the ARC_ROOT Environment Variable for
Ansys RSM Cluster (ARC) Job Submission (p. 103).
In a single-node ARC, the head node (submit host) and execution node are the same machine. This can
be a user's local machine (see The Default 'Localhost' Configuration (p. 68)), or a single remote machine
to which multiple users submit jobs.
A multi-node ARC consists of two or more machines. It is based on the head/compute model of com-
munication in which one machine or process has control over the other machines. In this scenario one
machine serves as the head node (submit host), and the other machines serve as execution nodes. The
submit host may also serve as an execution node if desired.
In order for an ARC to be operational (that is, able to accept, schedule and execute jobs), ARC services
must be running on the cluster nodes.
Important:
When a user submits a job to a single-node ARC, RSM will check to see if ARC services are
running on that node. If the services are not running, RSM will automatically start them as
the user who submitted the job. This is fine for a 'localhost' configuration where a user will
be submitting jobs locally to his or her own machine. However, if multiple users will be
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
64 of ANSYS, Inc. and its subsidiaries and affiliates.
Overview of an Ansys RSM Cluster (ARC) Configuration
The steps that you need to take to set up an ARC depend on whether the cluster will be a single- or
multi-node cluster, and who will be submitting jobs to it. The table below provides an overview of ARC
setup types and the tasks you need to perform for each setup.
You can perform most of these tasks using the automated ARC Configuration application or arcdeploy
command. See Creating and Deploying an Ansys RSM Cluster (ARC) (p. 70).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 65
Ansys RSM Cluster (ARC) Configuration
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
66 of ANSYS, Inc. and its subsidiaries and affiliates.
Overview of an Ansys RSM Cluster (ARC) Configuration
• Create additional
cluster queues (that
contain only certain
machines, for
example). Once
defined, you will be
able to import cluster
queues into an RSM
configuration so that
you can map them to
RSM queues (p. 49).
If using the ARC
Configuration
application or
arcdeploy command
for configuration, see
Creating and
Deploying an Ansys
RSM Cluster
(ARC) (p. 70). Or, to
perform this task
manually, see
Configuring ARC
Queues (arcconfig
queue) (p. 97).
• Configure resource
allocation on
individual nodes. See
Configuring ARC
Cluster Nodes
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 67
Ansys RSM Cluster (ARC) Configuration
Important:
• If the Master service or Node service does not start, consult the ArcMaster221-
<date>.log or ArcNode221-<date>.log file to find the possible cause. For more
information, see Accessing RSM Log Files (p. 199).
• If users will be running a large number of jobs (such as Workbench design point updates)
on a single-node Ansys RSM Cluster (ARC), the single node on which jobs are executed
could become overloaded, resulting in system issues such as memory usage errors. In this
case, it would be advisable to use a multi-node ARC or commercial third-party cluster in-
stead.
• If you have a firewall set up, this may prevent communication between the master node
and execution nodes. To resolve this issue, see Dealing with a Firewall in a Multi-Node
Ansys RSM Cluster (ARC) (p. 103).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
68 of ANSYS, Inc. and its subsidiaries and affiliates.
The Default 'Localhost' Configuration
The localhost configuration automatically makes each user's local machine a single-node cluster. It
provides immediate job submission capability for all users, enabling them to submit certain types of
jobs to their local machines using the Local queue that is defined in this configuration.
In the default localhost configuration, the Name and Submit host are both set to localhost, and
cannot be changed. The HPC type is set to ARC, indicating that the Ansys RSM Cluster (ARC) scheduling
system will be used to submit jobs to the cluster. Since jobs are only being submitted to the local machine
and not a remote one, only the Able to directly submit and monitor HPC jobs option is available.
When the Submit host is the local machine ('localhost'), only two file transfer options are available on
the File Management tab, because the HPC staging directory is on the local machine:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 69
Ansys RSM Cluster (ARC) Configuration
For an overview of the settings in an RSM configuration, refer to Specifying RSM Configuration Set-
tings (p. 35).
The application guides you easily through the following configuration tasks:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
70 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
Tasks such as starting ARC services on cluster nodes are achieved with the click of a button, with no
additional input required.
If the cluster includes multiple nodes, you can configure all execution nodes from the master node, or
run the ARC Configuration application on the execution nodes themselves. However, the master node
can only be configured directly on that machine.
Before you proceed with configuration, refer to Prerequisites for ARC Configuration (p. 71) to ensure
that all nodes are accessible and suitable for ARC configuration.
In this section:
4.4.1. Prerequisites for ARC Configuration
4.4.2. Configuring an Ansys RSM Cluster (ARC) Using the ARC Configuration Application
4.4.3. Configuring an Ansys RSM Cluster (ARC) Using the Command Line (arcdeploy)
• You must have full Administrator and/or sudo rights on all of the nodes in the cluster.
• All of the nodes in the cluster must have the same operating system.
• Ansys software must be installed on the Master node and all Execution nodes. The Ansys installation
directory should be the same on all machines (for example, /apps/ansys_inc or C:\Program
Files\ANSYS Inc).
For additional considerations, see Important Considerations and Requirements for the Ansys RSM
Cluster (ARC) (p. 64).
4.4.2. Configuring an Ansys RSM Cluster (ARC) Using the ARC Configuration
Application
To begin configuration, run the ARC Configuration application on the machine that will serve as the
cluster head node, or master node. You can configure other nodes from the master node while using
the application.
The instructions that follow focus on the deployment of a Windows cluster, but could be applied to
the deployment of a Linux cluster as well.
In this section:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 71
Ansys RSM Cluster (ARC) Configuration
• On Windows, select Start > Ansys 2022 R1 > ARC Configuration 2022 R1.
You can also launch the application manually by double-clicking the following executable:
[RSMInstall]\ARC\bin\arcConfigConsole.exe
<RSMInstall>/ARC/Config/tools/linux/arcconfigui
When launched, the ARC Configuration application will display the hostname of the current
machine:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
72 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
I am the only user running jobs Jobs will be submitted, scheduled and run on the same
on this cluster and they are local machine (the local machine), and only you will be
jobs. submitting jobs to this machine (using the Local queue).
The existing localhost configuration is an example of a
single-node, single-user cluster.
I am configuring this cluster for Jobs will be submitted by one or more users to a specific
either a multi-user or multi-node machine on the network. The machine specified as the
environment. master node (cluster submit host) may also serve as an
execution node, or have additional execution nodes
associated with it. Jobs may run on the master node or on
any of the execution nodes depending on the queue
chosen.
2. In the Master Node Service section, select Yes to specify that the current machine is the master
node, then click Start to install the Master Node Service on this node. When you do so, any exe-
cution nodes and/or queues previously associated with this node are displayed in the tree view.
3. In the Execution Node Service section, specify whether or not the current machine will be used
to run jobs. If you selected Yes, click Start to install the Execution Node Service on this node.
When you do so, the node is added to the Execution Nodes list.
Note that you can also use the ARC Configuration application to stop running ARC services when
desired:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 73
Ansys RSM Cluster (ARC) Configuration
Note:
You cannot define execution nodes unless the Master Node Service has been started on
the master node. This is done on the Cluster Management page.
1. If you specified that the Master node is going to run jobs, go to step 4 to specify execution
properties for this node. Otherwise, to add nodes to the cluster, right-click Execution Nodes in
the tree and select Add an execution node.
2. In the ARC Configuration Manager dialog box, enter the hostname of the machine that you
want to add as an execution node, then click Add.
3. Under Execution Nodes, select the node that you want to define:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
74 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
4. Click Start to install the Execution Node Service on the selected node (if not already started).
6. Click Apply to update the node with the properties that you have set.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 75
Ansys RSM Cluster (ARC) Configuration
Defining Queues
When users want to submit a job to a cluster, they must select a queue for the job. Each queue has
specific nodes associated with it, enabling you to control where and how jobs are run.
The ARC has a built-in local cluster queue for submitting jobs to the local machine, and a default
cluster queue that can submit jobs to any of the execution nodes. You can create additional queues
that submit jobs to specific execution nodes, or have custom properties.
1. To create a cluster queue, right-click Queues in the tree view and select Add a queue.
2. In the ARC Configuration Manager dialog box, enter a name for the queue in the edit box,
then click Add.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
76 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 77
Ansys RSM Cluster (ARC) Configuration
Set Active Time Range? If Set Active Time Range is disabled, jobs can be
submitted to this queue 24 hours a day.
Start Time/End Time
If Set Active Time Range is enabled, you can specify
a time range in which jobs can be submitted to the
queue. To define the range, specify a Start Time and
End Time.
Note:
Limit Concurrent Jobs? If Limit Concurrent Jobs is enabled, you can specify the
maximum number of jobs than can be run at one time from
Max Concurrent Jobs this queue.
4. Click Apply.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
78 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
To specify the execution nodes to which this queue can submit jobs, enable the check boxes of
the desired nodes in the Allowed Nodes table. If you have recently added a new execution node
to the cluster, and do not see it in the list, click Refresh to update the list.
By default, all users who have access permission on the cluster nodes will be able to use this
queue. To restrict queue usage to specific users, enable the Restrict usage to specific users
option, then select all in the table and replace it with the username of the first user to whom
you would like to grant access. Continue adding usernames to the list, one username per row.
6. Click Apply.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 79
Ansys RSM Cluster (ARC) Configuration
2. By default, any user with admin access on the master node, as well as members of the RSM
Admins group, have permission to make cluster configuration changes:
3. To restrict configuration access to specific users or groups, use the Allowed Users and Allowed
Groups tables to specify the desired user accounts and group names.
4. Click Apply.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
80 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
2. Enter the password that you want RSM to use when the cluster is accessed with the current user
account, then click Confirm:
Note:
If you want RSM to use a different account to log in to the cluster, you will need to
cache the password for that account using the arccredentials command instead. For
more information, see Caching Credentials for Cluster Job Submission (arccreden-
tials) (p. 100)
To set up an automatic refresh that occurs at a regular interval, select View > Auto Refresh (30s in-
terval).
4.4.3. Configuring an Ansys RSM Cluster (ARC) Using the Command Line
(arcdeploy)
The arcdeploy command provides the same functionality as the ARC Configuration application, but
delivers it via the command line for those who require or prefer this type of deployment.
Before using this command, refer to Prerequisites for ARC Configuration (p. 71) to ensure that all
nodes are accessible and suitable for ARC configuration.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 81
Ansys RSM Cluster (ARC) Configuration
The information that you enter is captured in a CSV file so that you can quickly and easily deploy the
same cluster again in the future.
%AWP_ROOT221%\RSM\ARC\tools\winx64\arcdeploy.cmd
2. Type y (for yes) to confirm that you have full Administrator and/or sudo rights on all cluster
nodes.
3. Type y to confirm that Ansys software is uniformly installed on the Master node and all Execution
nodes.
4. Type y to confirm that all nodes in the cluster are based on the same Operating System.
5. Type y to confirm that the current machine will serve as the cluster's Master node.
The path shown is the detected directory of the deployment application. This directory must
reside within the Ansys installation directory.
6. To record the name shown as the Master hostname, type y. Otherwise, to record a different
Master hostname type n (for no), then type the desired name. The specified hostname must exist
in DNS or other hostname resolution system. The recorded name is the one that will appear in
the command's output when performing upcoming steps.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
82 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
You are then prompted to enter the hostnames of the Execution nodes:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 83
Ansys RSM Cluster (ARC) Configuration
7. Type the hostname of an Execution node, then press Enter. Repeat this step for each remaining
Execution node (including the Master node if it will be used to execute jobs).
8. When you have completed the list of Execution nodes, press CtrlX and then press Enter.
9. Review the list for accuracy, then type y to confirm that it is correct.
The ARC will have a built-in local cluster queue for submitting jobs to the local machine, and
a default cluster queue that can submit jobs to any of the execution nodes. You can create
more queues in the next step.
a. Type the desired name for the new queue, then press Enter.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
84 of ANSYS, Inc. and its subsidiaries and affiliates.
Creating and Deploying an Ansys RSM Cluster (ARC)
All Execution nodes are added to the queue. When prompted, specify whether or not you
want to add the Master node to the queue as well, by typing either y or n:
The queue will be established, but will not have any nodes assigned to it. You will need
to do this using different commands after the cluster has been deployed. See Modifying
a Cluster Queue (p. 98).
11. When you have finished adding queues, type n and press Enter at the Do you want to add
another queue? (Y/N) prompt.
Each queue will be created and enabled, and the appropriate nodes will be assigned to each
queue according to the responses that you entered:
12. To save all of the information that you have entered to the default CSV file name, arcde-
ploy.config.csv, simply press Enter. Otherwise, type the desired file name and press Enter.
You can use this file to instantly deploy the same cluster in the future if needed. To do this, you
will use the -f argument when issuing the arcdeploy command.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 85
Ansys RSM Cluster (ARC) Configuration
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
86 of ANSYS, Inc. and its subsidiaries and affiliates.
Defining an RSM Configuration for an Ansys RSM Cluster (ARC)
The submit host is the machine that is responsible for submitting jobs to the cluster, and on which the
Master service is installed. You have two options when specifying the Submit host value:
• Specifying 'localhost' in the Submit host field indicates that the RSM client and the ARC submit host
are the same machine. This means that jobs will be submitted, scheduled and run on the same machine
(a user's local machine). The RSM configuration will exhibit the same behavior as the default 'localhost'
configuration described in The Default 'Localhost' Configuration (p. 68).
• Specifying a hostname or IP address in the Submit host field designates a particular machine in your
network as the ARC submit host. This could be the current machine (on which you are creating the
RSM configuration), or a remote one. Even if the current machine is the submit host, you must specify
the hostname and OS of this machine if other users will be submitting jobs to this machine:
For an overview of the settings in a cluster configuration, refer to Defining RSM Configurations (p. 33).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 87
Ansys RSM Cluster (ARC) Configuration
Many of the commands have options that enable you to modify the operation of the command.
To view the available options for a command, simply type the command name with no command to
run. For example:
C:\Program Files\ANSYS Inc\v221\RSM\ARC\tools\winx64>arcconfig node
Usage:
-? or -h: Display usage
In order to add a new execution node, simply start the execution node services on that host.
Basic job-related commands like arcsubmit and arcstatus are used in both single- and multi-node ARC
setups. Additional commands are available for administrators who would like to further customize an
ARC configuration.
The following sections provide details about how each command is used, and the options available for
command execution:
4.6.1. Installing ARC Cluster Services on Windows (installservice)
4.6.2. Uninstalling ARC Cluster Services on Windows (uninstallservice)
4.6.3. Installing ARC Cluster Services on Linux
4.6.4. Uninstalling ARC Cluster Daemon Services on Linux (uninstall_daemon)
4.6.5. Commands for ARC Job Management
4.6.6. Configuring ARC Cluster Nodes (arcconfig node modify)
4.6.7. Displaying Resource Availability on ARC Nodes (arcnodes)
4.6.8. Configuring ARC Queues (arcconfig queue)
4.6.9. Displaying the Status and Details of ARC Queues (arcqueues)
4.6.10. Caching Credentials for Cluster Job Submission (arccredentials)
4.6.11. Migrating an ARC Setup from a Previous Version (arcconfig migration)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
88 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
submit host, and install the Node service on every node that will be used for job execution (including
the submit host if it will be used for running jobs).
For Windows, use the installservice command and options to complete the following tasks:
4.6.1.1. Installing the ARC Master Service on a Windows Head Node
4.6.1.2. Installing the ARC Node Service on Windows Execution Nodes
%AWP_ROOT221%\RSM\ARC\tools\winx64\installservice -arcmaster
%AWP_ROOT221%\RSM\ARC\tools\winx64\installservice -arcnode
To uninstall the Master service on the head node, run the following command on that node:
%AWP_ROOT221%\RSM\ARC\tools\winx64\uninstallservice -arcmaster
To uninstall the Node service on an execution node, run the following command on that node:
%AWP_ROOT221%\RSM\ARC\tools\winx64\uninstallservice -arcnode
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 89
Ansys RSM Cluster (ARC) Configuration
When installing an ARC cluster service on Linux, you must determine if you want to install the service
as a daemon that will start the service automatically when the machine is booted, or if you want to
start the service manually via a startup script. Use only one of these methods.
When an ARC cluster service is started manually, it runs as a process for the user who initiated the
service. A manually started ARC cluster service is stopped each time the machine is rebooted. After
a reboot you must restart the ARC cluster service by running the startup script.
In a multi-node ARC on Linux, it is recommended that you install cluster services as daemons.
# The following examples show loading environment settings specific to ARC Advanced mode.
# When defined, ARC services will inherit the environment created here
# . /home/batch/environmentVariables/conf/settings.sh
Note:
• This profile only works on Linux. Windows users should modify their environment via
the environment interface in Windows.
• This profile will work for all Ansys RSM Cluster (ARC) jobs, but the shell is dependent
on what is chosen.
– For a basic ARC configuration in which no special setup has been performed, this
file must be written in /bin/sh. This applies, for example, when a user submits a
job to a single-node ARC, and RSM auto-starts ARC services on the node if they are
not running.
– For an ARC configuration where ARC services have been configured as daemon
services, arc_env_profile should be written in whatever shell is chosen in the
LinuxShellToUse setting in the Ans.Rsm.AppSettings.config file.
• Ensure that you set this profile carefully. Setting it incorrectly could prevent RSM and
ARC from working properly.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
90 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
4.6.3.3. Starting ARC Cluster Services Automatically at Boot Time for Linux (in-
stall_daemon)
You can configure an ARC cluster service to start automatically when the machine is booted by
configuring it as a “daemon” service (if the service is not configured to start automatically, then it
must be started manually, as described in Starting ARC Cluster Services Manually on Linux (arcmaster
| arcnode) (p. 90)). Daemon services are scripts or programs that run persistently in the background
of the machine, and which are usually executed at startup by the defined runlevel.
Once the daemon service is installed, the cluster service will be started automatically without re-
booting. The next time the machine is rebooted, the installed cluster service will be started auto-
matically.
$AWP_ROOT221/RSM/ARC/tools/linx64/install_daemon -arcmaster
$AWP_ROOT221/RSM/ARC/tools/linx64/install_daemon -arcnode
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 91
Ansys RSM Cluster (ARC) Configuration
To uninstall the Master daemon service on the head node, run the following command on that node:
$AWP_ROOT221/RSM/ARC/tools/linx64/uninstall_daemon -arcmaster
To uninstall the Node daemon service on an execution node, run the following command on that
node:
$AWP_ROOT221/RSM/ARC/tools/linx64/uninstall_daemon -arcnode
Note:
This is for reference or debugging purposes only. We do not recommend using these
commands directly for cluster job submission. RSM is designed to integrate with Ansys
RSM Cluster setups, eliminating the need to manually issue job submission commands.
Note:
The arcsubmit command cannot be issued independently. It only works when issued
via RSM.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
92 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 93
Ansys RSM Cluster (ARC) Configuration
arcstatus [jobId]
The following example shows the types of information that the arcstatus command retrieves:
C:\Program Files\ANSYS Inc\v221\RSM\ARC\tools\winx64>arcstatus
JobNumber Status UserName Cores Submit Time Queue Assigned Machines Info
==========================================================================================================
arckill [jobId]
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
94 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
The following are the key uses for this command when configuring ARC execution nodes:
4.6.6.1. Associating ARC Execution Nodes with the Master Node
4.6.6.2. Setting the Maximum Number of Cores to be Used on an Execution Node
4.6.6.3. Setting the Maximum Resource Allocation on an Execution Node
Important:
Once ARC cluster services have been started on cluster nodes, you do not have to go to
each execution node to configure it. You can configure any execution node from the
master node or any other node in the cluster as long as the firewall exemption has been
set up correctly.
For each execution node, run the following command in the ARC command directory (p. 88), replacing
<execHostName> and <masterHostName> with the machine name of the execution node and
master node:
Note:
This command can also be issued for a single-node ARC to limit the number of cores to
be used on the single cluster node. Individual users who use the localhost configur-
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 95
Ansys RSM Cluster (ARC) Configuration
ation can run the command on their own machines if they do not want to use all of the
cores on their machines to run jobs.
For example, to set the maximum number of cores to 2 on MACHINE2, specify arcconfig node
modify machine2 -c 2.
Exec Node Name Associated Master Port Max Cores Max Memory Max Disk
===============================================================================
* Indicates that resources have not been set up. Any resource request will be accepted.
Exec Node Name Associated Master Port Max Cores Max Memory Max Disk
===============================================================================
* Indicates that resources have not been set up. Any resource request will be accepted.
To use all cores that you have, specify arcconfig node modify <execHostName> -c
all.
Note the change in the Max Memory and Max Disk values in the example below:
C:\Program Files\ANSYS Inc\v221\RSM\ARC\tools\winx64>arcconfig node modify machine2 -r d:30tb,m:40mb
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
96 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
Exec Node Name Associated Master Port Max Cores Max Memory Max Disk
===================================================================================
* Indicates that resources have not been set up. Any resource request will be accepted.
Exec Node Name Associated Master Port Max Cores Max Memory Max Disk
===================================================================================
* Indicates that resources have not been set up. Any resource request will be accepted.
To view resource information for a specific node, append the node's machine name to the command:
arcnodes <nodeName>
If a node name is not provided, the command will list information for all execution nodes, as shown
below:
* Indicates that resources have not been set up. Any resource request will be accepted.
The available cores, memory and disk space will vary depending on how much is currently being
consumed by jobs.
Every multi-node ARC setup has a default cluster queue that can submit jobs to any machine(s)
in the cluster, and a local cluster queue for submitting jobs to the local machine. If you would like
certain types of jobs to be targeted to specific machines, you can create additional cluster queues to
address this desired behavior.
You can use the arcconfig queue command in the ARC command directory (p. 88) to add, remove
or modify ARC cluster queues.
In this section:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 97
Ansys RSM Cluster (ARC) Configuration
Once you have added a queue you can assign machines to it, enable it, and specify other custom
settings. For options see Modifying a Cluster Queue (p. 98).
Note:
For example, to enable a queue that you have created, run the following command: arcconfig
queue modify <queueName> -e true.
Note:
Modifying the built-in default and local cluster queues is not recommended. Even
if you were to modify the default queue, it would continue to automatically add new
machines.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
98 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
Name Status Priority Start Time End Time Max Jobs....Allowed Machines Allowed Users
=========================================================================================================
* Indicates that resources have not been set up. Any resource request will be accepted.
If a queue has been enabled, its Status will be either Active or Closed depending on whether or not
the current time falls within the queue's Start Time/End Time range.
If you did not enable a queue after creating it, or have disabled a queue, its Status will be Disabled
in the queue list. For information enabling queues, see Modifying a Cluster Queue (p. 98).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 99
Ansys RSM Cluster (ARC) Configuration
Password:
To change the account used to log in to execution nodes, run the following command:
arccredentials -a Username
This command works the same way as the general RSM migration utility (see Migrating RSM from a
Previous Version (p. 150)), but includes functionality to maintain your configuration of ARC cluster
nodes. This is done by migrating each node's settings, such as the cores and memory allocated for
job execution.
1. On the master node, log into an account with administrative privileges (in other words, as a
member of the rsmadmins group, or as root).
2. If you have not already done so, install the new product version.
3. If you have not already done so, start the ARC services of the new version.
You will need to install the Master service on the master node, and the Node service on each
node that will be used for job execution.
Note that ARC services of the previous version do not need to be running.
4. For a multi-node ARC, associate each execution node with the master node, as described in As-
sociating ARC Execution Nodes with the Master Node (p. 95).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
100 of ANSYS, Inc. and its subsidiaries and affiliates.
Ansys RSM Cluster (ARC) Command Usage and Options
Windows: %AWP_ROOT221%\RSM\ARC\tools\winx64
Linux: $AWP_ROOT221%/RSM/ARC/tools/linx64
6. Run the following command, appending the desired operator and options from the accompanying
tables:
Operator Usage
config Migrate ARC cluster databases, such as ARC
configurations and queues.
settings Migrate ARC-specific settings in the RSM\Con-
fig\Ans.Rsm.AppSettings.config file.
(To see which settings will be migrated, refer to
the new version's RSM\Config\migra-
tion\Arc.AppSettings.Migrate.con-
fig file.)
all Migrate everything (cluster configurations,
queues, and settings).
Option Usage
-v123 (Required) Specify the version that you are
migrating, so that the migration command
knows which files to look for. Replace the 123
with the version that you are migrating (for
example, enter -v212 for version 2021 R2).
Note:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 101
Ansys RSM Cluster (ARC) Configuration
Example
In the following example we are using the arcconfig migration command on the master node to
migrate ARC queues and master node settings previously defined in version 2021 R2 to a new, version
2022 R1 installation.
C:\Program Files\ANSYS Inc\v221\RSM\ARC\tools\winx64>arcconfig migration all -v212 -preview -verbose
Migrating
C:\ProgramData\Ansys\v221\ARC\..\..\v212\ARC\NodeCommunicationConfig.xml
to
C:\ProgramData\Ansys\v221\ARC\ARCMASTER_NodeCommunicationConfig.xml
Skip migrating queue: default - the built-in queue cannot be migrated.
Skip migrating queue: local - the built-in queue cannot be migrated.
Queue successfully merged: highmem
Queue successfully merged: night_only
Queue successfully merged: test
C:\Program Files\ANSYS Inc\v221\RSM\ARC\tools\winx64>
Note:
If a queue with the same name already exists in the new setup, it will not be migrated.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
102 of ANSYS, Inc. and its subsidiaries and affiliates.
Dealing with a Firewall in a Multi-Node Ansys RSM Cluster (ARC)
4.7. Setting the ARC_ROOT Environment Variable for Ansys RSM Cluster
(ARC) Job Submission
Important:
Although different versions of RSM can be installed side by side, RSM allows only one version
of ARC to be used on each node at one time. You cannot have two versions of an ARC (for
example, 2021 R2 and 2022 R1) running at the same time. This ensures that resources such
as cores, memory and disk space can be properly allocated on each node.
When multiple versions of RSM are running, it is recommended that you set the ARC_ROOT environment
variable on the ARC master node to ensure that the correct version of ARC is used when jobs are sub-
mitted to that machine.
The variable should point to the following directory, where xxx is the version that you want to use (for
example, 221):
Windows: %AWP_ROOTxxx%\RSM\ARC
Linux: $AWP_ROOTxxx/RSM/ARC
If you do not specify the ARC_ROOT variable, RSM will attempt to use the ARC from the current install-
ation.
CommandCommunicationPort: The port on the master and execution nodes that allows incoming
commands such as arcsubmit and arcstatus to be read. By default, port 11221 is used.
MasterCommunicationPort: The port on the master node that allows incoming traffic from
execution nodes. By default, port 12221 is used.
NodeCommunicationPort: The port on the execution node that allows incoming traffic from
the master node. By default, port 13221 is used.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 103
Ansys RSM Cluster (ARC) Configuration
For example, to set the value of the node communication port to 14221 on Windows, you would enter
the following:
Important:
• Port settings must be specified on the master node and each execution node. If you are
not using a network installation of RSM, this means that you will need to run the RSM
Utilities application (in other words modify the Ans.Rsm.AppSettings.config file)
on each node in the cluster.
• When specifying the three ports, make sure that each port is different, and is not being
used by any other service (such as the RSM launcher service).
Note:
Multi-node ARC configuration requires system administrator or root permission and should
only be performed by an IT administrator.
Scenario
Cluster Nodes
There are 4 machines available for scheduling and running jobs. Their names and roles are described
below.
• ARCMASTER: This is the machine to which jobs will be submitted from users' client machines for
scheduling. In other words, it is the cluster submit host, or master node.
At a minimum this machine has Workbench and RSM installed, as well as the RSM launcher service
(see Installing and Configuring the RSM Launcher Service for Windows (p. 21)).
• EXECHOST1, EXECHOST2 and EXECHOST3: These are high-capacity machines on which jobs will
run. In other words, they are execution nodes. They have Workbench, RSM, and Ansys solvers installed.
On EXECHOST1 and EXECHOST2 we will restrict the number of cores that can be used by ARC jobs
on those machines, and place no restrictions on EXECHOST3 so that it can handle larger jobs.
We will install the Node service on each of these machines, and associate them with the ARCMASTER
machine to essentially create a cluster.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
104 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
Note that if we wanted to use ARCMASTER to run jobs as well, we would simply need to install the
Node service on that machine. For this example, we will assume that only EXECHOST1, EXECHOST2
and EXECHOST3 will be used for running jobs.
Note:
All nodes in an Ansys RSM Cluster must run on the same platform (Windows or Linux). In-
structions for both platforms are provided in this example.
Cluster Queues
The ARC will already have a local cluster queue for submitting jobs to the local machine, and a de-
fault cluster queue that can submit jobs to any of the execution nodes.
We are going to create a custom cluster queue named high_mem that will be dedicated to running
jobs on EXECHOST3 only, which is the execution node with unrestricted resource allocation. We will
also set the maximum number of jobs that can be run on this queue to 100.
RSM Configuration
Once we have set up the ARC cluster, we will use the RSM Configuration application to create a config-
uration named ARC. We will import the ARC cluster queues (local, local and high_mem) into the
configuration and create RSM queues that map to these cluster queues.
Finally, we will make the ARC configuration available to users so that the RSM queues defined in the
configuration appear in client applications on their machines, enabling them to submit jobs to the RSM
queues (which map to the ARC cluster queues). Jobs will be sent to ARCMASTER, where the Master
service will dispatch jobs to the execution nodes.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 105
Ansys RSM Cluster (ARC) Configuration
• On Windows, select Start > Ansys 2022 R1 > ARC Configuration 2022 R1.You can also
launch the application manually by double-clicking the following executable:
[RSMInstall]\ARC\bin\arcConfigConsole.exe
<RSMInstall>/ARC/Config/tools/linux/arcconfigui
3. On the Local Service Management page, set the Cluster Usage and Service Management options
as follows:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
106 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
Remember that we are not using the master node for job execution (but could if we wanted to).
4. Click Start to start the Master Node Service on the current machine.
5. Once the Master Node Service is started, right-click Execution Nodes in the tree and select Add
an execution node.
6. In the ARC Configuration Manager dialog box, type exechost1, then click Add:
A connection is made to the EXECHOST1 machine, and it is added to the node list.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 107
Ansys RSM Cluster (ARC) Configuration
9. In the properties panel, set Max Cores to 4 to limit the number of cores that can be used by cluster
jobs on EXECHOST1.
11. In the tree, right-click Execution Nodes and select Add an execution node.
12. In the ARC Configuration Manager dialog box, type exechost2, then click Add:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
108 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
17. In the tree, right-click Execution Nodes and select Add an execution node.
18. In the ARC Configuration Manager dialog box, type exechost3, then click Add:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 109
Ansys RSM Cluster (ARC) Configuration
21. In the properties panel, delete the value in the Max Cores so that there are no restrictions on the
number of cores that cluster jobs can use on EXECHOST3.
Now that the cluster nodes have been established, we can create cluster queues. We will create a
queue named high-mem. This queue will run jobs on EXECHOST3 only. This is the only machine
with unrestricted resource allocation, making it ideal for larger jobs. We will also set the maximum
number of jobs that can run on the high_mem queue to 100.
24. In the ARC Configuration Manager dialog box, type high-mem, then click Add:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
110 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
26. In the properties panel, enable the queue and set the Max Concurrent Jobs to 100:
28. In the tree, select the high-mem queue’s Access Control setting. Then, in the Node Assignments
area, enable the exechost3 check box, and ensure that the other nodes are not checked:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 111
Ansys RSM Cluster (ARC) Configuration
This specifies that jobs submitted to this queue will only run on EXECHOST3.
30. In the tree, right-click Administrative Access and select Cache Password.
31. Enter the password that you want RSM to use when the cluster is accessed with the current user
account, then click Confirm:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
112 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
Note:
If we wanted RSM to use a different account to log in to the cluster, we would run the
arccredentials command to cache the credentials for that account. For more information,
see Caching Credentials for Cluster Job Submission (arccredentials) (p. 100).
At this point the Ansys RSM Cluster is set up and ready to receive job submissions from client machines.
1. Launch the RSM Configuration application by selecting Start > Ansys 2022 R1 > RSM Configuration
2022 R1.
2. Click , or right-click in the HPC Resources list and select Add HPC Resource.
3. On the HPC Resource tab, specify a name for the configuration, select ARC for the HPC type, and
specify the machine name of the cluster submit host.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 113
Ansys RSM Cluster (ARC) Configuration
4. Click Apply, then select the File Management tab. Referring to Specifying File Management
Properties (p. 41), select the desired file transfer method, then specify the job execution working
directory.
6. Click to import the cluster queues that are defined on the ARCMASTER machine (default,
local and high_mem).
7. Since we are accessing ARCMASTER for the first time, we may be asked to enter the credentials
that RSM will use to access that machine:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
114 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
8. Enter a User Name and Password that can be used to access ARCMASTER, then click OK. The
specified credentials will be cached, enabling any user using this configuration to submit jobs to
ARCMASTER.
9. For each ARC Queue in the list, you can specify a unique RSM Queue name if you want (by default,
the RSM queue name matches the cluster queue name). RSM queues are what users see in client
applications when they choose to submit jobs to RSM. You can also choose which queues you
want to enable for users, and submit a test job to each RSM queue by clicking Submit in the Test
column.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 115
Ansys RSM Cluster (ARC) Configuration
With either of these options, if users were to launch the RSM Configuration application on their ma-
chines, they would see the ARC configuration automatically added to their HPC Resources list. They
could then start submitting jobs to the RSM queues that were defined in this configuration. RSM queues
are linked to the ARC configuration defined in RSM, which enables jobs to be submitted to ARCMASTER
for scheduling.
There are two options for making the ARC.rsmcc and queues.rsmq files available to users:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
116 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
This method ensures that all users have the most accurate and up-to-date configuration information,
as files are centrally stored and managed.
• If you changed the RSM configuration directory to a share-friendly folder before creating configurations
(as described in Creating a Shareable RSM Configuration Directory (p. 33)), you can go ahead and
share that folder. Make sure that the folder has read-only permission to prevent others from modifying
your configurations.
• If you did not change the RSM configuration directory before creating cluster configurations, your
configurations are located in the default configuration directory (p. 32), which is a user-specific dir-
ectory that is not suitable for sharing.
Windows
1. Create a folder in a location that is not associated with a user account (for example,
C:\some\folder).
3. If the RSM service is currently running, stop it. As an administrator, run net stop
RSMLauncherService221.
5. Issue the following command, replacing the path with the desired value:
You can specify a local path if the directory is on the local machine, or a UNC path if the directory
is a network share.
%APPDATA%\ANSYS\v221\RSM
7. Copy the ARC.rsmcc and queues.rsmq files to the new RSM configuration directory (for ex-
ample, C:\some\folder\ANSYS\v221\RSM).
Linux
1. Create a folder in a location that is not associated with a user account (for example,
/some/folder).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 117
Ansys RSM Cluster (ARC) Configuration
You can specify a local path or a mounted file system depending on where the directory resides.
~/.ansys/v221/RSM
On Linux, ~ is the home directory of the account under which RSM is being run.
6. Copy the ARC.rsmcc and queues.rsmq files to new configuration directory (for example,
/some/folder/ANSYS/v221/RSM).
Once the configuration directory has been shared, users should set the configuration directory on their
local machines to the path of the shared configuration directory. For example, the share path might be
something like \\machineName\Share\RSM for Windows users, or /clusternode-
mount/share/RSM for Linux users. They will follow the steps in Specifying the Location of the RSM
Configuration Directory (p. 147).
Note:
One potential drawback of this method is that users may not be able to access to the shared
configurations if the host goes offline or cannot be accessed for some reason (for example,
if a user is working off-site and does not have access to the network). In this case RSM will
automatically switch the configuration directory back to the default configuration directory
on their local machines. This means that users will, at a minimum, be able to submit jobs to
ARC clusters already installed on their local machines using the localhost configuration
that is generated in the default configuration directory when RSM is installed.
If you are a user looking to access configurations that have been defined by your RSM or system admin-
istrator, you can do so by setting your configuration directory to the shared configuration directory
that was set by the administrator (see Option 1 above). Alternatively you can copy the configuration
database to the appropriate directory on your machine.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
118 of ANSYS, Inc. and its subsidiaries and affiliates.
Example: Setting Up a Multi-Node Ansys RSM Cluster (ARC)
1. Obtain the ARC.rsmcc and queues.rsmq files from the RSM or system administrator. If the ad-
ministrator has put the files in a shared directory that you can access, you can retrieve them directly
from there.
2. On your local machine, copy the files into your default configuration directory.
By default, the directory in which the configurations are stored resolves to the following location:
Windows: %APPDATA%\ANSYS\v221\RSM
Linux: ~/.ansys/v221/RSM
On Linux, ~ is the home directory of the account under which RSM is being run.
Note:
If any of the shared files that you are copying have the same name as files in your local
configuration directory, you will need to rename your local files if you do not want them to
be overwritten. For example, you may want to rename your localhost.rsmcc file to
mylocalhost.rsmcc to distinguish it from the remote resource's localhost.rsmcc
file, as its settings may be different.
4. Copy the configurations from the network share to your new configuration directory (for
example, C:\SharedRSMConfig\ANSYS\v221\RSM).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 119
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
120 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 5: RSM Integration with a Cluster or Cloud
Portal
When you want RSM to submit jobs to an HPC resource, whether it be an Ansys RSM Cluster (ARC), a
third-party cluster such as Microsoft Windows HPC or LSF, or a Cloud portal, you need to create a con-
figuration in RSM, as described in Defining RSM Configurations (p. 33). Additional steps may be required
to ensure successful cluster job submission.
Assumptions
• The cluster with which you want to integrate RSM has already been established and properly con-
figured. For information on setting up an Ansys RSM Cluster (ARC), refer to Ansys RSM Cluster (ARC)
Configuration (p. 63). The establishment of commercial clusters is beyond the scope of this user's
guide. For those steps, consult the documentation for the third-party scheduler you are using.
• You know the machine name of the cluster submit host (the node that performs job scheduling).
• If you are using a UGE (SGE) cluster, parallel environments have already been defined by your cluster
administrator.
• You are able to install and run Ansys, Inc. products, including Licensing, on the cluster nodes. For
information on product and licensing installations, see the Installation and Licensing Documentation
on the Ansys Help site.
• RSM has been installed on the cluster submit host. See RSM Software Installation (p. 19).
• The RSM launcher service is installed and running on the cluster submit host if it will be accepting
submissions from remote RSM clients. See Installing and Configuring the RSM Launcher Service (p. 20).
This chapter describes the additional steps you may need to take when integrating RSM with an estab-
lished cluster.
5.1. Configuring RSM to Use SSH for Job Submission and/or File Transfers to a Remote Linux Cluster
5.2. Integrating RSM with a Microsoft HPC or Windows-Based Cluster
5.3. Integrating RSM with a Cloud Portal
5.1. Configuring RSM to Use SSH for Job Submission and/or File Transfers
to a Remote Linux Cluster
SSH/SCP (Secure Shell/Secure Copy) can be used to establish communication between a Windows RSM
client machine and a Linux-based ARC, LSF, PBS Pro, SLURM, or UGE (SGE) cluster. The SSH application
is used instead of RSM to execute cluster commands, monitor jobs, and copy data to/from the Linux
cluster node.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 121
RSM Integration with a Cluster or Cloud Portal
RSM supports using SSH/SCP in custom job scripts. The built-in job scripts for the RSM job submissions
have been tested using the PuTTY SSH client (http://www.chiark.greenend.org.uk/~sgtatham/putty).
• RSM has been installed on both the Windows and Linux machines.
• PS, AWK, GREP, LS, and the ANSYS221 command must exist on the Linux machine.
• You are able to install and run Ansys, Inc. products, including Licensing, on both Windows and Linux
systems. For information on product and licensing installations, see the Installation and Licensing
documentation on the Ansys Help site.
1. Define a configuration in RSM which indicates that SSH will be used. See Defining a Configuration
for a Remote Linux Cluster (SSH) (p. 122).
2. Install and configure an SSH client (PuTTY SSH) on RSM client machines. See Configuring PuTTY
SSH (p. 123).
Note:
• Ansys recommends that you use SSH only if your IT policy requires it. The communication
overhead that SSH imposes results in slower processing times when launching solutions
and retrieving results.
• RSM handles job submission and file transfers independently. This means that you have
the flexibility to choose whether you want to use SSH for either or both of these functions.
See Defining a Configuration for a Remote Linux Cluster (SSH) (p. 122).
• To use SSH for job submission (communication with the submit host):
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
122 of ANSYS, Inc. and its subsidiaries and affiliates.
Configuring RSM to Use SSH for Job Submission and/or File Transfers to a Remote
Linux Cluster
On the HPC Resource tab, select Use SSH or custom communication to the submit host. Then,
in the Account name field, specify the account name that the Windows client will use to submit
jobs to the Linux submit host. For example:
Note:
The Account name specified is the SSH account and not the account that will be cached
by RSM. You will need to set up passwordless SSH access to the submit host. See Con-
figuring PuTTY SSH (p. 123).
On the File Management tab, select External mechanism for file transfer (SCP, Custom), and
select SCP via SSH as the Transfer mechanism. For information about this and all available file
transfer methods, see Specifying File Management Properties (p. 41).
Note:
SSH configuration must be completed by your IT administrator. This section provides in-
structions for a PuTTY SSH implementation. Other SSH implementations are possible, and
your IT administrator can determine which one is best for your site.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 123
RSM Integration with a Cluster or Cloud Portal
2. Change the Key comment to include your machine name and Windows username.
If you use a pass phrase, jobs will hang a prompt for you to enter the pass phrase. Be sure to secure
the private key file using some other means. For example, if only you will be using the key, save
it to a location where only you and administrators have access to the file, such as your My Docu-
ments folder. If multiple users share the same key, allow the owner full control, then create a
group and give only users in that group access to this file.
5. If your Linux cluster uses OpenSSH, convert the key to OpenSSH format by selecting Conversions
> Export Open SSH key in the PuTTY Key Generator dialog box.
6. Move the public portion of the key to the Linux machine. This requires you to edit the
~/.ssh/authorized_keys file on the Linux machine as follows:
a. Open an SSH session to one of your cluster nodes, cd into ~/.ssh, and open the author-
ized_keys file in your favorite editor (for example, vi or Emacs).
b. Copy all the text from the box under Public key for pasting and paste it into ~/.ssh/au-
thorized_keys. All of this text should be one line.
c. If the authorized_keys file does not exist, create one. Alternatively, paste it into a text file
and move that file to the Linux machine for editing.
2. On the Advanced tab, select Environment Variables. The Environment Variables dialog box
appears.
3. In the Environment Variables dialog box, locate the Path variable in the System variables pane.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
124 of ANSYS, Inc. and its subsidiaries and affiliates.
Configuring RSM to Use SSH for Job Submission and/or File Transfers to a Remote
Linux Cluster
4. Select the Path variable and then click the Edit button. The Edit System Variable dialog box
appears.
5. Add the PuTTY install directory to the Variable value field (for example, C:\Program
Files\putty) and then click OK.
6. In the System variables pane, click the New button. The New System Variable dialog box appears.
7. In the New System Variable dialog, create a new environment variable named KEYPATH with a
value containing the full path to the private key file (for example, <drive>:\Program
Files\Putty\id_rsa.ppk).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 125
RSM Integration with a Cluster or Cloud Portal
Use a user variable if the key file is used only by you. Use a system variable if other users are
sharing the key file. For example, if a Windows 7 user has a key file in My Documents, the variable
value should be %USERPROFILE%\My Documents\id_rsa.ppk (this expands to
<drive>:\Documents and Settings\<user>\My Documents\id_rsa.ppk).
8. Click OK.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
126 of ANSYS, Inc. and its subsidiaries and affiliates.
Integrating RSM with a Microsoft HPC or Windows-Based Cluster
Note:
• ~ (tilde) representation of the home directory is not supported when specifying paths
in RSM (for example, the path of the HPC staging directory when defining a configuration).
• Different shells use different initialization files than the account's home directory and
may have a different syntax than shown above. Refer to the Linux man page for the
specific shell or consult the machine administrator.
Following are additional considerations you may need to make when integrating with a Microsoft HPC
or Windows-based cluster.
Prerequisites Installation
The Ansys product installation requires the installation of several prerequisites. The installer will check
for these prerequisites on the machine where the installer is launched (for example, the head node). If
you plan to have a network installation of Ansys products in your Microsoft HPC or Windows-based
cluster, you must also install the prerequisites on each execution node. If you do not install the pre-
requisites on each cluster node, job execution may fail.
You can install the prerequisites separately by running ProductConfig.exe from the top-level directory
as an administrator. You can also install the prerequisites silently using the following command:
For more information, refer to the Ansys Installation and Licensing Documentation.
For a Microsoft HPC cluster, you may also be able to install the prerequisites on all nodes using the
clusrun utility that is part of the Microsoft HPC Pack installation. For more information, refer to the
Microsoft HPC documentation.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 127
RSM Integration with a Cluster or Cloud Portal
Passwords
RSM does not require users to manually cache their Windows password with Microsoft HPC. Each RSM
job runs the hpcutils.exe tool prior to submitting the job to the cluster. This tool programmatically
does the equivalent of cluscfg setcreds.
However, if you still see the error messages regarding the password in the RSM log, such as "Failed to
cache password with HPC" or "Account password MUST be cached with MS Compute Cluster," you may
need to verify that the Service Packs for Microsoft HPC Pack and Windows Server have been properly
installed. If you have not installed the Service Packs, you may still need to run cluscfg setcreds
command from cluster head node to cache the HPC password.
Mixed Domains
You can use RSM when the client computer and the cluster are different domains. The assumption is
that the client computer and user account are on the corporate domain and the cluster is its own domain.
In this case, the cluster domain must be configured to have a ‘one-way trust’ with the corporate domain.
That is, the cluster domain trusts the corporate domain but not vice-versa. Corporate domain users
must be able to use cluster resources (login as CORPORATE\user into a cluster node). If the cluster
administrator can add corporate domain accounts as cluster users, then this trust has likely been con-
figured when the cluster domain was created.
REM Set AWP_ROOT221 on all nodes so that they can use the shared installation, replacing the share name
REM "AnsysInc221" and install root directory "v221" with your actual share and installation root directory
clusrun setx AWP_ROOT221 "\\%COMPUTERNAME%\AnsysInc221\v221" /M
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
128 of ANSYS, Inc. and its subsidiaries and affiliates.
Integrating RSM with a Cloud Portal
You will need to create a configuration for Cloud portal job submission using the RSM Configuration
application. For more information, refer to Defining RSM Configurations (p. 33).
To enable portal job submission on a client machine, the configuration must reside in the default RSM
configuration directory on the client machine. Or, the RSM configuration directory on the client machine
must be set to a shared directory that contains the necessary configuration. See Sharing and Accessing
RSM Configurations (p. 54).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 129
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
130 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 6: RSM User Accounts and Passwords
RSM needs to acquire and cache your credentials to be able to submit jobs to an HPC resource on your
behalf. RSM's direct integration with Ansys applications and built-in caching capability automates this
process to a great extent, and eliminates the need to manually define user accounts in RSM.
Each user account is associated with a specific configuration defined in RSM, and therefore tied to the
RSM queues defined in the configuration. In addition to auto-account creation, you can create a user
account directly in RSM, and apply the account to an RSM configuration.
When you submit a job to RSM from Workbench for the first time, RSM will prompt you to specify cre-
dentials for accessing the HPC resource associated with the chosen RSM queue. RSM will validate and
cache the credentials, and there will be no need to specify credentials again unless your password
changes. See Credential Caching from Workbench (p. 132).
Important:
For improved account security, passwords are always cached on the client machine, even if
jobs will be submitted to a remote resource.
The way in which credentials are specified in client applications differs depending on the application.
Refer to the following:
6.1.1. Credential Caching from Workbench
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 131
RSM User Accounts and Passwords
If your credentials validate, the job will be submitted to the cluster or Cloud portal. If the credentials
do not validate, the job will be aborted.
2. In the right pane, click on the toolbar, or right-click and select Add Account.
3. In the Adding Account dialog box, specify the user name and password that will be used to submit
jobs the HPC resource, then verify the password. If jobs will be submitted to a Windows machine,
ensure that you include the Windows domain when specifying the user name.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
132 of ANSYS, Inc. and its subsidiaries and affiliates.
Adding a User Account
Note:
• The password cannot contain special characters, or characters with special marks such
as accents.
5. In the Apply to panel, select the cluster configuration that will use this account:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 133
RSM User Accounts and Passwords
If you are working in Workbench, RSM will automatically prompt you to specify updated credentials, as
shown in Credential Caching from Workbench (p. 132). Once you specify your new password, the creden-
tials cached in RSM will be updated.
You can also update your account password directly in RSM if needed.
4. In the Changing Password dialog box, the User Name field will be auto-populated with the DO-
MAIN\username of the selected account. Enter and verify the new password.
5. Click OK.
Note:
It is also possible to run the RSM password application manually. For details, see Manually
Running the Password Application (p. 135).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
134 of ANSYS, Inc. and its subsidiaries and affiliates.
Manually Running the Password Application
You can use the RSM Utilities application to run the password application. For more information see
Managing Credentials for RSM Queues (rsm.exe creds) (p. 148).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 135
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
136 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 7: RSM Settings and Utilities
The following sections describe actions or settings that control RSM behavior, and provide an overview
of the commands available in the RSM utilities application:
7.1. Specifying the Job Cleanup Period
7.2. Performing Administrative Tasks with the RSM Utilities Application
7.3. Refreshing the View
Clearing out jobs in a timely manner improves the performance of RSM and optimizes memory usage.
2. In the Job Cleanup Period pane, specify the desired time period for each job status. The following
values are acceptable:
• H (hours) = 0–23
• MM (minutes) = 0–59
• SS (seconds) = 0–59
You can enter only the number of days (without the zeros), only the hours/minutes/seconds, or
both.
Examples:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 137
RSM Settings and Utilities
• 00.15.00 = 15 minutes
For Windows, you can start the RSM Utilities application by opening a command prompt in the
[RSMInstall]\bin directory and running rsm.exe.
For Linux, you can start the RSM Utilities application by running the rsmutils shell script, located in
the [RSMInstall]/Config/tools/linux directory.
Note:
The Linux shell scripts are dependent on their relative location in the Ansys Workbench in-
stallation, so cannot be moved.
The commands shown below can be used on both Windows and Linux:
Usage: rsm.exe xmlrpc|config|appsettings|creds|migration [operator] [options] [arguments]
Where:
xmlrpc {operator} [options]: XmlRpc configuration commands.
config {operator} [options]: Configuration related commands.
appsettings {operator} [options] [arguments]: Appsetting related commands.
creds {operator} [arguments]: Credentials related commands.
migration {operator} [arguments]: Migrate from previous version (e.g. v180).
In this section:
7.2.1. Managing RSM Configurations and Queues (rsm.exe | rsmutils config)
7.2.2. Editing RSM Application Settings (rsm.exe | rsmutils appsettings)
7.2.3. Managing Credentials for RSM Queues (rsm.exe creds)
7.2.4. Migrating RSM from a Previous Version
A configuration contains information about the HPC resource to which jobs will be submitted, and
how RSM will work with the resource.
RSM queues are the queues that users will see in client applications when submitting jobs to RSM.
Each RSM queue maps to an HPC queue and configuration.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
138 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
Configurations are saved to .rsmcc files in the RSM configuration directory. To determine the
location of this directory, refer to Specifying a Directory for RSM Configuration Files (p. 147).
To manually create a configuration, run the appropriate command below, appending options from
the accompanying table to specify configuration settings:
For the [hpc type], specify one of the following to create a configuration for that HPC type: default
(see Example 7.1: Default Ansys RSM Cluster (ARC) Configuration (p. 140)), ARC | LSF | PBS | SLURM
| SGE | UGE | MSHPC.
RSM = RSM uses TCP sockets to stream files from the client machine
to the submit host. Use when the HPC staging directory is in a remote
location that is not visible to client machines.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 139
RSM Settings and Utilities
SCP = SSH/SCP will be used to transfer files from the client machine
to the submit host.
-stagingDir | -sd path The HPC staging directory as the RSM client sees it. A Windows client will
see the shared file system as a UNC path (for example, \\machine\shar-
eName). A Linux client may mount the HPC staging directory such that
the path appears different than it does on the cluster (for example,
/mounts/cluster1/staging). Leave empty if using the
no-file-transfer method.
-stagingMapDirs | The path to the shared file system as the cluster sees it (for example,
-sdmap path;path;... /staging on a Linux machine). This maps the client-visible path
to the cluster-visible path. For example, the Windows client sees
\\machine\STAGING which is a Linux Samba share of /staging.
Multiple paths are only supported when all of the following are true:
-localScratch | -ls path Local scratch path if jobs will run in a scratch directory local to the
execution node. Leave empty to run jobs in the HPC staging directory.
-scratchUnc | -su path (Windows clusters only): UNC share path of -localScratch path not
including the '\\machine\' portion.
-peSmp | -ps name (UGE/SGE only): Parallel Environment (PE) names for Shared Memory
Parallel. If not specified, default will be 'pe_smp'.
-peMpi | -pm name (UGE/SGE only): Parallel Environment (PE) names for Distributed Parallel.
If not specified, default will be 'pe_mpi'.
-noCleanup | -nc Keep job files in the HPC staging directory after the job has run.
The default Ansys RSM Cluster (ARC) configuration is used to submit jobs to the local machine (on
which the RSM configuration resides). Every RSM installation has a basic ARC cluster already con-
figured.
Running the command rsm.exe | rsmutils config create cluster -type default
is the equivalent of running rsm.exe | rsmutils config create cluster -type
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
140 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
ARC -name localhost -rq Local, where the name of the RSM configuration is localhost,
and the RSM queue that is associated with this configuration is named Local.
To configure RSM to use an LSF queue, and run jobs in the local scratch directory, you would run
the following command:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 141
RSM Settings and Utilities
<useSsh>False</useSsh>
<sshAccount />
<useSshForLinuxMpi>True</useSshForLinuxMpi>
<deleteStagingDirectory>True</deleteStagingDirectory>
<readonly>False</readonly>
</ClusterConfiguration>
In this example, an RSM queue name (LSF-SCRATCH) is specified. This will be the queue name
displayed in client applications. If an RSM queue name is not included in the command line (for
example, -rsmQueue LSF-SCRATCH), the actual cluster queue name will be displayed instead.
If you were to open the queues.rsmq file, you would see the LSF-SCRATCH queue added there:
<?xml version="1.0" encoding="utf-8"?>
<Queues>
<Queue version="1">
<name>LSF-SCRATCH</name>
<clusterConfigurationName>LSFSCRATCH</clusterConfigurationName>
<clusterQueueName>normal</clusterQueueName>
<enabled>True</enabled>
</Queue>
</Queues>
The clusterConfigurationName value, LSFSCRATCH in this example, is what links the queue
to the actual RSM configuration.
In the RSM client, the SGE queue sgeshare1 is referred to as SGE_SHARE. all.q is not aliased
and is referred to as all.q.
Local scratch setup: Jobs submitted to RSM all.q queue will run in a local scratch folder /rsmtmp.
No local scratch. Jobs submitted to RSM’s SGE_SHARE queue will run in shared cluster staging
directory.
rsmutils config create cluster -type SGE -name SGESHARE -peSmp myPE
-clusterQueue sgeshare1 -rsmQueue SGE_SHARE
Note that you can specify UGE or SGE for config create. They are the same.
Local scratch setup. Jobs submitted to RSM’s HPC-SCRATCH queue will run in a local scratch folder
C:\RSMTemp. Note that the cluster nodes will all share this folder as \\[Execution-
Node]\RSMTemp.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
142 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
No local scratch. Jobs submitted to RSM’s HPC-SHARE queue will run in shared cluster staging
directory.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 143
RSM Settings and Utilities
Queues:
Default [WinHPC Cluster, Default]
High_mem [ARC, high_mem]
LM-WinHPC [WinHPC, LM]
LSF-SCRATCH [LSFSCRATCH, normal]
Local [localhost, local]
XLM-WinHPC [WinHPC, XLM]
Configurations:
ARC
localhost
LSFSCRATCH
WinHPC
Queues:
LSF-SCRATCH [LSFSCRATCH, normal]
<ClusterConfiguration version="2">
<name>LSFSCRATCH</name>
<type>LSF</type>
<machine>lsfheadnode</machine>
<submitHostPlatform>allLinux</submitHostPlatform>
<stagingDirectory />
<networkStagingDirectory />
<localScratchDirectory>/rsmtmp</localScratchDirectory>
<fileCopyOption>None</fileCopyOption>
<nativeSubmitOptions />
<useSsh>False</useSsh>
<sshAccount />
<useSshForLinuxMpi>True</useSshForLinuxMpi>
<deleteStagingDirectory>True</deleteStagingDirectory>
<readonly>False</readonly>
</ClusterConfiguration>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
144 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
The appsettings command has two possible operators: get (for querying) and set (for editing).
To query or edit a setting in the Ans.Rsm.AppSettings.config file, you must know the name
of the section in which the setting is located (SectionName), and the name of the setting (SettingName).
In the following excerpt from the Ans.Rsm.AppSettings.config file, the SectionName is
Global, and the first SettingName is DiskSpaceLowWarningLimitGb:
<appSettings name="Global">
<add key="DiskSpaceLowWarningLimitGb" value="2.0" />
<add key="IdleTimeoutMinutes" value="10" />
<add key="PingServerTimeout" value="3000" />
<add key="PingServerMaxRetries" value="4" />
<add key="PortInUseTimeout" value="5000" />
<add key="RemotingSecureAttribute" value="false" />
<add key="EnablePerformanceLogging" value="false" />
Important:
If you edit settings in the Ans.Rsm.AppSettings.config file, you may need to restart
RSM services in order for the changes to take effect.
Windows:
To query a setting, run the following command in the [RSMInstall]\bin directory:
Linux:
To query a setting, run the following command in the [RSMInstall]/Config/tools/linux
directory:
Examples
The appsettings command is used in the following sections in this user's guide:
• Configuring a Computer with Multiple Network Interface Cards (NICs) (p. 60)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 145
RSM Settings and Utilities
• Dealing with a Firewall in a Multi-Node Ansys RSM Cluster (ARC) (p. 103)
For example, to set a port range of 2000-9000, you would enter the following:
7.2.2.2. Specifying a Port Range for User Proxy Socket File Transfers
When the cluster submit host is a remote machine, the RSM launcher service launches a user proxy
process on the submit host which transfers files to the cluster on the user's behalf. When this occurs,
a port is opened for each file being transferred. By default, RSM will randomly select a port that is
free. If you want to control which ports RSM can choose, you can specify a port range using the
RSM Utilities application, which modifies the user proxy SocketTransfererPortRange value
in the RSM\Config\Ans.Rsm.AppSettings.config file.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
146 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
Since the default directory is associated with a specific user account, it may not be an appropriate
directory for storing configuration files if you plan to share the directory with other users. For more
information, see Sharing and Accessing RSM Configurations (p. 54).
You can change this to a different directory using the RSM Utilities application (p. 138), or by editing
the RSM\Config\Ans.Rsm.AppSettings.config file.
In this section:
7.2.2.3.1. Querying the Location of the RSM Configuration Directory
7.2.2.3.2. Specifying the Location of the RSM Configuration Directory
If you want to share RSM configurations with other users you can make the chosen configuration
directory a shared directory so that users can retrieve configurations from it. Users can then map
or mount the shared directory on their local machines and point their ConfigurationDirect-
ory setting to it using the same steps presented in this section.
Follow the appropriate set of instructions below to change the RSM configuration directory.
Windows
1. If the RSM launcher service is currently running, stop it. As an administrator, run net stop
RSMLauncherService221.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 147
RSM Settings and Utilities
3. Issue the following command, replacing the path with the desired value:
You can specify a local path if the directory is on the local machine, or a UNC path if the
directory is a network share.
Note that RSM will add the required sub-folders ANSYS\v221\RSM to the specified folder
if they do not already exist (for example, if you specified the path c:\some\folder). Any
new configurations that you create will be saved to this sub-location (for example,
C:\some\folder\ANSYS\v221\RSM). This should also be your target location if you are
copying existing configuration files from another source.
Linux
You can specify a local path, UNC path on Windows, or mounted file system on Linux de-
pending on where the directory resides.
Note that RSM will add the sub-folders ANSYS/v221/RSM to the specified folder if they do
not already exist (for example, if you specified the path /some/folder). Any new config-
urations that you create will be saved to this sub-location (for example,
/some/folder/ANSYS/v221/RSM). This should also be your target location if you are
copying existing configuration files from another source.
The following arguments are used with operators of the creds command:
Argument Usage
-a Account name to be used for job submission to RSM queue. The default is the current
account account.
-rq The RSM queue to which the credentials will be applied. For caching, the default is all
queue queues.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
148 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
Argument Usage
-l Launch UserProxy after validation. (Validation only)
Refer to Table 7.2: Arguments Used for Managing Credentials (p. 148) when running the caching
command.
To cache credentials:
When you run the caching command, you will be prompted for the password. For example:
C:\Program Files\ANSYS Inc\v221\RSM\bin>rsm.exe creds cache -a ANSYS\atester -rq LSF-SCRATCH
Caching password for: ANSYS\atester
Enter password:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 149
RSM Settings and Utilities
Important:
• If you have previously set up a custom Ansys RSM Cluster (ARC), and would like to migrate
ARC-related settings and ARC node configurations, you will need to run the ARC migration
command in addition to running the RSM migration utility described here. For more in-
formation see Migrating an ARC Setup from a Previous Version (arcconfig migra-
tion) (p. 100).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
150 of ANSYS, Inc. and its subsidiaries and affiliates.
Performing Administrative Tasks with the RSM Utilities Application
To perform a migration:
1. On the cluster submit host, log into an account with administrative privileges.
2. If you have not already done so, install the new product version.
3. Run the appropriate command below, appending the desired operator and options from the
accompanying tables.
Operator Usage
config Migrate RSM configurations and queues.
settings Migrate RSM settings in the RSM\Con-
fig\Ans.Rsm.AppSettings.config file.
(To see which settings will be migrated, refer to
the new version's RSM\Config\migra-
tion\Rsm.AppSettings.Migrate.con-
fig file.)
all Migrate everything (configurations, queues, and
settings).
Option Usage
-v123 (Required) Specify the version that you are
migrating, so that the migration command
knows which files to look for. Replace the 123
with the version that you are migrating (for
example, enter -v212 for version 2021 R2).
Note:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 151
RSM Settings and Utilities
Example
In the following example we are migrating cluster configurations, queues and settings from version
18.0 to version 2022 R1.
By using the -preview option we can see that 4 configurations, 8 queues and 1 setting will be migrated:
C:\Program Files\ANSYS Inc\v221\RSM\bin>rsm.exe migration all -v212 -preview -verbose
v221 settings located at C:\Program Files\ANSYS Inc\v221\RSM\Config\Ans.Rsm.AppSettings.config
v212 settings located at C:\Program Files\ANSYS Inc\v212\RSM\Config\Ans.Rsm.AppSettings.config
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
152 of ANSYS, Inc. and its subsidiaries and affiliates.
Refreshing the View
Note:
If a configuration, queue or setting already exists in the new version, and the content or
value is different in the new version, that item will be overwritten. Otherwise, if the content
or value is the same, migration of that item will be skipped.
To verify that configurations and queues have been successfully migrated, you can issue the config
list (p. 144) command, as described in Listing RSM Configurations and Queues (p. 144), or simply open
the RSM Configuration application and review the configurations in the HPC Resources list. To verify
that settings have been migrated, refer to Editing RSM Application Settings (rsm.exe | rsmutils
appsettings) (p. 145).
When the view is refreshed, the following items are updated in the application window:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 153
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
154 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 8: RSM Custom Integration
This section discusses various methods of customizing Ansys Remote Solve Manager.
This section briefly describes the types of files used in the customization:
8.1.1. Job Templates
8.1.2. Job Scripts
8.1.3. HPC Commands File
8.1.4. Job Configuration File
RSM job templates are located in the [RSMInstall]\Config\xml directory. Examples of job
templates in this directory are GenericJob.xml, Workbench_ANSYSJob.xml, and Work-
bench_CFXJob.xml.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 155
RSM Custom Integration
RSM job script files are located in the [RSMInstall]\Config\scripts directory. There is a set
of script files for each cluster type.
Specialized job scripts for integrating RSM with third-party job schedulers or Cloud portals are invoked
based upon the HPC type property in the RSM configuration:
• For example, if you set the HPC type to LSF, then LSF will be your “keyword”, and RSM will look
in the jobConfiguration.xml file for the corresponding hpc_commands file. It will find
hpc_commands_LSF.xml, which invokes the scripts necessary to run a test job on an LSF cluster.
<keyword name="LSF">
<hpcCommands name="hpc_commands_LSF.xml"/>
</keyword>
• If you set the HPC type to Custom, then “Custom” is not used as the keyword. Rather, the value
that you specify in the Custom HPC type field in the RSM configuration will become your “keyword”,
and you can customize job execution on the cluster by adding a section for your new keyword in
the jobConfiguration.xml file. The section will list the files that you want to apply to the
custom job type.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
156 of ANSYS, Inc. and its subsidiaries and affiliates.
Understanding RSM Custom Architecture
The commands inside the HPC command file can point directly to cluster-specific commands (like
bsub or qstat). When the operations are more complex, the commands can reference scripts or
executables that call the cluster software functions internally. These scripts can be in any language
that can be run by the cluster node. The HPC command file is described in greater detail in Custom
Cluster Integration Setup (p. 159).
The jobConfiguration.xml file contains keyword entries for the following job types:
• MSCC
• STORAGE_SSH (for file transfers to/from the HPC staging directory using SSH)
The file also contains entries for a Custom Integration Sample (CIS), and various example clusters.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 157
RSM Custom Integration
Below is the standard keyword entry for jobs running on an LSF cluster using SSH. The keyword for
this job type is LSFSSH. Commands from the uppermost file in the list, hpc_commands_LSF.xml,
will be used first. Subsequently, commands from the included hpc_commands_exten-
sions_SSH.xml file will be used. Duplicate commands in included file will be ignored (because it
is not first in the list), and commands that are not duplicates will be combined with those in the first
file to form the complete HPC commands file.
<keyword name="LSFSSH">
<hpcCommands name="hpc_commands_LSF.xml">
<include name="hpc_commands_extensions_SSH.xml"/>
</hpcCommands>
<protocol name="protocol_definition_SSH.xml"/>
</keyword>
If you want to customize the way jobs are run on an HPC resource, you will need to modify the
jobConfiguration.xml file, either by editing existing keyword entries or adding new ones for
each custom HPC job type.
For example, assuming that your cluster is most like a standard LSF cluster that uses SSH, you could
modify the existing LSFSSH keyword entry in the jobConfiguration.xml file. The modified
entry below references a new hpc_commands file, MyCustomChanges.xml, that contains changes
to only one or a few commands. The <include> option is used to include the original hpc_com-
mands_LSF.xml and hpc_commands_extensions_SSH.xml files to make use of any standard
commands that are not present in the custom file. Commands in the MyCustomChanges.xml file
will be used first, because that file is at the top of the list. Any non-duplicate commands in the two
included files will be merged with those in the custom file to form the complete HPC commands file.
<keyword name="LSFSSH">
<hpcCommands name="MyCustomChanges.xml"> <!-- First file in the list will override commands in the
included files below -->
<include name=hpc_commands_LSF.xml/>
<include name="hpc_commands_extensions_SSH.xml"/>
</hpcCommands>
<protocol name="protocol_definition_SSH.xml"/>
</keyword>
If significant changes are to be made to HPC job execution, or there are different customization re-
quirements for different projects, groups, and so on, then you will need to add a custom keyword
entry to the jobConfiguration.xml file.
In the example below, a new entry has been created for jobs running on a custom cluster that has
been assigned the keyword CUSTOM. It references a custom HPC commands file, hpc_commands_CUS-
TOM.xml, as well as the standard hpc_commands_PBS.xml file.
<keyword name="CUSTOM">
<hpcCommands name="hpc_commands_CUSTOM.xml">
<include name="hpc_commands_PBS.xml"/>
</hpcCommands>
</keyword>
For more information see Modifying the Job Configuration File for a New Cluster Type (p. 161).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
158 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
RSM also provides a custom cluster integration mechanism that allows third parties to use custom
scripts to perform the tasks needed to integrate Workbench with the cluster. The custom integration
scenarios can be grouped into the following categories in order of complexity:
• Commercial clusters (listed above) for which you need some additional operation to be performed
as part of the RSM job execution. This is a type of cluster-side integration.
• “Unsupported” clusters, not included in the list above, that you want to use for executing a job via
RSM. This is also a type of cluster-side integration.
• You have specialized requirements that need to fully replace RSM functionality with 3rd-party scripts
for handling all aspects of job submission, including file transfer. This is called client-side integration.
The terms cluster-side and client-side integration refer to the location (in the RSM architecture) where
the custom script files are going to be located.
If the RSM client will be submitting jobs to a remote cluster, RSM files will be customized on the cluster
submit host. This is referred to as cluster-side integration. For cluster-side integration, RSM must be
installed on the cluster head node and file transfers should be handled by RSM, using either the internal
RSM transfer mechanism or operating system file transfer to an existing network share. The methods
of file transfer discussed in Setting Up Job Directories and File Transfers (p. 57) are available, except
for Using SSH for File Transfers (p. 61) and Custom Client Integration (p. 61).
If the RSM client is the cluster submit host, RSM files will be customized on the RSM client. This is referred
to as client-side integration. In this scenario, the RSM functionality is completely replaced by the 3rd-
party scripts. However, only a thin layer of the RSM architecture is involved, in order to provide the APIs
for execution of the custom scripts, which are located on the client machine.
Note that for supported clusters it is also possible to include additional job submission arguments to
the command executed by the cluster. The addition of custom submission arguments does not require
the creation of custom scripts.
The following sections describe the general steps for customization with cluster-side and client-side
integration. The detailed instructions for writing the custom code are similar for the two cases. They
are addressed in Writing Custom Code for RSM Integration (p. 172).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 159
RSM Custom Integration
RSM allows you to customize your integration with supported cluster types (LSF, PBS Pro, SLURM,
Windows HPC, and SGE) by starting with examples of production code for one of the standard cluster
types and then changing command lines or adding custom code where necessary. If an unsupported
cluster is being used, the recommended procedure is still to start from the example files for one of
the supported clusters.
When customizing files, you must choose a "keyword" that represents your custom cluster type. This
is a short word or phrase that you will append to the file names of your custom files, and use when
defining an RSM configuration to map the configuration to your custom files. The name is arbitrary,
but you should make it simple enough to append to file names. For example, if you are creating a
customized version of an LSF cluster, your keyword might be "CUS-LSF". The only requirement is that
you consistently use the same capitalization in all places where the keyword is referenced.
For a cluster-side RSM installation, you will need to log into the remote cluster submit host to perform
the following steps:
1. Create a copy of the HPC commands file that most closely matches your custom cluster
type, and replace the keyword with your custom cluster type “keyword” (for example,
hpc_commands_<keyword>.xml).
2. Add an entry to the job configuration file that associates your custom “keyword” with the
cluster-specific hpc_commands_<keyword> file.
3. Edit the cluster-specific hpc_commands_<keyword> file to reference the code that you
want to execute.
The following sections discuss the steps needed for custom cluster-side integration:
8.2.1.1. Creating Copies of Standard Cluster Code Using a Custom Cluster Keyword
8.2.1.2. Modifying the Job Configuration File for a New Cluster Type
8.2.1.3. Modifying the Cluster-Specific HPC Commands File
8.2.1.4. Creating a Configuration for the Custom Cluster
You may also want to refer to the Configuring Custom Cluster-Side Integration tutorial on the Ansys
Help site.
Note that all the actions listed below should be performed on the cluster installation.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
160 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
• Locate the HPC commands file that pertains to your cluster type (for instance, if you are using
PBS Pro, the file is hpc_commands_PBS.xml).
• Copy the content of the hpc_commands_<hpcType>.xml file into a new file, and name the
new file hpc_commands_<YOURKEYWORD>.xml (where <YOURKEYWORD> is a short word or
phrase that you have chosen to represent the custom cluster type). For example, if your keyword
for the custom cluster is “CUS_PBS”, the new file should be called hpc_com-
mands_CUS_PBS.xml.
Note:
Do not rename or make changes to the standard templates that ship with RSM (LSF, PBS
Pro, and so on). This can cause those standard cluster setups to fail and will make it
harder to start over if you need to change something later on. Here we have created a
custom cluster type, but used copies of a standard template from which to start. This is
the recommended method.
8.2.1.2. Modifying the Job Configuration File for a New Cluster Type
As part of the setup, you must add an entry for your custom cluster keyword in the jobConfig-
uration.xml file, and reference the files that are needed for that cluster job type.
Note that all the actions listed below should be performed on the cluster installation.
• Open the jobConfiguration.xml file and add an entry that follows the pattern shown in
the sample code below. This code corresponds to the example in preceding sections which as-
sumes your cluster is most like a PBS cluster.
<keyword name="YOURKEYWORD">
<hpcCommands name="hpc_commands_YOURKEYWORD.xml">
</hpcCommands>
</keyword>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 161
RSM Custom Integration
<arg>%RSM_HPC_CORES%</arg>
</arguments>
<outputs>
<variableName>RSM_PBS_MEMORY_AMOUNT</variableName>
</outputs>
<condition>
<env name="RSM_HPC_MEMORY">ANY_VALUE</env>
<!--Important : We only need to run this pre command for non-distributed jobs because of a
difference in how pbs wants memory reservation to be formatted-->
<env name="RSM_HPC_DISTRIBUTED">FALSE</env>
</condition>
</command>
</precommands>
<primaryCommand name="submit">
<application>
<app>qsub</app>
</application>
<arguments>
<arg>
<value>-q %RSM_HPC_QUEUE%</value>
<condition>
<env name="RSM_HPC_QUEUE">ANY_VALUE</env>
</condition>
</arg>
<arg>
<value>-l select=%RSM_HPC_CORES%:ncpus=1:mpiprocs=1</value>
<condition>
<env name="RSM_HPC_DISTRIBUTED">TRUE</env>
</condition>
</arg>
<arg>
<value>-l select=1:ncpus=%RSM_HPC_CORES%:mpiprocs=%RSM_HPC_CORES%</value>
<condition>
<env name="RSM_HPC_DISTRIBUTED">FALSE</env>
</condition>
</arg>
<!-- Caution, memory reservation must always be defined immediately after the core info because
it is part of the value for the -l argument-->
<arg noSpaceOnAppend="true">
<value>:mem=%RSM_HPC_MEMORY%mb</value>
<condition>
<env name="RSM_HPC_MEMORY">ANY_VALUE</env>
<env name="RSM_HPC_DISTRIBUTED">TRUE</env>
</condition>
</arg>
<arg noSpaceOnAppend="true">
<value>:mem=%RSM_PBS_MEMORY_AMOUNT%</value>
<condition>
<env name="RSM_PBS_MEMORY_AMOUNT">ANY_VALUE</env>
<env name="RSM_HPC_DISTRIBUTED">FALSE</env>
</condition>
</arg>
<arg>
<value>-l place=excl</value>
<condition>
<env name="RSM_HPC_NODE_EXCLUSIVE">TRUE</env>
</condition>
</arg>
<arg>-N "%RSM_HPC_JOBNAME%" %RSM_HPC_NATIVEOPTIONS% -V -o "%RSM_HPC_STAGING%/%RSM_HPC_STDOUTFILE%"
-e "%RSM_HPC_STAGING%/%RSM_HPC_STDERRFILE%" "%RSM_HPC_STAGING%/%RSM_HPC_COMMAND%"</arg>
</arguments>
</primaryCommand>
<postcommands>
<command name="parseSubmit">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/pbsParsing.py</pythonapp>
</application>
<arguments>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
162 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
<arg>-submit</arg>
<arg>
<value>%RSM_HPC_PARSE_MARKER%</value>
<condition>
<env name="RSM_HPC_PARSE_MARKER">ANY_VALUE</env>
</condition>
</arg>
</arguments>
<outputs>
<variableName>RSM_HPC_OUTPUT_JOBID</variableName>
</outputs>
</command>
</postcommands>
</submit>
The section in bold text is the section that provides the Submit action, which we want to customize
in this example. In the original version the Submit command invokes the cluster qsub with argu-
ments determined via environment variables. The actual executable that is submitted to the cluster
is determined by RSM during runtime and can be specified via an environment variable named
RSM_HPC_COMMAND. For details, see Submit Command (p. 176).
The example below shows the same section after it is customized to execute the Python file sub-
mit_PBS_EXAMPLE.py. In this example, we defined the type of application to execute (runpy-
thon, accessed from the Ansys installation) and the name of the Python file to be executed (sub-
mit_PBS_EXAMPLE.py).
<submit>
<primaryCommand name="submit">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<app>%AWP_ROOT221%/commonfiles/CPython/2_7_13/linx64/Release/runpython</app>
</application>
<arguments>
<arg>
<value>%RSM_HPC_SCRIPTS_DIRECTORY%/submit_PBS_EXAMPLE.py</value>
</arg>
</arguments>
</primaryCommand>
<postcommands>
<command name="parseSubmit">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/pbsParsing.py</pythonapp>
</application>
<arguments>
<arg>-submit</arg>
<arg>
<value>%RSM_HPC_PARSE_MARKER%</value>
<condition>
<env name="RSM_HPC_PARSE_MARKER">ANY_VALUE</env>
</condition>
</arg>
</arguments>
<outputs>
<variableName>RSM_HPC_OUTPUT_JOBID</variableName>
</outputs>
</command>
</postcommands>
</submit>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 163
RSM Custom Integration
The custom Submit command appears much simpler than the original one. However, the details
of the submission are handled inside the Python file, which contains the same arguments used in
the original section. The Python file will also contain any custom code to be executed as part of
the submission.
Note:
Other commands or queries can be overridden using the same procedure. You can find the command
name in the cluster-specific hpc_commands file and replace the application that needs to be ex-
ecuted and the arguments needed by the application. Details on how to provide custom commands,
as well as the description of the environment variables, are provided in Writing Custom Code for
RSM Integration (p. 172).
A “custom HPC integration” means that you are running in non-SSH mode (RSM is able to commu-
nicate directly with the cluster). Therefore, when specifying how the client communicates with the
cluster, you need to select Able to directly submit and monitor HPC jobs.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
164 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
For the File Management tab, see Specifying File Management Properties (p. 41) for details on
the different file transfer scenarios.
The mechanism and operations for custom client-side integration are very similar to the ones for
custom cluster-side integration. However, the underlying architecture is different. In the cluster-side
integration, the customization affects the scripts used for RSM execution on the cluster side. In the
client-side integration, only a thin layer of RSM on the client side is involved. The layer provides the
APIs for the execution of the custom scripts, which are located on the client machine. It is the respons-
ibility of the custom scripts to handle all aspects of the job execution, including transfer of files to
and from the HPC staging directory (if needed).
The RSM installation provides some prototype code for client integration that can be tailored and
modified to meet specific customization needs.
When customizing files, you must choose a "keyword" that represents your custom HPC type. This is
a short word or phrase that you will append to the file names of your custom files, and use when
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 165
RSM Custom Integration
defining a configuration in RSM to map the configuration to your custom files. The name is arbitrary,
but you should make it simple enough to append to file names. For example, if you are creating a
customized version of an LSF cluster, your keyword might be "CUS-LSF". The only requirement is that
you consistently use the same capitalization in all places where the keyword is referenced.
For client-side integration, you will be using the local client machine to perform the following steps:
1. Create a copy of the HPC commands file that most closely matches your custom cluster type,
and replace the keyword with your custom cluster type “keyword” (for example, hpc_com-
mands_<keyword>.xml).
2. Add an entry to the job configuration file that associates your custom cluster “keyword” with
the cluster-specific hpc_commands_<keyword> file.
3. Edit the cluster-specific hpc commands_<keyword> file to reference the custom commands.
4. Provide cluster-specific script\code\commands that perform the custom actions and return the
required RSM output.
Once you have completed the steps above, you can create a configuration in RSM that will use your
customized files.
You may also want to refer to the Configuring Custom Client-Side Integration tutorial on the Ansys
Help site.
For this example we will use the sample file marked with the suffix CIS (Client Integration Sample),
which is based on integration with an LSF cluster.
Note that all the actions listed below should be performed on the client machine.
1. Using the RSM installation on your client machine, locate the directory [RSMInstall]\Con-
fig\xml.
3. Copy the content of the hpc_commands_CIS.xml file into a new file, hpc_commands_<YOUR-
KEYWORD>.xml. If, for example, your keyword is “CUS_LSF”, the new file should be called
hpc_commands_CUS_LSF.xml.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
166 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
The client-side integration requires a custom implementation to be provided for all the commands
to be executed on the cluster. The standard RSM installation includes sample scripts for all these
commands, which should be used as a starting point for the customization. The sample scripts are
named submitGeneric.py, cancelGeneric.py, statusGeneric.py, scp.py, and
cleanupSSH.py. They are located in the [RSMInstall]\RSM\Config\scripts directory.
While it is not absolutely necessary to create a copy and rename the scripts, we have done so for
consistency; in the rest of the example, it is assumed that they have been copied and renamed to
add the same keyword chosen for the custom cluster (for example, submit_CUS_LSF.py, can-
cel_CUS_LSF.py, status_CUS_LSF.py, and cleanup_CUS_LSF.py). These scripts will have
to be included in the custom job template, as shown in the following section, Modifying the Job
Configuration File for a New Cluster Type (p. 167).
These scripts are actually sample scripts that use a fully custom client integration on a standard
LSF cluster, for example only. Generally, custom client integrations do not use standard cluster
types, and thus there are no samples for custom client integrations on other cluster types.
Note:
Any additional custom code that you want to provide as part of the customization should
also be located in the [RSMInstall]\RSM\Config\scripts directory corresponding
to your local (client) installation. Alternatively, a full path to the script must be provided
along with the name.
8.2.2.2. Modifying the Job Configuration File for a New Cluster Type
As part of the setup, you must add an entry for your custom cluster “keyword” in the jobConfig-
uration.xml file, and reference the files that are needed for that cluster job type.
Note that all the actions listed below should be performed on your client machine.
2. Open the jobConfiguration.xml file and add an entry that follows the pattern shown in
the sample code below. This code corresponds to the example in preceding sections which as-
sumes your cluster is most like an LSF cluster.
Note:
In our example we have been using “CUS_LSF” as the keyword, but you still must replace
“YOURKEYWORD” with the actual custom cluster keyword you have defined.
<keyword name="YOURKEYWORD">
<hpcCommands name="hpc_commands_YOURKEYWORD.xml"/>
</keyword>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 167
RSM Custom Integration
This example script is set up to be run on a modified LSF cluster. If you are running on a different
cluster type, you will need to choose a different parsing script (or write a new one) depending on
the cluster type that you have chosen. Parsing scripts are available for supported cluster types: LSF,
PBS Pro, SLURM, UGE, and MSCC. They are named lsfParsing.py, pbsParsing.py, ugePars-
ing.py, and msccParsing.py respectively. If you are using an unsupported cluster type, you
will need to write your own parsing script. For details refer to Parsing of the Commands Out-
put (p. 173).
The hpc_commands file provides the information on how commands or queries related to job
execution are executed. The file can also refer to a number of environment variables. Details on
how to provide custom commands, as well as the description of the environment variables, are
provided in Writing Custom Code for RSM Integration (p. 172).
<jobCommands version="3" name="Custom Cluster Commands">
<environment>
<env name="RSM_HPC_PARSE">LSF</env>
<env name="RSM_HPC_PARSE_MARKER">START</env> <!-- Find "START" line before parsing according to parse type -
<env name="RSM_HPC_SSH_MODE">ON</env>
<env name="RSM_HPC_CLUSTER_TARGET_PLATFORM">Linux</env>
<!-- Still need to set RSM_HPC_PLATFORM=linx64 on Local Machine -->
</environment>
<submit>
<primaryCommand name="submit">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/submit_CUS_LSF.py</pythonapp>
</application>
<arguments>
</arguments>
</primaryCommand>
<postcommands>
<command name="parseSubmit">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/lsfParsing.py</pythonapp>
</application>
<arguments>
<arg>-submit</arg>
<arg>
<value>%RSM_HPC_PARSE_MARKER%</value>
<condition>
<env name="RSM_HPC_PARSE_MARKER">ANY_VALUE</env>
</condition>
</arg>
</arguments>
<outputs>
<variableName>RSM_HPC_OUTPUT_JOBID</variableName>
</outputs>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
168 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
</command>
</postcommands>
</submit>
<queryStatus>
<primaryCommand name="queryStatus">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/status_CUS_LSF.py</pythonapp>
</application>
<arguments>
</arguments>
</primaryCommand>
<postcommands>
<command name="parseStatus">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/lsfParsing.py</pythonapp>
</application>
<arguments>
<arg>-status</arg>
<arg>
<value>%RSM_HPC_PARSE_MARKER%</value>
<condition>
<env name="RSM_HPC_PARSE_MARKER">ANY_VALUE</env>
</condition>
</arg>
</arguments>
<outputs>
<variableName>RSM_HPC_OUTPUT_STATUS</variableName>
</outputs>
</command>
</postcommands>
</queryStatus>
<cancel>
<primaryCommand name="cancel">
<properties>
<property name="MustRemainLocal">true</property>
</properties>
<application>
<pythonapp>%RSM_HPC_SCRIPTS_DIRECTORY_LOCAL%/cancel_CUS_LSF.py</pythonapp>
</application>
<arguments>
</arguments>
</primaryCommand>
</cancel>
</jobCommands>
Note:
Any custom code that you want to provide as part of the customization should also be
located in the [RSMInstall]\RSM\Config\scripts directory corresponding to
your local (client) installation. Alternatively, a full path to the script must be provided
along with the name.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 169
RSM Custom Integration
A “custom client integration” means that you are running in SSH mode (or non-RSM communication).
Thus, when specifying how the client communicates with the cluster, you need to select Use SSH
or custom communication to the submit host, and specify the account name that the Windows
RSM client will use to access the remote Linux submit host.
For the File Management tab, see Specifying File Management Properties (p. 41) for details on
the different file transfer scenarios.
If you are using a custom mechanism for file transfers, you will need to create a custom HPC storage
commands file. You will then need to add an entry to the jobConfiguration.xml file which
references these files.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
170 of ANSYS, Inc. and its subsidiaries and affiliates.
Custom Cluster Integration Setup
1. When creating a configuration, on the File Management tab, select External mechanism for
file transfer (SCP, Custom), then select either SCP via SSH or Custom from the Transfer
mechanism dropdown. Fill out the fields as described in File Management Properties for a
Cluster (p. 41).
If you are using a Custom mechanism, make note of the keyword specified in the Custom
transfer type field, as it will be used in upcoming steps.
2. To define custom commands for transferring files to/from the HPC staging directory, edit or
create a custom HPC storage commands file in the [RSMInstall]\Config\xml directory.
• If you are using SSH-SCP, edit the hpc_storage_SSH.xml file. You may want to create a
backup of the original file before editing it.
Some commands use specific scripts for execution. For example, if you are using SSH, the up-
loadToStorage and downloadFromStorage commands use the scp.py script. If you are
using a custom mechanism, you will need to create custom command scripts and reference them
in the command definitions in the HPC storage commands file. For example, if your custom
keyword is AnsSecureTransfer, you might create a script called AnsSecureTransfer.py,
and reference that script in the uploadToStorage and downloadFromStorage commands.
RSM will set the following variables when file transfer commands are executed:
• RSM_HPC_FILEDIRECTION
• RSM_HPC_FILECONTEXT
• RSM_HPC_FILELIST
• RSM_HPC_FILEUPLOADTRACKING
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 171
RSM Custom Integration
• RSM_HPC_FILE_LOCAL_DIR
• RSM_HPC_FILE_REMOTE_DIR
For more information about these variables, see Environment Variables Set by RSM (p. 179).
3. If you are using a custom mechanism, edit the jobConfiguration.xml file in the [RSMIn-
stall]\Config\xml directory.
Create an entry for your custom transfer type which references the custom HPC storage commands
file (and optionally the custom protocol definition file) that you created in the previous steps.
Be sure to append your keyword (that you specified in step 1) to the keyword name. For example:
<keyword name="STORAGE_AnsSecureTransfer">
<hpcCommands name="hpc_storage_AnsSecureTransfer.xml"/>
</keyword>
Note that there is already an entry for the SSH-SCP protocol in the jobConfiguration.xml
file named STORAGE_SSH-SCP.
The custom code can be in any form convenient to you, typically in the form of scripts or executables.
Generally, scripts are used to wrap the underlying cluster software (for example, LSF) commands. You
can review sample Python scripts in the [RSMInstall]\Config\scripts directory.
The scripts have access to standard login environment variables as well as environment variables that
are dynamically set by RSM to provide information about job- and storage-related actions. A detailed
description of the environment variables that the scripts can access is given in Custom Integration En-
vironment Variables (p. 178).
Each command has some environment variables provided that RSM expects the command will need in
order to function as expected. Some commands have required outputs as well. If a command is being
written as an entirely new script or executable to be used solely for RSM integration, then the desired
inputs and outputs can be directly read and written from that script or executable. However, in most
cases, the command is often something that is already written, and that has only predetermined com-
mand line options such as mkdir, bsub, qstat, and cp. In this case, RSM allows the use of pre and
post commands in order to set up or parse some information from a basic command in order to provide
the RSM output required.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
172 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
Parsing scripts are already implemented for supported cluster types (LSF, PBS, and so on). They are
named lsfParsing.py, pbsParsing.py, and so on.
If you are not using the default implementation of a supported cluster, or you are trying to integrate
with an unsupported cluster, then you will need to write your own parsing script(s), or write a wrapper
script that includes the parsing internally.
In order to parse the output of the primary command and return an answer to the RSM code, you
need to use a few standardized variables. Every post command has access to these variables.
• RSM_HPC_PRIMARY_STDOUT
• RSM_HPC_PRIMARY_STDERR
These environment variables are set by RSM when running the post commands (parsing) code.
They contain all of the standard output and standard error, respectively, from the associated primary
command. The post command (parsing) script simply needs to make a call to the environment
variables above to get the necessary output from the associated command. Then, after obtaining
the standard output of the primary command, it can be parsed using any method desired in order
to create the required RSM output.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 173
RSM Custom Integration
• Unknown
• Queued
• Running
• Finished
• Failed
• Cancelled
getAll- parseAll- RSM_HPC_OUT- The job statuses that are on the cluster in
Status Status PUT_GENERIC a Python dictionary format like
{JobId1:Status1, JobId2:Status2, ….. }, where
the jobID is the string from the cluster and
the Status is one of the statuses from the
list provided for parseStatus.
queryQueues check- RSM_HPC_OUT- Must be set to TRUE or FALSE based on
QueueEx- PUT_QUEUE_DEFINED whether or not the queue was found.
ists
get- parseQueueL-RSM_HPC_OUT- The queues that are available to the user
AllQueues ist PUT_GENERIC for job submission, in a Python list format
such as [queue1, queue2, queue3, …..].
listStor- parseList- RSM_HPC_OUT- The files that are contained in
ageCon- ing PUT_GENERIC RSM_HPC_FILE_REMOTE_DIR in a
tents Python dictionary format such as
{FileName:[Size, ModifiedDate],
FileName2:[Size2, ModifiedDate2],
…….}, where:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
174 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
The handling of the command output depends on the post commands parseSubmit, pars-
eStatus and checkQueueExists discussed previously.
Error Handling
Error messages and warnings information are written to stdout as necessary. If they are properly
labeled as indicated below, they will appear in the RSM log as orange for warnings and bold red
for errors.
Output format:
• RSM_HPC_ERROR=<errormessage>
• RSM_HPC_WARN=<warning>
Debugging
Debugging information, typically used for troubleshooting purposes, is shown in the RSM job log
only if the Debug Messages option is selected from the job log context menu. (To access this option,
right-click anywhere inside the job log pane of the RSM application main window.)
Output format:
RSM_HPC_DEBUG=<debugmessage>
Job-Specific Commands
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 175
RSM Custom Integration
The parsing of the output of the submit command is handled by the post command parseSub-
mit, which will look through all of the output from the submit script, find the Job ID, and return
it in a formal manner. For details see Parsing of the Commands Output (p. 173).
The custom integration infrastructure provides the Python script, ClusterJobs.py, in the
[RSMInstall]\Config\scripts directory. The script serves as a layer of abstraction that allows
a user-selected operation (such as a component update for one or more of the applications or a
design point update) to be invoked without the need to be aware of the command line arguments
and options required for the appropriate submission of the job.
In the Submit command, the ClusterJobs.py script should be invoked (rather than executing
the individual applications). This Python script should be considered as a layer that builds the ap-
propriate command line and sets the appropriate environment variables for the remote execution.
The usage of application specific command line in the Submit script is strongly discouraged and
cannot be properly supported in a general way.
For user convenience, the complete Python command that contains the job to be executed by the
Submit command (for instance, by LSF bsub) is provided through the environment variable
RSM_HPC_COMMAND.
Examples:
• Custom server examples for LSF, PBS Pro, SGE, and MSCC are located in the [RSMIn-
stall]\Config\scripts\EXAMPLES directory.
• A generalized custom client example (for all cluster types) is provided in the file sub-
mitGeneric.py, located in the [RSMInstall]\Config\scripts directory.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
176 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
The output of the queryStatus command should be direct output from the cluster. Any parsing
of this output can be done by the post command parseStatus, which will look through all of
the output from the queryStatus command, find the job status, and return it in a formal manner.
For details see Parsing of the Commands Output (p. 173).
Examples:
• A generalized custom client example (for all cluster types) is provided in the file
statusGeneric.py, located in the [RSMInstall]\Config\scripts directory.
The command should query the cluster for the status of all the jobs available and return the job
status for each in a Python dictionary format, or add a post (parsing) command to convert the
output into the format described in Required Output from Commands (p. 173). If no jobs are running,
a dictionary result should still be returned, but it should be “{-1:UNKNOWN}”.
Given a queue name, RSM_HPC_QUEUE, the command should check if that queue is valid for this
user and return TRUE or FALSE as described in Outputting Variables from the Parsing Scripts (p. 173)
and Required Output from Commands (p. 173) (checkQueueExists).
No output is required from the cancel command. However, an output statement should be given
for verification in the RSM log.
Examples:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 177
RSM Custom Integration
• A generalized custom client example (for all cluster types) is provided in the file can-
celGeneric.py, located in the[RSMInstall]\Config\scripts directory.
No output is required from the createStorage command. However, an output statement should
be given for verification in the RSM log.
No output is required from the deleteStorage command. However, an output statement should
be given for verification in the RSM log.
The format of the output should be a that of a stringified Python dictionary of the filenames along
with other required information as shown in Required Output from Commands (p. 173).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
178 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
the custom scripts or executables in the HPC commands file. Other environment variables can be set
by your RSM administrator, if appropriate to your job management process.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 179
RSM Custom Integration
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
180 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
RSM_HPC_STAGING Path for the cluster’s central staging area for job files.
Typically needed when client and cluster platforms
are different.
Additionally, the user can set any number of variables that follow in Providing Custom Client Inform-
ation for Job Submission (p. 181).
• The username of the submitter (which, for instance, provides the ability to monitor jobs submitted
by a particular user for accounting purposes)
• The license necessary to execute the job, which can be used to integrate with cluster resource
management to check Ansys license availability before a job starts running
For more information on how to integrate licensing with cluster software, contact your cluster admin-
istrator or Ansys customer support.
As an example, we’ll pass the submitter’s username from the client to a PBS Pro cluster.
The following sections detail the steps for providing custom information for job submissions to
clusters.
8.3.4.1. Defining the Environment Variable on the Client
8.3.4.2. Passing the Environment Variable to the Cluster
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 181
RSM Custom Integration
In the example below, we’ve defined the environment variable RSM_CLIENT_USERNAME. The
name is arbitrary as long as it begins with the RSM_CLIENT_ prefix.
In the code sample below, you can see that the environment variable is added to the qsub com-
mand. Note, also, that it is preceded by –A, which defines the account string associated with the
job for the PBS Pro cluster.
<primaryCommand name="submit">
<application>
<app>qsub</app>
</application>
<arguments>
<arg>
<value>-q %RSM_HPC_QUEUE%</value>
<condition>
<env name="RSM_HPC_QUEUE">ANY_VALUE</env>
</condition>
</arg>
<arg>
<value>-A %RSM_CLIENT_USERNAME%</value>
<condition>
<env name="RSM_CLIENT_USERNAME">ANY_VALUE</env>
</condition>
</arg>
<arg>
<value>-l select=%RSM_HPC_CORES%:ncpus=1:mpiprocs=1</value>
<condition>
<env name="RSM_HPC_DISTRIBUTED">TRUE</env>
</condition>
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
182 of ANSYS, Inc. and its subsidiaries and affiliates.
Writing Custom Code for RSM Integration
</arg>
<arg>
<value>-l select=1:ncpus=%RSM_HPC_CORES%:mpiprocs=%RSM_HPC_CORES%</value>
<condition>
<env name="RSM_HPC_DISTRIBUTED">FALSE</env>
</condition>
</arg>
<arg>-N "%RSM_HPC_JOBNAME%" %RSM_HPC_NATIVEOPTIONS% -V -o "%RSM_HPC_STAGING%/%RSM_HPC_STDOUTFILE%" -e
"%RSM_HPC_STAGING%/%RSM_HPC_STDERRFILE%" "%RSM_HPC_STAGING%/%RSM_HPC_COMMAND%"</arg>
</arguments>
</primaryCommand>
To view a sample of this file before the addition of custom information, see Modifying the Cluster-
Specific HPC Commands File (p. 168).
Since we added the environment variable to the qsub command in the PBS Pro commands file, it
will also show up in the area of the job log indicating that the qsub command has been run.
qsub -q %RSM_HPC_QUEUE% -A %RSM_CLIENT_USERNAME% -1
select=1:ncpus=%RSM_HPC_CORES%:mpiprocs=%RSM_HPC_CORES% ...
qsub -q WB_pbsnat -A myname -1 select=1:ncpus=1:mpiprocs=1 ...
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 183
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
184 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 9: RSM Job Monitoring
The RSM Job Monitoring application enables you to monitor jobs that you have submitted to a cluster
from Workbench. Although jobs can be monitored directly in Workbench, the RSM Job Monitoring
application enables you to monitor jobs if the client application is closed or inaccessible. This standalone
application is included in every RSM installation for your convenience.
Note:
The RSM Job Monitoring application can be used to monitor the jobs of the current user
only.
In this chapter:
9.1. Launching the RSM Job Monitoring Application
9.2. Monitoring Jobs in the RSM Job Monitoring Application
9.3. Viewing a Job Log
9.4. Managing Jobs
• On Windows, select Start > Ansys 2022 R1 > RSM Job Monitoring 2022 R1.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 185
RSM Job Monitoring
You can use context-sensitive menus in the upper and lower panes to perform a variety of actions. For
example, you can customize the job list display, and perform actions on a job such as Abort or Interrupt.
In this section:
9.2.1. Viewing the Status of Jobs
9.2.2. Enabling Live Job Monitoring
9.2.3. Controlling the Job List Display
9.2.4. Filtering the Job List
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
186 of ANSYS, Inc. and its subsidiaries and affiliates.
Monitoring Jobs in the RSM Job Monitoring Application
By default, the RSM Job Monitoring application does not perform a real-time status refresh. This is
because the client application already performs a real-time status refresh, and it is assumed that you
will be monitoring your jobs in the client application. Rather, the RSM Job Monitoring application
refreshes job status based on the cached job status in RSM, which is the status that was obtained the
last time that a query was made, and saved in the job's history. This prevents unnecessary querying
of the HPC resource (since the client application is already doing that), and improves monitoring
performance.
If the client application is shut down, however, and you want to use the RSM Job Monitoring applic-
ation to monitor your jobs, it is recommended that you enable real-time refresh mode in the RSM
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 187
RSM Job Monitoring
Job Monitoring application to get real-time job status updates from the HPC resource. If you do not
enable real-time refresh mode, the reported status of running jobs may not be an up-to-the-minute
reflection of the true job status.
If you subsequently return to monitoring jobs from the client application, you can disable real-time
refresh mode in the RSM Job Monitoring application by clicking on the toolbar.
If you right-click any column header, a context menu is displayed which contains the following options:
To display a log for a particular job, simply click the job in the upper pane. Details about the job are
displayed in the lower pane:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
188 of ANSYS, Inc. and its subsidiaries and affiliates.
Viewing a Job Log
The log automatically scrolls to the bottom to keep the most recent messages in view. If necessary you
can copy and paste the log's content into a text editor, or save the log to a file.
In this section:
9.3.1. Controlling the Job Log Display
9.3.2. Copying Text in the Job Log Display
9.3.3. Saving a Job Report
9.3.4. Hiding/Showing the Job Log Pane
View Line Numbers Toggle the display of line numbers at the beginning of each line in the job
log.
View Time Stamps Toggle the display of time stamps at the beginning of each line in the job
log.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 189
RSM Job Monitoring
View Exclusively Only display file-related messages in the job log, such as those relating to
File Messages file transfers.
View Debug Toggle the display of debugging information in the job log.
Messages
SubTask Filters Filters the content of the job log by design point sub-task. This option is only
displayed when the job contains design points.
Scroll to Top Go to the top of the job log.
Scroll to Bottom Go to the bottom of the job log.
To copy text:
• To select all text, right-click in the job log pane and choose Select All.
As an alternative you can use the Save Job Report action to instantly save the entire job log to a
file. See Saving a Job Report (p. 190).
1. In the upper pane, select the job whose log you want to save.
2. In the lower pane, right-click in the job log view and select Save Job Report.
3. In the Save Job Report dialog box, specify the desired save location and file name:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
190 of ANSYS, Inc. and its subsidiaries and affiliates.
Managing Jobs
You can also specify whether or not you want to display the following in the report:
• Debug Messages
• Line Numbering
4. Click Save.
To hide or show the job log pane, click the toggle on the toolbar.
When the job log pane is hidden, the job list pane is maximized.
If you right-click a job in the upper pane, a context-sensitive menu is displayed that contains Abort,
Interrupt and Remove actions.
The Abort action stops the calculation immediately without any regard for available generated data.
Jobs terminated via this option will have a status of Cancelled.
The Interrupt action stops the calculation at the next point where data can be safely stored for later
use. When a job is interrupted, any available generated data will be retained. Jobs terminated via this
option will have a status of Finished.
To abort a job, right-click the job in the list view and select Abort.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 191
RSM Job Monitoring
To interrupt a job, right-click the job in the list view and select Interrupt.
Note:
The Abort and Interrupt actions are enabled for running jobs only.
To delete a job:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
192 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 10: RSM Cluster Load Monitoring
When RSM is integrated with a cluster using configurations defined in RSM, you can use the RSM Cluster
Load Monitoring application to retrieve and display the following information from the cluster:
• Available cores
• Available nodes
This information can help you monitor cluster usage, determine available resources, and troubleshoot
issues.
A refresh feature enables you to get up-to-the-minute status information when needed.
In this chapter:
10.1. Launching the RSM Cluster Load Monitoring Application
10.2. Monitoring a Cluster in the RSM Cluster Load Monitoring Application
10.3. Viewing Cluster Job Information
10.4. Controlling the Display of Cluster Monitors
To launch it, select Start > Ansys 2022 R1 > RSM Cluster Monitoring 2022 R1.
1. In the RSM Cluster Load Monitoring application, select the HPC configuration with which the
cluster is associated. For more information about HPC configurations, see RSM Configuration (p. 31).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 193
RSM Cluster Load Monitoring
When you select a configuration, information about the associated cluster's Submit host and HPC
type are displayed below the HPC configuration drop-down list.
2. To retrieve current information from the selected cluster, click on the toolbar, or click
The Overall Status section lets you know at a glance how many cores and nodes are currently available
for use in the cluster. Use the color legend to understand the meaning of colored sections in the dia-
grams.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
194 of ANSYS, Inc. and its subsidiaries and affiliates.
Viewing Cluster Job Information
The Nodal Heat Map helps you visualize the activity and availability on each node, by displaying the
number of cores being used by jobs versus the total number of cores available.
The Recent Jobs section displays a list of jobs that users have recently submitted to the cluster, and
details about each job. See Viewing Cluster Job Information (p. 195).
Note that for an Ansys RSM Cluster (ARC), all jobs are displayed in the Recent Jobs list. For other cluster
types, however, you may only see a partial listing of jobs if the cluster administrator has placed restrictions
on the types of jobs that you can view (for example, you may only be able to view your own jobs). If
only a partial job list is displayed, the information reported in the Overall Status and Nodal Heat Map
sections represents the resources being consumed by all jobs, not just the ones you see in the job list.
Note:
To minimize the querying of the HPC resource, the RSM Cluster Load Monitoring application
does not perform a continuous, real-time status refresh. To get up-to-the-minute status in-
formation, you must use the application's Refresh function.
Status Description
The job has been placed in the cluster queue,
Queued and is waiting to run.
The job is running.
Running
The job has been terminated via a cancel or
Abort action.
Cancelled
Also applies to jobs that have been aborted
because you exited a project without first
performing one of the following actions:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 195
RSM Cluster Load Monitoring
You can sort the job list by any other criterion (Cores, Queue, Owner, Status) by clicking on the
appropriate column header in the Recent Jobs table. Clicking on the same column header a second
time reverses the sort order of that column (from ascending to descending).
Note:
Performing a refresh of the RSM Cluster Load Monitoring window returns the table to
the default sort order (ascending Job Id), and removes any filters that you have applied.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
196 of ANSYS, Inc. and its subsidiaries and affiliates.
Viewing Cluster Job Information
To sort the column in either ascending or descending order, select either Sort Ascending or Sort
Descending in the drop-down menu. Selecting either of these options makes the current column
the one by which the table is sorted.
To return the table to the sort order that was in place prior to selecting a sort option, select Remove
Sort.
1. Click the drop-down arrow in the column that contains the desired filter criteria. For example, if
you only wanted to see jobs that used 3 or more cores, you would display the drop-down in the
Cores column.
2. In the drop-down, disable the criteria that you do not want to include in the job list display:
When filters have been applied to a column, a filter icon is displayed next to the column's drop-down
arrow to indicate that a filtered view is currently displayed:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 197
RSM Cluster Load Monitoring
Note that the filters will remain active even if you choose to sort the table by a different column.
However, filters will be cleared if you perform a Refresh.
To clear filters that you have applied to the job list, display the appropriate column drop-down and
click Clear.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
198 of ANSYS, Inc. and its subsidiaries and affiliates.
Chapter 11: RSM Troubleshooting
In this chapter:
11.1. Accessing RSM Log Files
11.2.Troubleshooting RSM-Related Issues
11.3.Troubleshooting Product-Related Issues
11.4. Known Issues and Limitations
Linux
/tmp
rsm_user- Windows Provides service information and
name_<timestamp>_pid.log errors from the RSM UserProxy
User %TEMP% process (Ans.Rsm.UPHost.exe). There
will be a UserProxy process for each
Linux user.
/tmp
ArcMaster221-<date>.log Windows When configuring an Ansys RSM
Cluster (ARC), this provides a
If running as a Windows transcript of what has occurred while
service: C:\Win- starting the ARC Master Service on
dows\Temp the submit host.
If not running as a
Windows service: %USER-
PROFILE%\App-
Data\Local\Temp
Linux
/tmp
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 199
RSM Troubleshooting
If not running as a
Windows service: %USER-
PROFILE%\App-
Data\Local\Temp
Linux
/tmp
Alternatively you can run generate_service_script with the -launcher command line option:
tools/linux> ./generate_service_script
Usage: generate_service_script -launcher
Options:
-launcher: Generate RSM Launcher service script.
Configuring RSM for Mapped Drives and Network Shares for Windows
If RSM is used to solve local or remote jobs on mapped network drives, you may need to modify security
settings to allow code to execute from those drives because code libraries may be copied to working
directories within the project.
You can modify these security settings from the command line using the CasPol utility, located under
the .NET Framework installation:
C:\Windows\Microsoft.NET\Framework64\v2.0.50727
In the example below, full trust is opened to files on a shared network drive to enable software to run
from that share:
C:\Windows\Microsoft.NET\Framework64\v2.0.50727\CasPol.exe
-q -machine -ag 1 -url "file://fileserver/sharename/*"
FullTrust -name "Shared Drive Work Dir"
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
200 of ANSYS, Inc. and its subsidiaries and affiliates.
Troubleshooting RSM-Related Issues
For more information on configuring RSM clients and cluster nodes using a network installation, refer
to Network Installation and Product Configuration.
Firewall Issues
Error: 'LauncherService at machine:9221 not reached'
1. Ensure that the launcher service is running. For instructions refer to Installing and Configuring the
RSM Launcher Service (p. 20).
2. If you have a local firewall turned on, you need to add port 9221 to the Exceptions List for the
launcher service (Ans.Rsm.Launcher.exe).
3. Allow a ping through the firewall (Echo Request - ICMPv4-In). Enable "File and Printer Sharing" in
firewall rules.
A user proxy process is created for every user who submits a job to RSM. Each user proxy process will
use a separate port chosen by RSM. By default, RSM will randomly select any port that is free. If you
want to control which ports RSM can choose, ensure that a range of ports are available for this purpose,
and specify the port range in the RSM application settings. See Specifying a Port Range for User Proxy
Processes (p. 146).
When the user proxy process is transferring files, a port is opened up for each file being transferred. If
you want to control which ports RSM can choose, ensure that a range of ports are available for this
purpose, and specify the port range in the RSM application settings. See Specifying a Port Range for
User Proxy Socket File Transfers (p. 146).
When a firewall is in place, traffic from the master node to the execution nodes (and vice versa) may
be blocked. To resolve this issue, you must enable ports on cluster nodes to allow incoming traffic, and
then tell each node what port to use when communicating with other nodes. For details see Dealing
with a Firewall in a Multi-Node Ansys RSM Cluster (ARC) (p. 103).
1. Open Control Panel > User Accounts > Change User Account Control settings.
2. On the User Account Control settings dialog box, use the slider to specify your UAC settings:
Note:
Disabling UAC can cause security issues, so check with your IT department before changing
UAC settings.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 201
RSM Troubleshooting
If you are not running a Microsoft HPC cluster, test the localhost configuration by opening a command
prompt and running the command, ping localhost. If you get an error instead of the IP address:
2. Verify that localhost is not commented out (with a # sign in front of the entry). If localhost is
commented out, remove the # sign.
Note:
If you are running on a Microsoft HPC cluster with Network Address Translation (NAT) enabled,
Microsoft has confirmed this to be a NAT issues and is working on a resolution.
Initially, the following error may be displayed in the RSM job report:
Job was not run on the cluster. Check the cluster logs and check if the
cluster is configured properly.
If you see this error, you will need to enable debug messages in the RSM job report in Workbench to
get more details about the failed job. Look for an error similar to the following:
259 5/18/2017 3:10:52 PM '\\jsmithPC\John-Share\WB\InitVal_pending\UDP-2'
260 5/18/2017 3:10:52 PM CMD.EXE was started with the above path as the current directory.
261 5/18/2017 3:10:52 PM UNC paths are not supported. Defaulting to Windows directory.
262 5/18/2017 3:10:52 PM 'clusterjob.bat' is not recognized as an internal or external command,
263 5/18/2017 3:10:52 PM operable program or batch file.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
202 of ANSYS, Inc. and its subsidiaries and affiliates.
Troubleshooting RSM-Related Issues
Alternatively, for a Microsoft HPC cluster, you can gather diagnostic information by running the HPC
Job Manager (supplied as part of the Microsoft HPC Pack), selecting the failed job, and examining the
output section of the job’s tasks.
Solution: Modify the registry on Windows compute nodes to enable the execution of commands via
UNC paths.
1. Create a text file of the following contents and save to a file (for example, commandpromp-
tUNC.reg).
Windows Registry Editor Version 5.00
[HKEY_LOCAL_MACHINE\Software\Microsoft\Command Processor]
"CompletionChar"=dword:00000009
"DefaultColor"=dword:00000000
"EnableExtensions"=dword:00000001
"DisableUNCCheck"=dword:00000001
For a Microsoft HPC cluster, the task of executing this on the compute nodes may be automated
using the clusrun utility that is part of the Microsoft HPC Pack installation.
This can occur if the file transfer method in the RSM configuration is set to No file transfer needed,
which requires that client files be located in a shared file system that is visible to all cluster nodes.
• Change the file transfer method to RSM internal file transfer mechanism, and enter the path of
the cluster staging directory.
OR
• Ensure that the cluster staging directory is visible to client machines, and that client working direct-
ories are created within the shared file system.
For information on file management options, see Specifying File Management Properties (p. 41).
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 203
RSM Troubleshooting
1. First, try to cancel the job using the arckill <jobId> command. See Cancelling a Job (arck-
ill) (p. 94).
2. If cancelling the job does not work, stop the ARC services, and then clear out the job database and
load database files on the Master node and the node(s) assigned to the stuck job. Delete the backups
of these databases as well.
On Linux, the database files are located in the service user's home directory. For example,
/home/rsmadmin/.ansys/v221/ARC.
Once the database files are deleted, restart the ARC services. The databases will be recreated auto-
matically.
Tip:
Clearing out the databases will fix almost any issue that you encounter with an Ansys
RSM Cluster. It is the equivalent of a reinstall.
Use the arcstatus (p. 93) command to view any errors related to the job (or check the ArcNode log).
You may see an error similar to the following:
2017-12-02 12:04:29 [WARN] System.ComponentModel.Win32Exception: The directory name is invalid
["\\MachineName\RSM_temp\tkdqfuro.4ef\clusterjob.bat"] (CreateProcessAsUser)
To resolve this issue you may need to open the network share of the cluster staging directory (\\Ma-
chineName\RSM_temp in the example) and grant Read/Write permissions on one of the following
accounts:
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
204 of ANSYS, Inc. and its subsidiaries and affiliates.
Troubleshooting RSM-Related Issues
2. Remove /tmp/UserProxyLauncherLock.lock.
3. Restart RSM.
At a minimum, an administrator should check to make sure that the RSM service user (for example,
rsmadmin) can see user proxy processes spawned for other users.
If this occurs, we recommend increasing the value of the User Proxy timeout setting to more than 10
minutes.
You will need to edit the following setting in the Ans.Rsm.AppSettings.config file:
<add key="IdleTimeoutMinutes" value="10" />
This setting appears twice in the configuration file. We recommend changing it in both places.
For more information see Editing RSM Application Settings (rsm.exe | rsmutils appsettings) (p. 145).
• ulimit -u (bash)
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 205
RSM Troubleshooting
• ulimit -H -u (bash)
You cannot increase the limit amount beyond this hard limit value. Contact your system administrator
to increase the hard limit.
To solve this issue, open the sge_pe file on the cluster and change the daemon_forks_slaves
parameter to FALSE.
CFD-Post Errors
If a project schematic contains CFD-Post, submitting a project or design point update to a Linux cluster
via RSM may result in errors, and the update of the CFD-Post results cell may fail completely.
When launching CFD-Post on a remote Linux machine, the DISPLAY variable must be either unset or
set to a valid X display before running in batch mode. For more information, see Running in Batch Mode
in the CFD-Post User's Guide.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
206 of ANSYS, Inc. and its subsidiaries and affiliates.
Known Issues and Limitations
To avoid this issue, ensure that any CFX coupling participant has its Solution | Update Option property
set to Run in Foreground before submitting a coupling job to RSM.
Once an update has been attempted with a CFX participant’s Solution | Update Option property set
to Run in Background, you may find that it is not possible to change the value of this setting for CFX.
To address this issue, first reset System Coupling’s Solution cell and then set CFX to perform updates
using Run in Foreground.
• When using an application's My Computer, Background setting to solve a SMART Crack Growth
problem on the local machine, the solution may fail as a result of a failed remeshing attempt. Installing
ARC services as daemons resolves this issue.
• Currently the RSM Job Monitoring application shows the jobs of the current user only. It does not
show the jobs of all users.
– Coupling jobs must be run in the foreground, so all coupling participants must be configured for
foreground solution updates.
– Coupling jobs that are submitted to Microsoft HPC 2012 or 2016 may fail if the cluster staging
directory is not shared from the head node.
– The coupling process may fail to start for coupling jobs that are submitted to Microsoft HPC Pack
2019.
• A solution submitted to RSM may fail to start if the name of a folder in the project path contains
special characters, or characters with special marks such as accents. Removing these characters from
the folder name resolves this issue.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 207
RSM Troubleshooting
• Currently, the RSM Cluster Load Monitoring application is supported on Windows only.
• When running remote design point updates through RSM to a Linux cluster, some jobs may be shown
as failed in the RSM Job Monitoring application, even though the design point calculation completed
and all results were retrieved. This will be accompanied with an error such as "[ERROR] FATAL UN-
HANDLED EXCEPTION: System.ExecutionEngineException: ExecutionContext_ExceptionInAsyncLocal-
Notification ---> System.Threading.ThreadAbortException" in the RSM job log. In this situation, the
results that have been retrieved are correct, and the failure status can be ignored.
• Job submission from Windows RSM clients to a Linux cluster may fail at the user authentication stage
if user accounts do not follow the account@machine format.
• Passwords that are cached by RSM and used to submit jobs to HPC resources cannot contain special
characters, or characters with special marks such as accents.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
208 of ANSYS, Inc. and its subsidiaries and affiliates.
Glossary
Ansys RSM Cluster (ARC) The built-in cluster type provided by RSM. An ARC cluster operates
in the same way that a commercial cluster does, running Ansys ap-
plications in local or distributed mode, but uses its own scheduling
capability rather than a third-party job scheduler.
client application A client application is the Ansys application run on the local RSM
client machine and is used to submit jobs to RSM. Examples include
Ansys Workbench, Ansys Fluent, Ansys CFX, and so on.
HPC queue An HPC queue determines the machine(s) on which jobs will run
when jobs are submitted to that queue. HPC queues are defined on
the HPC side (for example, on a cluster submit host), and can be
imported into the RSM Configuration application so that you can
map them to RSM queues when defining configurations.
HPC staging directory The HPC staging directory is the directory in which job input files
are placed by the client application when a job is submitted to RSM.
When defining a configuration in RSM, you specify whether the job
will execute in the HPC staging directory, or in a local scratch direct-
ory on the execution node(s). If you choose the former option, the
HPC staging directory will also serve as the job execution directory.
code template A code template is an XML file containing code files (for example,
C#, VB, JScript), references, and support files required by a job.
custom cluster integra- A custom cluster integration refers to the mechanism provided by
tion RSM that allows third parties to use custom scripts to perform the
tasks needed to integrate Ansys Workbench with the cluster. Both
client-side and cluster-side customizations are possible.
daemon services Daemon services are scripts or programs that run persistently in
the background of the machine, and which are usually executed at
startup. It is recommended that you install the RSM launcher service
as a daemon service. This allows the launcher service to be started
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 209
Glossary
head node The head node is the machine in a cluster that is configured as the
control center for communications between RSM and the cluster.
Typically it serves as the submit host and distributes jobs across the
cluster for execution.
job A job consists of a job template, a job script, and a processing task
submitted from a client application such as Ansys Workbench. An
example of a job is the update of a group of design points for an
Ansys Mechanical simulation.
job execution directory The job execution directory is the solver working directory. If you
specify that jobs will run in the HPC staging directory, the HPC sta-
ging directory will serve as the job execution directory. If you specify
that jobs will run in a local scratch directory on the execution node(s),
job input files will be transferred from the HPC staging directory to
the local scratch directory, and files generated by the job will be
transferred to the HPC staging directory so that client applications
can access them.
job template A job template is a component of an RSM job. It is an XML file that
specifies input and output files of the client application.
non-root privileges Non-root privileges give the user a limited subset of administrative
privileges. With RSM, non-root privileges are conferred by an
rsmadmin account (that is, membership to the rsmadmins user
group. It is recommended that non-root privileges are used for
starting and running the RSM launcher service.
parallel processing In parallel processing, jobs are executed on multiple CPU cores
simultaneously.
parallel environment (PE) A parallel environment allows for parallel execution of jobs. By
default, RSM is configured to support Shared Memory Parallel and
Distributed Parallel environments for SGE clusters.
PBS Pro Altair PBS Professional is a batch queuing system supported by RSM.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
210 of ANSYS, Inc. and its subsidiaries and affiliates.
queue A queue is a list of execution hosts that are suited to run a particular
class of jobs. When you submit a job to RSM, you submit it to an
RSM queue, which maps to an HPC queue. The HPC queue determ-
ines when and where the job will run based on resource requests
and current available resources. Queue definitions are part of the
configurations that are defined in RSM.
root privileges Root privileges give the user administrative access to all commands
and files on a Linux system. It is recommended that root privileges
are not used for starting and running the RSM launcher service.
RSM Admins group The RSM Admins group is a Windows user group that confers ad-
ministrative privileges for RSM. Also refers to the privileges conferred
on members of this group (that is, “RSM Admins privileges”).
RSM client The RSM client is the local machine from which RSM jobs are sub-
mitted to an HPC resource. It runs both RSM and a client application
such as Ansys Workbench.
RSM configuration A set of properties defined in RSM which specify information about
an HPC resource, and how RSM will communicate with that resource.
For example, it specifies the network name of a cluster submit host,
file transfer method to be used, and RSM queues. Configurations
are saved in .rsmcc files. If you store configurations in a shared
location, RSM users can retrieve them and use them on their own
machines.
RSM queue You define RSM queues when you define configurations in the RSM
Configuration application. When users submit jobs to RSM in client
applications, they submit them to RSM queues. Each RSM queue
maps to a specific RSM configuration and HPC queue. RSM queue
definitions are saved as .rsmq files.
rsmadmin user account An rsmadmin user account is a Linux account with membership in
the rsmadmins user group; as such, the account has RSM adminis-
trative privileges.
rsmadmins user group The rsmadmins user group is a Linux user group that confers ad-
ministrative privileges for RSM.
scratch directory Using a scratch directory is the practice of storing solver files in a
local directory on the execution node(s). Recommended to optimize
performance when there is a slow network connection between
execution nodes and the HPC staging directory, or when the solver
used produces many relatively large files.
serial processing In serial processing, jobs are executed on only one CPU core at a
time.
SGE Sun Grid Engine is not technically supported by RSM because UGE
is the latest version, though many SGE installations will still work
without modification. See UGE.
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
of ANSYS, Inc. and its subsidiaries and affiliates. 211
Glossary
submit host The submit host is the machine or cluster node that performs job
scheduling. In most cases, the cluster submit host is a remote ma-
chine, but it can also be your local machine ("localhost").
Release 2022 R1 - © ANSYS, Inc. All rights reserved. - Contains proprietary and confidential information
212 of ANSYS, Inc. and its subsidiaries and affiliates.