How to run the thermal solver on clusters?

This article explains two different methods on how to run the thermal solver on clusters for thermal-flow and multiphysics analyses.

Introduction

Running jobs on a cluster provides enhanced processing power, efficient resource management, parallel processing capabilities, scalability, remote access, and monitoring. The thermal solver uses Distributed Memory Parallel (DMP) implementation to run jobs in parallel on multiple machines. There are two methods to run the thermal solver on cluster:
  • Using a script that executes the thermal solver in parallel mode.
  • Using the dedicated ND argument in the TMG Executive Menu, specifically designed for cluster execution.

Running through the custom script

The most common method to run the thermal solver is through a custom script tailored to the specific scheduler. Typically, you submit a script requesting specific resources. Once the job begins execution, it launches the script and identifies the nodes and CPUs allocated for the job by the job manager. This information is then used to generate the parallel configuration file with the correct cluster nodes. Finally, you launch the thermal solver from the command line in the script using the input file and the updated parallel configuration file.

Running through the dedicated ND argument in theTMG Executive Menu

The second method allows you to run jobs through the TMG Executive Menu. In this mode, the monitor submits a series of cluster jobs that execute sequentially for each solver module. Additionally, parallel modules such as VUFAC and ANALYZ can be executed as separate parallel jobs.

The primary difference between the two methods lies in how resources are allocated and the parallel configuration file is formed. In the first method, the custom script handles resource allocation and forms the parallel configuration file, all within a single job. In contrast, when using the ND argument, the thermal solver creates the configuration file with the node names and launches separate cluster jobs for each TMG module. Running a simulation through the custom script offers more flexibility, however requires more scripting skills.

Note:
You cannot run multiphysics thermal-structural analyses using the TMG Executive Menu with the ND argument in parallel on cluster.

With the Simcenter 3D Thermal/Flow DMP license, you manage DMP processing using a process scheduler, such as Sun Grid Engine Software (SGE), or Load Sharing Facility (LSF), which organize the queues and resources to perform the analysis in parallel.