These settings are saved to The Preferences Dialog Box in the Multicore and Cluster Computing section.
|
After making these settings, click the Save as Default () button in the Settings window toolbar to save the current directory settings as the default preference.
|
•
|
The entry in the Host file field specifies the host file used for the job. If left empty, MPD looks for a file mpd.hosts in the Linux home directory.
|
•
|
Select which bootstrap server should be used by MPI using the Bootstrap server setting.
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
Select a Node granularity: Node (the default), Socket, or Core. Node allocates one process on each host, Socket allocates one process on each socket, and Core allocates one process on each core.
|
•
|
The Exclusive nodes check box is selected by default. Keep it selected if a compute node is used exclusively for a job — that is, no other jobs can interfere. Click to clear it if you want to run on nodes shared by other users.
|
•
|
Set the names of Requested nodes to use specified nodes only. The job scheduler only allocates jobs on the nodes listed by you.
|
•
|
Enter the Node group. As above, but can use node group names if defined by the system administrator. The job scheduler only allocates jobs on the nodes belonging to the group.
|
•
|
Enter the minimum required Cores per node — the number of concurrent threads per physical node. The default is 0, which means all cores. The job scheduler only allocates jobs to nodes with at least as many cores as set.
|
•
|
Enter the minimum required Memory per node (MB). The default is 0, which means no limit. The job scheduler only allocates jobs to nodes with at least as much memory as set.
|
•
|
Enter the Runtime (minutes) before the job is canceled. The default is Infinite; that is, the job is never canceled.
|
•
|
Select a Priority — Highest, Above normal, Normal (the default), Below normal, or Lowest — for the scheduled job based on the standard scheduler.
|
•
|
If you want to include scheduler arguments, add them to the Additional scheduler arguments field (for example, for mpiexec).
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
The entry in the Scheduler field is the IP address of the enterprise adapter of the head node or the DNS name of the head node. The default is localhost, and it should only be changed if you submit your job from a different computer than the head node.
|
•
|
The entry in the User field is the user account that COMSOL Multiphysics uses for submitting the job. You provide the password in a separate command window that opens at execution time with the possibility to save the credentials.
|
•
|
Select a Slot granularity — Host, Slot, or Manual — to specify if COMSOL Multiphysics should parallelize on the physical Host level or on the OGS/GE-allocated Slot level. For Host and Slot, specify the Number of slots to allocate. The Manual setting can be used to control the granularity more. In this case set the number of computational nodes to use in the Number of nodes. For Slot and Manual, the number of processes on each node is set in the Number of processes on host field; usually this is 1.
|
•
|
Enter the Number of processes on host to set the number of processes on each host.
|
•
|
The Sun Grid Engine priority is set in the Priority value field. The default is 0.
|
•
|
If you want to include scheduler arguments, add them to the Additional scheduler arguments field (for example, for mpiexec).
|
•
|
Select the Bootstrap server that should be used by MPI.
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
Enter the Queue name to set the name of the Sun Grid Engine.
|
•
|
The Exclusive nodes check box is selected by default. Keep it selected if a compute node is used exclusively for a job — that is, no other jobs can interfere. Click to clear it if you want to run on nodes shared by other users.
|
•
|
Set the names of Requested nodes to use specified nodes only. The job scheduler only allocates jobs on the nodes listed by you.
|
•
|
Enter the minimum required Memory per node (MB). The default is 0. The job scheduler only allocates jobs to nodes with at least as much memory as set.
|
•
|
Enter the Runtime (minutes) before the job is canceled. The default is Infinite; that is, the job is never canceled.
|
•
|
The SLURM priority value is set in the Priority value field. The default is 0.
|
•
|
If you want to include scheduler arguments, add them to the Additional scheduler arguments field (for example, for mpiexec).
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
Enter the Queue name to set the name of the SLURM job scheduler.
|
•
|
The entry in the Scheduler field is the IP address of the enterprise adapter of the head node or the DNS name of the head node. The default is localhost.
|
•
|
The entry in the User field is the user account that the COMSOL software uses for submitting the job. You provide the password in a separate command window that opens at execution time with the possibility to save the credentials.
|
•
|
Enter the minimum required Memory per node (MB). The default is 0. The job scheduler only allocates jobs to nodes with at least as much memory as set.
|
•
|
Enter the Runtime (minutes) before the job is canceled. The default is Infinite; that is, the job is never canceled.
|
•
|
The PBS priority value is set in the Priority value field. The default is 0.
|
•
|
If you want to include scheduler arguments, add them to the Additional scheduler arguments field (for example, for mpiexec). To designate nodes as exclusive, use the Additional scheduler arguments -n. This means that a compute node is used exclusively for a job — that is, no other jobs can interfere.
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
The entry in the Scheduler field is the IP address of the enterprise adapter of the head node or the DNS name of the head node. The default is localhost.
|
•
|
The entry in the User field is the user account that the COMSOL software uses for submitting the job. You provide the password in a separate command window that opens at execution time with the possibility to save the credentials.
|
•
|
If you want to include scheduler arguments, add them to the Additional scheduler arguments field (for example, for mpiexec). To designate nodes as exclusive, use the Additional scheduler arguments -n. This means that a compute node is used exclusively for a job — that is, no other jobs can interfere.
|
•
|
Specify the installation directory for the MPI in the MPI directory field. Click Browse to browse for a directory on the file system.
|
•
|
If you must provide extra arguments to MPI, use the Additional MPI arguments field.
|
•
|
The entry in the Scheduler field is the IP address of the enterprise adapter of the head node or the DNS name of the head node. The default is localhost.
|
•
|
The entry in the User field is the user account that the COMSOL software uses for submitting the job. You provide the password in a separate command window that opens at execution time with the possibility to save the credentials.
|
Micromixer — Cluster Version: Application Library path COMSOL_Multiphysics/Cluster_and_Batch_Tutorials/micromixer_cluster.
|