Accessing the cluster
The cluster can be accessed via SSH to one of the front end hosts, soemaster1.hpc.rutgers.edu
or soemaster2.hpc.rutgers.edu
, with your Engineering user account, same as on DSV/EIT lab computers. Both soemaster1.hpc.rutgers.edu
and soemaster2.hpc.rutgers.edu
can be acessed from the RU networks only. On the front end hosts, you can develop, compile, and debug your code, then submit it to the queue system, SLURM. The queue system will dispatch the job to run on the compute nodes, soenode03
, soenode04
, etc when the resources become available.
For faster data file transfer to and from the cluster, it is recommended to scp
or sftp
directly to the NFS file server, soenfs1.hpc.rutgers.edu
, bypassing the front end hosts. Your data will be available to all the cluster computers through the shared NFS file system, see the diagram below.
You can check your home directory quota usage by running command below on the NFS server: chkquota
However, there is no software available for code development on soenfs1
, so please use it only for transferring your data. There are shared file systems available on all the cluster computers, NFS with home directories, and a scratch space on Lustre. There is also a local /tmp
on each compute node. During the run time, please have your job to use either the local /tmp
or the Lustre scratch directory for efficient I/O. See the discussion on using the file systems on SOE cluster.