SOE HPC Cluster


Due to a planned power outage in Hill Center on Saturday, 1/27/2018, the SOE HPC will be brought down on Friday night at around 6 pm. During the down time, technical services and upgrades will be done to some compute nodes. The cluster will resume operation on Monday, 1/29/2018.

The operating system on SOE HPC cluster is being upgraded to Ubuntu 16.04. The updated nodes can be accessed through Access to the old part of the cluster, soemaster1.hpc is closed now.

Hardware Architecture

The cluster hardware is based on Intel Sandy Bridge 2670 CPUs, 16 cores and 128 GB of RAM per node. 

There are 60 nodes in the cluster available for general use. There are also 30 nodes dedicated to Engineering research groups.

The nodes are interconnected over FDR infiniband and Gbit networks.

Each of the two interactive front-end machines have two Nvidia Kepler K20 GPU cards installed.

Software Installed 

At the moment, the following apps have been installed and tested on the

  •   PGI 13.7 C/C++/Fortran compilers
  •   GNU GCC/G++/gfortran 4.7.3, 4.9.3 compilers
  •   Open MPI (compiled with GNU and PGI, and Intel compilers)
  •   LAMMPS
  •   OpenFoam
  •   Matlab
  •   CUDA


The front end hosts and the storage server on the cluster can be accessed via SSH from the RU networks. See the details here.

The cluster resources are managed by SLURM queue system. Users will need to submit all their applications through batch scripts to run on the cluster. Here is a reference to the most common SLURM commands.

The environment for the applications, such as path to the binaries and shared libraries as well as the licenses, is loaded/unloaded through the environment modules. For example, to see what the modules are available on the cluster, run command  module avail, to load a module, say matlab, run command   module load matlab. More information on using the environment modules can be found at this link.

There are three file systems available for computations: local /tmp on the nodes, Lustre, and NFS. Both Lustre and NFS are shared file systems and provide the same file system image to all the computational nodes. If you run a serial or parallel computational job, utilizing only one node, always use the local /tmp file system. For multi-node MPI runs, you need a shared file system, so please use Lustre. Avoid using NFS as its I/O saturates quickly at multiple parallel runs and becomes extremely slow. For details on using the file systems on the cluster, follow this link

Account on the Cluster

Who is eligible for having an account: Engineering faculty, postdocs, graduate students

Account Requirements: EIT/DSV Account

Application for SOE HPC Cluster Account: (NOTE: Must be on a Rutgers-networked computer/VPN to access! ) Undergraduate and graduate students should have their faculty advisor to send a confirmation e-mail to Alexei about requested computing resources. 

Those with accounts may access the SOE HPC cluster via SSH to and

ACCESS NOTE: The two aforementioned hosts and the application URL are only accessible from WITHIN Rutgers' networks, so you must be logged in on a Rutgers computer physically or remotely via SSH or VPN to access these hosts.

© 2011 Rutgers, The State University of New Jersey - This site is XHTML 1.0 valid.