Sabine is the latest addition to the RCDC shared campus resource pool housing public CPU and GPU nodes with shared access to storage resources. It hosts a total of 5704 CPU cores in 169 compute and 12 GPU nodes.
Sabine is operated under the new RCDC policy. It is housed in the RCDC (Research Computing Data Core), and went into production mid January 2018. A second phase has been added in Spetember of 2018. If you plan to use this system, please make sure your PI has been granted an allocation and you have requested an account on the system. Please refer to the User Guide to find out about specifics of running jobs, selecting resources and available software on this cluster.
Theoretical peak performance is approximately 600 TFlops.
|Node Type||CPU Type||CPU Socket Count||Total Cores||Memory||Disk Space||Node Count|
|Login HP DL 380||Intel Xeon E5-2680v4||2||28||128GB||36TB||1|
|Compute ProLiant HPE XL170r||Intel Xeon E5-2680v4||2||28||256GB||1 TB||68|
|Compute ProLiant HPE XL170r||Intel Xeon E5-2680v4||2||28||128GB||1 TB||48|
|GPU Accelarator HPE ProLiant XL190r||Intel Xeon E5-2680v4
|Compute HPE Proliant XL170r||Intel Xeon G6148||2||40||192GB||1 TB||52|
|GPU Accelerator HPE Proliant XL270d||Intel Xeon E5-2680v4
|Large Memory HPE DL 360||Intel Xeon G6148||2||40||768GB||4TB||1|
Interconnect: Sabine nodes are connected via Intel OmniPath switch with a 100Gb Line Rate.
Storage: Sabine has a shared ~250TB Lustre file storage and ~725TB NFS storage.