RESOURCES

c3ddb03

C3DDB Compute, Storage, and Networking Resources

The C3DDB compute cluster currently consists of 133 compute nodes with a total of 7200 cores 61 terabytes of main memory, an FDR infiniband interconnect, a 2 Pebibyte (2.24 Petabyte) Lustre file system, and a 54 Terabyte NAS.

Compute:

  • 100 compute nodes, each consisting of a Dell PowerEdge R715 chassis with:
    Four 16-Core AMD Abu Dhabi Opteron processors (AMD 6376) running at 2.3Ghz
    256GB main memory
    500GB scratch storage
  • 32 compute nodes, each consisting of a an HP DL580 chassis with:
    Two 10-Core / 20 Thread Intel Ivy Bridge processors (Intel E7-4830v2) running at 2.2Ghz
    1 Terabyte main memory
    3 X 3 x nvidia K40c GPU accelerators
  • 1 large-memory compute node consisting of an SGI UV2000 with:
    Twenty 8-core Intel Ivy Bridge Processors (Intel E5-4650 v2) running at 2.4GHz
    Four Terabytes of main memory
  • Master and gateway nodes to support login, management, and network access
  • License server
  • All of the nodes in the system run RHEL 6.5, and are managed by a SLURM resource manager.

Storage:

  • 2.24 Petabytes (2 Pebibytes) of storage for scratch and project directories, managed by a high performance LUSTRE file system.
  • 54 Terabytes of NAS storage for home directories, supported by snapshot and backup services.
  • dbGap-compliant directories for sensitive data.

Network:

  • For ordinary use:
      • The C3DDB is accessible via SSH, SFTP, and SCP from anywhere on the public internet.
  • For high speed data transfers:
    • All computing systems at the MGHPCC, including the C3DDB, are accessible via 10Gbps links to the five MGHPCC member universities.
    • The C3DDB is also accessible via the Northern Crossroads node at the MGHPCC.
    • A Globus Connect Data Transfer Node is also available for those who need a highly optimized data transfer path.