HPC/Carbon Cluster - Overview
< HPC
Jump to navigation
Jump to search
|
Primary Usage
- Modeling and Simulation
- CNM User Community
- CNM Theory and Modeling group
- Real-time on-demand data processing
- Nanoprobe Beamline
- E-beam lithography control
- Other high-data intensive instruments
Hardware
- 380 compute nodes, 8-12 cores/node (Xeon)
- total: 3200 cores, 8 TB RAM, 100 TB disk
- Infiniband interconnect
- Performance 32 TFLOPS (aggregrate)
- More details on a separate page
Software
- Redhat Enterprise Linux 5/CentOS 5
- Moab Cluster Suite
- GNU and Intel compilers