HPC/Carbon Cluster - Overview: Difference between revisions
< HPC
Jump to navigation
Jump to search
mNo edit summary |
mNo edit summary |
||
Line 13: | Line 13: | ||
=== Hardware === | === Hardware === | ||
[[Image:HPC Compute Rack-up.png|noframe|right|200px|]] | [[Image:HPC Compute Rack-up.png|noframe|right|200px|]] | ||
* 380 compute nodes, 8 cores/node (Xeon) | * 380 compute nodes, 8-12 cores/node (Xeon) | ||
* total: | * total: 3200 cores, 8 TB RAM, 100 TB disk | ||
* Infiniband interconnect | * Infiniband interconnect | ||
* Performance | * Performance 32 TFLOPS (aggregrate) | ||
* More [[HPC/Hardware Details|details on a separate page]] | * More [[HPC/Hardware Details|details on a separate page]] | ||
Revision as of 15:13, March 20, 2014
|
Primary Usage
- Modeling and Simulation
- CNM User Community
- CNM Theory and Modeling group
- Real-time on-demand data processing
- Nanoprobe Beamline
- E-beam lithography control
- Other high-data intensive instruments
Hardware
- 380 compute nodes, 8-12 cores/node (Xeon)
- total: 3200 cores, 8 TB RAM, 100 TB disk
- Infiniband interconnect
- Performance 32 TFLOPS (aggregrate)
- More details on a separate page
Software
- Redhat Enterprise Linux 5/CentOS 5
- Moab Cluster Suite
- GNU and Intel compilers