HPC Resources/InfraGRID

From VI-SEEM Wiki
Jump to: navigation, search

InfraGRID Cluster consists of 50 compute nodes powered by IBM BladeCenter-H technology. The entire solution is built up from 4 blade center chassis, each with 14 HS21 blade server. Each blade server has a dual quad core Intel Xeon E5504 CPU (clocked at 2.00Ghz) and 10GB RAM memory. The connectivity is delivered by: (a) Infiniband for interconnect and storage, (b) fiber channel for dedicated storage and (c) Gigabit for service networking. This cluster was initially setup in 2009. In 2011 the InfraGRID Cluster received an update by adding a new IBM BladeCenter-H chassis powered with 7 dual CPU/GPU HS22 blade servers. The CPU is powered by Intel XEON technology with a clock speed at 3.46Ghz and 32GB RAM memory. The GPU cards are NVidia Tesla M2070Q (448 GPU cores and 6GB GDDR5 RAM memory).

The cluster is managed by two service nodes, one dedicated to user access (also called head node) and one dedicated exclusively for service actions and cluster management. The storage is shared using GPFS file system over Infiniband using two dedicated NSD nodes. The service management is conducted using IBM BladeCenter the advanced management module (AMM) that is built in into both blade center chassis and blade servers (IMM). This feature allows remote administration and monitoring of any of the installed hardware. Infragrid Cluster is cooled with air. The displacement of the cooling units is in-row with a well delimited hot-cold area. Currently three APC InRow cooling units are installed and working in a cluster configuration to obtain cooling unit high-availability behavior.

Hpc-infragrid.jpg

Technical Specification

Administrative Data
Name InfraGRID Cluster
Short Description UVT HPC Center – InfraGRID Cluster
Owner Universitatea de Vest din Timisoara (UVT)
Country Romania
Computational Power
Number of servers 57
Server specification IBM BladeCenter HS21
CPU per server 2
RAM per server 10 GB
Total number of CPU-cores 400
Max number of parallel processes 800
Interconnect type QDR Infiniband
Interconnect latency 2.5 μs
Interconnect bandwidth 40 Gbps
Local filesystem type GPFS
Total storage (TB) 50
Accelerators type NVIDIA TESLA M2070Q
Number of cores 448
Accelerators per server 1
Servers equipped with accelerators 7
Peak performance CPU (Tflops) 3.5
Peak performance accelerators (Tflops) 3.5
Peak performance (Tflops) 2.11
Real performance (Tflops) 3.5
Operating system Red Hat Enterprise Linux
Version 6
Batch system/scheduler LoadLeveler
Development tools Intel Compilers (C/C++, FORTRAN), GNU Compilers, OpenMPI, CUDA, gdb, pgdbg
Libraries Intel MKL, HDF5, FFTW, NetCDF, GSL, LAPACK, Boost, BLAS
Applications Misc.
Dedication to VI-SEEM
CPU (percent) 10%
Storage (percent) 0%
Accelerators (percent) 20%
CPU (core-hours per year) 798,912
Storage in TB 0
Accelerators (hours per year) 5,494,272
Integration
System operational since Sep 2009
Available to the project from PM04
Expected date system to be phased out N/A
Interfaces SSH