Upload
dominick-harris
View
215
Download
0
Embed Size (px)
Citation preview
Background Information
Recent trends in the marketplace of high performance computing
Strohmaier, Dongarra, Meuer, Simon
Parallel Computing 2005
Short history of parallel machines
• 1970s: vector computers• 1990s: Massively Parallel Processors (MPPs)
– Standard microprocessors, special network and I/O
• 2000s: – Cluster computers (using standard PCs)– Advanced architectures (BlueGene)– Comeback of vector computer
(Japanese Earth Simulator)– IBM Cell/BE
• 2010s:– Multi-cores, GPUs– Cloud data centers
Clusters
• Cluster computing– Standard PCs/workstations connected by fast network– Good price/performance ratio– Exploit existing (idle) machines or use (new) dedicated
machines
• Cluster computers vs. supercomputers (MPPs)– Processing power similar: based on microprocessors– Communication performance was the key difference– Modern networks have bridged this gap
• (Myrinet, Infiniband, 10G Ethernet)
Overview
• Cluster computers at our department– DAS-1: 128-node Pentium-Pro / Myrinet cluster (gone)– DAS-2: 72-node dual-Pentium-III / Myrinet-2000 cluster– DAS-3: 85-node dual-core dual Opteron / Myrinet-10G– DAS-4: 72-node cluster with accelerators (GPUs etc.)
• Part of a wide-area system:– Distributed ASCI Supercomputer
DAS-2 Cluster (2002-2006)
• 72 nodes, each with 2 CPUs (144 CPUs in total)
• 1 GHz Pentium-III• 1 GB memory per node• 20 GB disk• Fast Ethernet 100 Mbit/s• Myrinet-2000 2 Gbit/s (crossbar)• Operating system: Red Hat Linux• Part of wide-area DAS-2 system
(5 clusters with 200 nodes in total)Myrinet switch
Ethernet switch
DAS-3 Cluster (Sept. 2006)
• 85 nodes, each with 2 dual-core CPUs(340 cores in total)
• 2.4 GHz AMD Opterons (64 bit)• 4 GB memory per node• 250 GB disk• Gigabit Ethernet • Myrinet-10G 10 Gb/s (crossbar)• Operating system: Scientific Linux• Part of wide-area DAS-3 system (5 clusters; 263 nodes), using
SURFnet-6 optical network with 40-80 Gb/s wide-area links
DAS-3 NetworksNortel 5530 + 3 * 5510ethernet switch
85 compute nodes
85 * 1 Gb/s ethernet
Myri-10G switch
85 * 10 Gb/s Myrinet
10 Gb/s ethernet blade
8 * 10 Gb/s eth (fiber)
Nortel OME 6500with DWDM blade
80 Gb/s DWDMSURFnet6
1 or 10 Gb/s Campus uplink
Headnode(10 TB mass storage)
10 Gb/s Myrinet
10 Gb/s ethernet
DAS-4 (Sept. 2010)• 72 nodes (2 quad-core Intel Westmere Xeon E5620,
24 GB memory, 2 TB disk)
• 2 fat nodes with 94 GB memory
• Infiniband network + 1 Gb/s Ethernet
• 16 NVIDIA GTX 480 graphics accelerators (GPUs)
• 2 Tesla C2050 GPUs
DAS-4 performance
• Infiniband network:
• - One-way latency: 1.9 microseconds
• - Throughput: 22 Gbit/s
• CPU performance:
• - 72 nodes (576 cores): 4399.0 GFLOPS
Blue Gene/L
2.8/5.6 GF/s4 MB
2 processors
2 chips, 1x2x1
5.6/11.2 GF/s1.0 GB
(32 chips 4x4x2)16 compute, 0-2 IO cards
90/180 GF/s16 GB
32 Node Cards
2.8/5.6 TF/s512 GB
64 Racks, 64x32x32
180/360 TF/s32 TB
Rack
System
Node Card
Compute Card
Chip
Blue Gene/L Networks3 Dimensional Torus
– Interconnects all compute nodes (65,536)
– Virtual cut-through hardware routing
– 1.4Gb/s on all 12 node links (2.1 GB/s per node)
– 1 µs latency between nearest neighbors, 5 µs to the farthest
– Communications backbone for computations
– 0.7/1.4 TB/s bisection bandwidth, 68TB/s total bandwidth
Global Collective
– One-to-all broadcast functionality
– Reduction operations functionality
– 2.8 Gb/s of bandwidth per link
– Latency of one way traversal 2.5 µs
– Interconnects all compute and I/O nodes (1024)
Low Latency Global Barrier and Interrupt
– Latency of round trip 1.3 µs
Ethernet
– Incorporated into every node ASIC
– Active in the I/O nodes (1:8-64)
– All external comm. (file I/O, control, user interaction, etc.)
Control Network