Upload
others
View
2
Download
0
Embed Size (px)
Citation preview
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation1
Stockholm Brain Institute Blue Gene/L
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation2
Stockholm Brain Institute Blue Gene/L
© 2002 IBM Corporation
IBM Systems & Technology Group and IBM Research
Copyright 2007– IBM Corporation
IBM® Blue Gene®/P - An Overview of a PetaflopCapable System
Carl G. TengwallIBM Storage & Technology [email protected]
4 October 18, 2007
Blue Gene Technology Roadmap
1GB VersionAvailable 1Q06
Blue GeneOn Demand
Center
Blue Gene/Q(Power architecture)Targeting 10 PFlops
Perf
orm
ance
2004 2011
Blue Gene/P(PPC 450 @ 850MHz)Scalable to 3 PFlops
2007
Blue Gene/L(PPC 440 @ 700MHz)
Scalable to 360 TFlops
Top500 list June 2007
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation5
Blue Gene/P Architectural Highlights
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation6
Blue Gene/P Architectural HighlightsLightweight kernel (CNK) on Compute NodesLinux on I/O Nodes handling syscallsOptimized MPI library for high speed messagingControl system on Service Node with private control networkCompilers and job launch on Front End Nodes
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation7
Blue Gene/P compared to Blue Gene/L
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation8
13.6 GF/s8 MB EDRAM
4 processors
1 chip, 20 DRAMs
13.6 GF/s2.0 (or 4.0) GB DDRSupports 4-way SMP
32 Node Cards1024 chips, 4096 procs
14 TF/s2 TB
72 Racks
1 PF/s144 TB
Cabled 8x8x16Rack
System
Compute Card
Chip
435 GF/s64 GB
(32 chips 4x4x2)32 compute, 0-1 IO cards
Node Card
Offering Schematic – Blue Gene/P
Front End Node / Service NodeJS21 / Power5Linux SLES10
Blue Gene/P continues Blue Gene’s leadership performance in a space-saving, power-efficient package for the most demanding and scalable
high-performance computing applications
HPC SW:CompilersGPFSESSLLoadleveler
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation9
Blue Gene/P Interconnection Networks3 Dimensional Torus
– Interconnects all compute nodes (73,728)– Virtual cut-through hardware routing– 3.4 Gb/s on all 12 node links (5.1 GB/s per node)– 0.5 µs latency between nearest neighbors, 5 µs to the farthest– MPI: 3 µs latency for one hop, 10 µs to the farthest– Communications backbone for computations– 1.7/3.9 TB/s bisection bandwidth, 188TB/s total bandwidth
Collective Network– One-to-all broadcast functionality– Reduction operations functionality– 6.8 Gb/s of bandwidth per link– Latency of one way tree traversal 1.3 µs, MPI 5 µs – ~62TB/s total binary tree bandwidth (72k machine)– Interconnects all compute and I/O nodes (1152)
Low Latency Global Barrier and Interrupt– Latency of one way to reach all 72K nodes 0.65 µs,
MPI 1.6 µs
Other networks– 10Gb Functional Ethernet– I/O nodes only– 1Gb Private Control Ethernet– Provides JTAG access to hardware. Accessible only from Service
Node system
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation10
Blue Gene/P System in a Complete Configuration
Complete Blue Gene/P Solution
BG/Pcompute
nodes1024 per Rack
BG/PI/O nodes
8 to 64per Rack
Fede
rate
d 10
Gig
abit
Eth
erne
t Sw
itch
Front-end nodes
Service node
WAN
visualization
archive
FS
1Gb/s - 4 per Rack
Control network
10Gb/s per I/ONode
1 to 7
2 Rac
ks
Storage Nodes File Storage
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation11
Blue Gene/P has wider rack dimensions and slightly longer vents than Blue Gene/L, but is otherwise similar.
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation12
BPC ASIC
4 cores, 8MB
Cu Heatsink
SDRAM-DDR2
1 of 20 sites.
DRAM address termination (Vtt),
EEPROM, decoupling, monitoring
13
PPC 450
FPUL1 PrefetchingL2
4MBeDRAM
L3
DDR-2Controller
PPC 450
FPUL1 PrefetchingL2
PPC 450
FPUL1 PrefetchingL2
4MBeDRAM
L3
DDR-2Controller
PPC 450
FPUL1 PrefetchingL2
Torus Collective BarrierJTAG 10Gb Ethernet
DMA
Blue Gene/P node
Arb
6*3.4Gb/sbidirectional
3*6.8Gb/sbidirectional
6*3.4Gb/sbidirectional
ControlNetwork
2*16B bus @435Mb/s
To 10Gbphysical layer
(Shares I/O with Torus)
Multiplexing sw
itchM
ultiplexing switch
13.6GB/s read(each), 13.6GB/s write(each)
Data read @ 6.8GB/sData write @ 6.8GB/sInstruction @ 6.8GB/s
13.6GB/s external DDR2 DRAM bus
4 symmetric ports for Tree, torus and global barriers
DMA module allowsRemote direct “put”/ “get”
14
Main Memory Bandwidth per Rack
0
2000
4000
6000
8000
10000
12000
14000
NASAItanium
Cray X1E
Cray XT4
ASCIPurple
BG/P NEC SX-8
15 October 18, 2007
0.230
0.350
0.0290.020 0.022 0.018
0.00
0.05
0.10
0.15
0.20
0.25
0.30
0.35
BG/L BG/P OpteronCluster
Cray RedStorm
Thunderbirddual-Xeon
Purple
System Power Efficiencygi
gaFL
OP/
s pe
r Wat
t
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation16
Blue Gene Software Hierarchical OrganizationCompute nodes dedicated to running user application, and almost nothing else - simple compute node kernel (CNK)I/O nodes run Linux and provide a more complete range of OS services – files, sockets, process launch, signaling, debugging, and terminationService node performs system management services (e.g., heart beating, monitoring errors) - transparent to application software
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation17
Supporting Equipment
Front End Nodes– IBM system-p 64-bit servers (not Blue Gene)
– Used for development of application code
– XL compilers, GNU compilers, libraries availableStorage Nodes
– A parallel filesystem is generally shared between to Blue Gene (the I/O nodes) and the Front End Nodes
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation18
Blue Gene SoftwareCompilersMessage Passing LibraryESSL & MASS LibrariesGPFS File SystemLoadLever schedulerHPC ToolkitManagement software on Management NodeCompute Node Kernel on Compute NodesLinux on I/O nodes
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation19
Materials Science
Climate Modeling
Genome Sequencing BiologicalModeling
Pandemic Research
Fluid Dynamics
Drug Discovery
Geophysical Data Processing
What is Blue Gene used for?
Financial Modeling
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation20
Applications already tested and running on BG/PBenchmarks
– Linpack, HPC Challenge– NAS Serial, NAS OpenMP, NAS Parallel– PALLAS, STREAM, MPPTest– UMT2K, SPPM, SPHOT
Applications– Physics: SPHOT, QCD, MILC, ICEPIC– Weather/Climate: HOMME, HYCOM, WRF, POP– Astrophysics: FLASH– CFD: Raptor, AVUS, CTH, AMR– Molecular Dynamics: NAMD, LAMMPS– Quantum Chemistry: CPMD, GAMESS– Materials Science: ParaTEC
All applications from BG/L will run on BG/P after recompile– General performance improvement of 2.4 going from BG/L to BG/P– Some applications have ratio >2.4 due to “superscaling”
• SMP mode with 4 threads on BG/P can achieve better performance than BG/L nodes
Many more applications in progress of being ported over and tested
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation21
Linpack on Blue Gene/P
ISC2007 Top 500 submission– 2 Racks, 1 MPI Process / Core, 8192 MPI Processes Total
• used basic NETLIB HPL with tweaks• used BG/L version of ESSL
– 20.86 TF, 74.89% of peak
– More performance to come
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation22
Scalability – Continues Extraordinary Blue Gene Behavior
LAMMPS Scalability
0
1
2
3
4
5
6
7
8
0 1000 2000 3000 4000 5000 6000 7000 8000 9000
CPU Count
Rela
tive
Perfo
rman
ce
HYCOM Scalability
0
1
2
3
4
5
6
7
0 500 1000 1500 2000 2500 3000 3500 4000 4500
CPU Count
Rel
ativ
e P
erfo
rman
ce
Two Examples– LAMMPS - molecular dynamics
– HYCOM - ocean modeling
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation23
Blue Gene is addressing six critical issues on the path to Petaflop computing
1. Power2. Floor space3. Cost4. Single processor performance5. Network scalability6. Reliability
Summary
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation24
Thank you!
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation25
Source: www.top500.org
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation26
IBM supercomputing leadership
Source: www.top500.org
Semiannual independent ranking of top 500
supercomputers in the world
IBM leads several key TOP500 metrics ...#1 System – LLNL – Blue Gene/L (280.6 TF)Most installed aggregate throughput with over 2,060 Teraflops (41.7%)Most in TOP10 with 6 systems (40%)Most in TOP100 systems with 45 (45%)Fastest machines in USA (BG/L)Fastest machine in Europe (MareNostrum)Fastest machine in China (Sinopec)
June 2007 Aggregate PerformanceIBM: 2,060 of 4,946 TF (41.9%)
IBM41,7%
SGI5,7%
HP24,2%
Other7,2%Dell
8,8%
Sun0,9%
Cray7,3%
Linux Networx
2,1%
NEC2,1%
June 2007 TOP100IBM is clear leader with 46%
IBM; 45
HP; 7 Dell; 13
Apple; 2
SGI; 9
Cray; 9
Other; 11Linux
Networx; 4
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation27
More Information
IBM Redbooks for Blue Gene ibm.com/redbooks
– Application Development Guide
– System Administration Guide
– Performance ToolsOpen Source Communities
IBM® System Blue Gene®/P Solution © 2007 IBM Corporation28
Read more about it…
http://www.ibm.com/servers/deepcomputing/bluegene.html
Sign up for newsletter
http://www-fp.mcs.anl.gov/bgconsortium/default.htm
Sign up for a consortium membership
http://www.top500.org
Latest TOP500 list
http://www.ibm.com/servers/deepcomputing/bluegene.htmlhttp://www-fp.mcs.anl.gov/bgconsortium/default.htmhttp://www.top500.org/
IBM® Blue Gene®/P - An Overview of a Petaflop Capable System Blue Gene Technology RoadmapBlue Gene/P Architectural HighlightsBlue Gene/P Architectural HighlightsBlue Gene/P compared to Blue Gene/LOffering Schematic – Blue Gene/PBlue Gene/P Interconnection NetworksBlue Gene/P System in a Complete ConfigurationBlue Gene/P has wider rack dimensions and slightly longer vents � than Blue Gene/L, but is otherwise similar. Main Memory Bandwidth per RackBlue Gene Software Hierarchical OrganizationSupporting EquipmentBlue Gene SoftwareApplications already tested and running on BG/PLinpack on Blue Gene/PScalability – Continues Extraordinary Blue Gene BehaviorBlue Gene is addressing six critical issues on the path to Petaflop computingThank you!More InformationRead more about it…