Upload
benedict-nash
View
213
Download
0
Embed Size (px)
Citation preview
Advanced User Support
Amit Majumdar5/7/09
OutlineThree categories of AUS
Update on Operational Activities
AUS.ASTA
AUS.ASP
AUS.ASEOT
Three Categories of AUS
I. Advanced Support for TeraGrid Applications (AUS.ASTA) AUS staff work on a particular user’s code (usually) Guided/initiated by the allocations process
I. Advanced Support for Projects (AUS.ASP) Project initiated by AUS staff (jointly with users) – impacts many users
I. Advanced Support for EOT (AUS.ASEOT) Advanced HPC/CI part of EOT
Update on Operational Activities Bi-weekly telecon among AUS POCs from every RP site
o Matching of AUS staff to ASTA projects, discussion about ASP, EOT activities
Web/Tele-con among AUS technical staffo Biweekly technical tele/web-conference on ASTA and other projects - using
readytalko 13 presentation sessions (including today), 22 technical presentations
ASTAs Started in October 2008 - TRAC
# PI name Field AUS staff
1 Jordan/USC Geo sciences
Cui (SDSC ), Urbanic (PSC ), Kim(TACC ), Wong (NICS), Bock (NCSA)
2 Mahesh/U. Minn CFD Tatineni (SDSC), Crosby (NICS),Cazes (TACC )
3 Roitberg/U. Florida Biochem Walker (SDSC ), Pamidighantam (NCSA ), Halloy (NICS )
4 Voth/U. Utah Bioechem Blood (PSC ), Crosby (NICS), Peterson (TACC )
5 Buehler/MIT Civil Engr Walker (SDSC)
6 Durisen/IU Astro Henschel, Berry (IU)
7 Papavassiliou/ OU CFD Reddy (PSC )
8 Van Einde/UCSD Str. Engr Choi (SDSC), Halloy (NICS), Cazes (TACC)
Startup/Supplemental ASTAs# PI Name Field AUS Staff
1 Schulten/UIUC Bio Chem Koesterke , Liu , Milfeld, TACC
2 Ferrante/U. Washington CFD/DNS/turbulence Taha (NCSA), Pekurovsky (SDSC,), Bock (NCSA), Peterson (TACC), Crosby (NICS)
3 Tabakin/U. Pitt Quantum Computing Gomez ( PSC)
4 Karimabadi/ UCSD Space Physics Tatineni (SDSC)
5 Mashayek/UIC CFD Taha (NCSA)
6 Malamataris/GM CFD Brook (NICS)
7 Hu/CSULB Physics Barth (TACC)
8 Yeung/Gtech CFD/DNS Wong(NICS), Pekurovsky (SDSC)
9 Finol/CMU CFD Jana (PSC)
10 Geoffrey/CMU CS Jana (PSC)
11,12,13 Roberts/RTI;Karniadakis/Brown;Coveney/UCL
CS/CFD/Chemistry Karonis (NIU, ANL)
14 Cobb/ORNL Neutron Science Lynch, Chen (ORNL)
15 Kanamitsu/SIO-UCSD Climate Science Peterson (TACC)
16 Burke/Pitt Medicine Brown (PSC)
ASTAs Started in January 2009 - TRAC
# PI Name Field AUS staff
1 Aidun/GT CFD Dillman (Purdue), Khurram (NCSA)
2 Cheatham/U.Utah Biochem Pamidighantam (NCSA), Crosby (NICS), Milfeld (TACC), Madrid (PSC), Walker (SDSC)
3 Gygi/UCD Materials Science Fahey (NICS), Kim (TACC)
4 Jansen/RPI CFD Wong (NICS), Peterson (TACC), O’neal (PSC)
5 Van de Walle/UCSB
Materials Science Liu (TACC), Vanomoer (NCSA)
6 Yip/MIT Materials Science Halloy (Yip)
7 Scheraga/Cornell Chemistry Blood, Mahmoodi (PSC)
8 Oleynik/U. South Florida
Materials Science Barth (TACC), Crosby (NICS), Jundt (LONI)
ASTAs Started in April 2009 - TRAC# PI Name Field AUS staff
1 Batista/LANL Materials Science Liu (TACC)
2 Schnetter/LSU Astrophysics Mahmoodi (PSC), Kufrin, Liu (NCSA), Pfeiffer (SDSC)
3 Helly/SIO-UCSD Climate Science Cheeseman (Purdue), Jundt (LONI)
4 Liu/UKY QCD Reddy (PSC), Fahey (NICS)
5 Smith/Gtech CFD TBD upon PI’s response
6 Wheeler/UT CFD Khamra (TACC)
7 Cen/Princeton Astrophysics Brook (NICS), Kharma(TACC), Chourasia (SDSC)
8 Latour/Clemson BioChem TBD upon PI’s response
9 Apte/Oregon St. CFD Rosales (TACC)
10 Sandholm/CPU CS Welling (PSC)
Total number of active ASTAs ~40
ASTA Update - PI: Durisen, IU, AstrophysicsAUS staff: Henschel, Berry (IU)
• Legacy code – OpenMP parallel – benchmarked on three Altix systems (PSC, NCSA, ZIH) – different performance!
• Optimized subroutines (calculates the gravitational potential) – 1.8X speedup
• Simulations generate several TBs of data which is then analyzed interactively using IDL
• Transferring this amount of data via traditional methods (ftp, scp, etc.) to IU is extremely time consuming and tedious
• By mounting the Data Capacitor at PSC on Pople the user can write their data directly to IU and then access from servers in their department
• Extensive profiling and optimization of I/O performance on local scratch vs DataCapacitor – eventually ~30% speedup in I/O
• Files appear locally as they are generated by simulation
ASTA Update - PI: Scheraga, Cornell, BiophysicsAUS Staff : Blood, Mahmoodi (PSC)
• Identified a serial bottleneck and load imbalance problem that limited parallel scaling
• Eliminated serial bottleneck and restructured code to eliminate imbalance
• Resulting code performing 4 times faster for large systems
• Never ending optimization– In the new code computation/communication balance is
changed – further profiling ongoing using CrayPAT, TAU
ASTA Update - PI: Van de Walle, UCSB, Condensed Matter Physics
AUS Staff : Liu (TACC), Vanmoer (NCSA)
• Main effort was to identify performance issues of VASP• Identified only routine that needed lower level (-O1)
compilation; others used –O3 : resulted in ~10 % performance improvement
• MKL on Ranger had SMP enabled with default OMP_NUM_THREADS of 4; caused overhead – fixed with proper wayness and threads
• Proper setting of NPAR (determines process grouping for band diagonalization and FFT) showed 3-4 times speedup
Advanced Support Projects• Two projects ongoing
– Benchmarking Molecular Dynamics codes– Benchmarking Materials Science codes
• Other potential ones we are looking into– Multi-core performance analysis– Usage-based perf/profiling tools
Comparison of MD Benchmark on TeraGrid Machines at Different Parallel Efficiencies
Advanced Support EOT
• Advanced HPC classes at various RP sites
• TG09– AUS staff participating in organizing TG09;
reviewing papers– AUS staff will be presenting papers at TG09;
presenting tutorials– Joint US-AUS-XS working group meeting at TG09