21

Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled
Page 2: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

• Typically applied in clusters and grids• Loosely-coupled applications with sequential

jobs• Large amounts of computing for long periods of

times• Measured in operations per month or years

2

Page 3: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

• Bridge the gap between HPC and HTC• Applied in clusters, grids, and supercomputers• Loosely coupled apps with HPC orientations• Many activities coupled by file system ops• Many resources over short time periods

– Large number of tasks, large quantity of computing, and large volumes of data

[MTAGS08 Workshop] Workshop on Many-Task Computing on Grids and Supercomputers 2008[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”[MTAGS08] “Many-Task Computing for Grids and Supercomputers”

3

Page 4: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

4

Number of Tasks

Input Data Size

Hi

Med

Low1 1K 1M

HPC(Heroic

MPI Tasks)

HTC/MTC(Many Loosely Coupled Tasks)

MapReduce/MTC(Data Analysis,

Mining)

MTC(Big Data and Many Tasks)

[MTAGS08] “Many-Task Computing for Grids and Supercomputers”

Page 5: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

5

• Goal: enable the rapid and efficient execution of many independent jobs on large compute clusters

• Combines three components:– a streamlined task dispatcher– resource provisioning through multi-level scheduling

techniques– data diffusion and data-aware scheduling to leverage the

co-located computational and storage resources• Integration into Swift to leverage many applications

– Applications cover many domains: astronomy, astro-physics, medicine, chemistry, economics, climate modeling, etc

[SciDAC09] “Extreme-scale scripting: Opportunities for large task-parallel applications on petascale computers”[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”[Globus07] “Falkon: A Proposal for Project Globus Incubation”[SC07] “Falkon: a Fast and Light-weight tasK executiON framework”[SWF07] “Swift: Fast, Reliable, Loosely Coupled Parallel Computation”

Page 6: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

6

• Falkon is a real system– Late 2005: Initial prototype, AstroPortal– January 2007: Falkon v0– November 2007: Globus incubator project v0.1

• http://dev.globus.org/wiki/Incubator/Falkon

– February 2009: Globus incubator project v0.9• Implemented in Java (~20K lines of code) and C

(~1K lines of code)– Open source: svn co https://svn.globus.org/repos/falkon

• Source code contributors (beside myself)– Yong Zhao, Zhao Zhang, Ben Clifford, Mihael Hategan

[Globus07] “Falkon: A Proposal for Project Globus Incubation”[CLUSTER10] “Middleware Support for Many-Task Computing”

• Workload• 160K CPUs• 1M tasks• 60 sec per task

• 2 CPU years in 453 sec• Throughput: 2312 tasks/sec• 85% efficiency

Page 7: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

7[TPDS09] “Middleware Support for Many-Task Computing”, under preparation

Page 8: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

8[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”

Page 9: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

High-speed local disk

Falkon

9

Slower distributed storage

ZeptOS

[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”

Page 10: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

10

• Wide range of analyses– Testing, interactive analysis,

production runs– Data mining– Parameter studies

[SC07] “Falkon: a Fast and Light-weight tasK executiON framework”[SWF07] “Swift: Fast, Reliable, Loosely Coupled Parallel Computation”

Improvement:up to 90% lower end-to-end run time

Page 11: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

11

B. Berriman, J. Good (Caltech)J. Jacob, D. Katz (JPL)

[SC07] “Falkon: a Fast and Light-weight tasK executiON framework”[SWF07] “Swift: Fast, Reliable, Loosely Coupled Parallel Computation”

Improvement:up to 57% lower end-to-end run time

Within 4% of MPI

Page 12: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

• Determination of free energies in aqueous solution– Antechamber – coordinates– Charmm – solution– Charmm - free energy

12[NOVA08] “Realizing Fast, Scalable and Reliable Scientific Computations in Grid Environments”

Improvement:up to 88% lower end-to-end run time

5X more scalable

Page 13: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

13

• Classic benchmarks for MapReduce– Word Count– Sort

• Swift and Falkon performs similar or better than Hadoop (on 32 processors)

Sort

4285

733

25

83

512

1

10

100

1000

10000

10MB 100MB 1000MBData Size

Tim

e (s

ec)

Swift+Falkon

Hadoop

Word Count

221

11431795

863

46887860

1

10

100

1000

10000

75MB 350MB 703MBData Size

Tim

e (s

ec)

Swift+PBSHadoop

Page 14: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

14

0

500

1000

1500

2000

2500

3000

3500

4000

4500

5000

0

200000

400000

600000

800000

1000000

Thro

ughp

ut (t

asks

/sec

)

Task

s C

ompl

eted

Num

ber o

f Pro

cess

ors

Time (sec)

ProcessorsActive TasksTasks CompletedThroughput (tasks/sec)

• CPU Cores: 130816• Tasks: 1048576• Elapsed time: 2483 secs• CPU Years: 9.3

Speedup: 115168X (ideal 130816)Efficiency: 88%

[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”

Page 15: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

start

report

DOCK6Receptor

(1 per protein:defines pocket

to bind to)

ZINC3-D

structures

ligands complexes

NAB scriptparameters

(defines flexibleresidues, #MDsteps)

Amber Score:1. AmberizeLigand3. AmberizeComplex5. RunNABScript

end

BuildNABScript

NABScript

NABScript

Template

Amber prep:2. AmberizeReceptor4. perl: gen nabscript

FREDReceptor

(1 per protein:defines pocket

to bind to)

Manually prepDOCK6 rec file

Manually prepFRED rec file

1 protein(1MB)

6 GB2M

structures(6 GB)

DOCK6FRED ~4M x 60s x 1 cpu~60K cpu-hrs

Amber ~10K x 20m x 1 cpu~3K cpu-hrs

Select best ~500

~500 x 10hr x 100 cpu~500K cpu-hrsGCMC

PDBprotein

descriptions

Select best ~5KSelect best ~5K

For 1 target:4 million tasks

500,000 cpu-hrs(50 cpu-years)15

[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”

Page 16: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

16

CPU cores: 118784Tasks: 934803Elapsed time: 2.01 hoursCompute time: 21.43 CPU yearsAverage task time: 667 secRelative Efficiency: 99.7%(from 16 to 32 racks)Utilization: • Sustained: 99.6%• Overall: 78.3%

[SC08] “Towards Loosely-Coupled Programming on Petascale Systems”

Page 17: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

17

• Purpose– On-demand “stacks” of

random locations within ~10TB dataset

• Challenge– Processing Costs:

• O(100ms) per object

– Data Intensive: • 40MB:1sec

– Rapid access to 10-10K “random” files

– Time-varying load

AP SloanData

+

+++

+

+

=

+

Locality Number of Objects Number of Files1 111700 111700

1.38 154345 1116992 97999 490003 88857 296204 76575 191455 60590 1212010 46480 465020 40460 202530 23695 790

[DADC08] “Accelerating Large-scale Data Exploration through Data Diffusion”[TG06] “AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis”

Page 18: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

• AstroPortal– Makes it really easy for astronomers to create

stackings of objects from the Sloan Digital Sky Servey (SDSS) dataset

18

• Throughput– 10X higher than GPFS

• Reduced load– 1/10 of the original GPFS load

• Increased scalability• 8X

Page 19: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

• There is more to HPC than tightly coupled MPI, and more to HTC than embarrassingly parallel long jobs

• Data locality is critical at large-scale

19

Page 20: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

20

• Embarrassingly Happily parallel apps are trivial to run– Logistical problems can be tremendous

• Loosely coupled apps do not require “supercomputers”– Total computational requirements can be enormous– Individual tasks may be tightly coupled– Workloads frequently involve large amounts of I/O– Make use of idle resources from “supercomputers” via backfilling – Costs to run “supercomputers” per FLOP is among the best

• Loosely coupled apps do not require specialized system software– Their requirements on the job submission and storage systems can be extremely large

• Shared/parallel file systems are good for all applications– They don’t scale proportionally with the compute resources– Data intensive applications don’t perform and scale well– Growing compute/storage gap

Page 21: Typically applied in clusters and grids • Loosely-coupled ...cs.iit.edu/~iraicu/teaching/CS595-F10/lecture04.pdf• Applied in clusters, grids, and supercomputers • Loosely coupled

21