28
1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

Embed Size (px)

Citation preview

Page 1: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

1

Computing platformAndrew A. Chien

Mohsen Saneei

University of Tehran

Page 2: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

2

outline

Basic Elements– Computing Elements– Communication Elements– Storage Elements

Simple Composite Elements (SCE): Local Grids– High Throughput SCEs– High Reliability SCEs– Dedicated High Performance SCEs– Shared Controllable Performance SCEs

Illinois HPVM Project and Similar Efforts

Page 3: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

3

Basic ElementsComputing Elements

Gordon Moore’s low: the number of transistor and performance of chip, will be double every tow years.

Transistor = 20 x 2(yr-1965)/1.5

Example:– 1975: intel 8080 with 4500 transistors– 1998: pentium II with 7,500,000 tran.

Performance: increases by 1.5 times each year– In 1975: 0.1 MIPS– In 1998: 1,000 MIPS

Page 4: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

4

Basic Elements Communication Elements

Networking element with terabit performance is available today

But networks had slower advances:– Cost– Fundamental nature

Over the past few years we have seen a rapid advance from 10 Mb/s networks to 100 Mb/s

We focus on local networks (Cluster networks)

Page 5: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

5

Basic Elements Communication Elements (cluster networks)

cluster networks:– Physically localized– High speed– Generally low-volume products

Some cluster networks:– Myricom’s Myrinet– Compaq/Tandem’s Servernet

Servernet I Servernet II

Page 6: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

6

Basic Elements Communication Elements (cluster networks)

Myrinet:– High speed local network– Full duplex 1.28 Gb/s links– Derived from multicomputer routers– Wormhole routing– Switch latency below 1 µs– Myrinet open new horizon for research and produced

messaging layersFast messageActive message

Page 7: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

7

Basic Elements Communication Elements (cluster networks)

0102030405060708090

band

wid

th (M

B/s

)

Message size (byte) in Fast Message on Myrinet

Page 8: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

8

Basic Elements Communication Elements (cluster networks)

Compaq/Tandem’s Servernet I– Full duplex 50 MB/s– Wormhole routing– Reliable communication– 64 byte packets– A few microsecond latency

Compaq/Tandem’s Servernet II– Full duplex 125 MB/s– 512 byte packet size– 64 bit network addressing

Page 9: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

9

Basic Elements Storage Elements

Capacity and cost-performance improved at an exponential rate

Density:– In 1970 to 1988: 29% every year– After 1988: 60% every year

Cost-per-byte:– In 1970 to 1988: 40% every year– After 1988: 100% every year

But seek times are improving very slowly (average seek timea of 7-10 ms remain typical)

Page 10: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

10

Basic Elementsfuture

Machine Computing Memory Disk Network

2003 PC 8 GIPS 1 GB 128 GB 1 Gb/s

Supercomputer 80 TIPS 10 TB 1,280 TB 10 Tb/s

2008 PC 64 GIPS 16 GB 2 TB 10 Gb/s

Supercomputer 640 TIPS 160 TB 20,000 TB 100 Tb/s

Page 11: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

11

Simple Composite Elements (SCE): Local Grids

SCEs: collections of basic elements, aggregated with software and special hardware.

They are often single administrative domains. SCEs are study for these reasons:

– They can reduce the number of problems higher-level grids must solve.

– SCEs use resources and software to implement the external properties

– SCEs form the basis for the larger computational grid

Page 12: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

12

Simple Composite Elements (SCE): Local Grids (cont.)

A national computational grid:– Reliable CEs to

management (access control & scheduling) and basic services (naming and routing)

– Other CEs to resource pools (data caching, storage, …)

Page 13: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

13

Simple Composite Elements (SCE): Local Grids (cont.)

SCEs are defined by:– Their external interface– Their internal hardware requirements– Their ability to deliver efficient and flexible use of the

hardware to application Their external interface:

– Capacity – Aggregate performance– Reliability – Predictability– sharability

Page 14: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

14

Simple Composite Elements (SCE): Local Grids (cont.)

hardware requirements– Heterogeneity– Network requirements (special hardware, link

length limited, bandwidth, …)– Distributed resources (links tens of meters or

thousands of kilometers)– Changes in constituent system– Scalability (number of nodes)

Page 15: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

15

SCE: Local Grids High Throughput SCEs

Pooled resources are utilized to achieve high throughput on a set of sequential compute jobs– Example: Condor, Utopia , Symbio

External interface:– high capacity for computation and a sharable resource– Interface for some parallel computing such as PVM

are available

Page 16: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

16

SCE: Local Grids High Throughput SCEs (cont.)

Hardware requirements:– Running on a wide range of processor and network

environment – Tolerating both processor and network heterogeneity

in type and speed– Can scale to larger number of processors (hundreds

to thousands)

Page 17: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

17

SCE: Local Grids High Throughput SCEs (cont.)

High Throughput SCEs in grids– Flexible and powerful systems for achieving high

throughput on large numbers of sequential jobs.– Thus, they are well-matched grid elements for such

tasks.

No supported :– Aggregate performance– Reliability (partial reliable)– Predictability

Page 18: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

18

SCE: Local Grids High Reliability SCEs (reliable clusters)

Provide computational resources with extremely low probability of service interruption and data loss

Limited scalability is used to increase system capacity

Used of sharable resource Prefer compatible hardware to enable

failover and data sharing

Page 19: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

19

SCE: Local Grids High Reliability SCEs (reliable clusters) (cont.)

Can used of lower performance standby system to reduce cost

Can be physically localized or distributed over a wide area network

Traditionally used special operating system

Page 20: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

20

SCE: Local Grids Dedicated High Performance SCEs

Merge basic element into a single resource, to be applied to a single computation.

Used of collection of microprocessors or entire systems (scalable network of workstation)

Initially applied to supercomputers tasks Scalable to connect hundreds or thousands of

node with limited physical extent (tens of meters)

Page 21: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

21

SCE: Local Grids Dedicated High Performance SCEs (cont.)

Predominant programming model: message passing (such as MPI)

Support both sequential jobs and parallel computation but focus on highest single-job performance

don’t support: – Reliability – Predictability– Sharable

Page 22: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

22

SCE: Local Grids Dedicated High Performance SCEs (cont.)

Berkeley Network Of Workstations (NOW) is a project o the Dedicated High Performance SCEs

IBM SP-2 and Intel/Sandia are 2 example:– Use high volume microprocessors as their

basic computation engines– Use custom high performance interconnect

delivering 5-100 MB/s of network bandwidth to each node

Latencies of 20-100 µs

Page 23: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

23

SCE: Local Grids Dedicated High Performance SCEs (cont.)

IBM SP-2– Employed entire workstation as the basic building

block– Standard AIX workstation operating system– Allowing a single job to each node

Intel/Sandia– Employed special system boards and packaging as

the basic building block– A custom operating system PUMA– Multitasking and virtual memory are not provided on

the compute node

Page 24: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

24

SCE: Local Grids Shared Controllable-Performance SCEs

Aim: deliver predictable high performance in a shared-resource, heterogeneous, distributed environment.

This SCEs combine the capabilities of all the SCEs except reliability.

This SCEs are High-Performance Virtual Machine (HPVM)

Page 25: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

25

SCE: Local Grids Shared Controllable-Performance SCEs (cont.)

HPVM simplifies programming task by allowing programmers to focus on the complexity of the application

Construction of effective HPVM requires meeting a number of research challenges in:– High performance predictable communication– Management of heterogeneity– Performance models– Adaptive resource management

Page 26: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

26

SCE: Local Grids Shared Controllable-Performance SCEs (cont.)

To achieve efficient tight coupling, the network hardware will need to support both low latency and high bandwidth.

It most be scalable to thousands of node, because HPVM execute on distributed resources

Physically limited geographic distribution

Page 27: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

27

SCE: Local Grids Illinois HPVM Project and Similar Efforts

Aim: develop shared controllable-high-performance SCEs.

Basic parameters :– Computing nodes: x86 and PCI computing systems– Operating systems: Windows NT and Linux– Networks: Myrinet, servernet, …

Real World Computing Project (RWCP) & Berkeley Network Of Workstations II (NOW II) are another project on HPVM

Page 28: 1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran

28

summary

Element type Scalable Aggregatable Reliable Predictable Sharable

Basicelements

Basic compute No -- No No Yes

Basic storage No -- No No yes

Basic network No -- No No yes

Local

Grids

(SCE)

High Throughput yes No partial No yes

High Reliability Limited No yes No yes

Dedicated High Performance

yes Yes No No No

Shared Controllable Performance

yes yes No yes yes