39
CLUSTER COMPUTING

cluster computing

Embed Size (px)

DESCRIPTION

this is a full pledge presentation of clusting computing.

Citation preview

  • 1. Cluster Computers in view2

2. ABSTRACTVery often applications need more computing power than a sequentialcomputer can provide. One way of overcoming this limitation is to improvethe operating speed of processors and other components so that they can offerthe power required by computationally intensive applications. Even thoughthis is currently possible to certain extent, future improvements areconstrained by the speed of light, thermodynamic laws, and the high financialcosts for processor fabrication. A viable and cost-effective alternative solutionis to connect multiple processors together and coordinate their computationalefforts. The resulting systems are popularly known as parallel computers, andthey allow the sharing of a computational task among multiple processors. 3. CLUSTERS HAS ARRIVED 4. IntroductionHistoryArchitectureData CommunicationBenefitsCluster ManagementTypes of ClustersApplications 5 5. A computer cluster consists of a set of loosely connected or tightlyconnected computers that work together so that in many respects theycan be viewed as a single system.Cluster consists of: Nodes(master+computing) Network OS Cluster middleware: It permitscompute clustering programsto be portable to a wide varietyof clusters.Cluster Middle wareCPU CPU CPU6High Speed Local NetworkCluster 6. INTRODUCTION A computer cluster is a group of tightly coupled or looselyconnected computers that work together closely so that it can beviewed as a single computer. Clusters are commonly connected through fast LAN(local areanetworks). Clusters are usually deployed to improve speed and/or reliabilityover that provided by a single computer, while typically beingmuch more cost effective than single computer the of comparablespeed or reliability. 7. CONT. Cluster computing can also be used as a relatively low-cost form ofparallel processing for scientific and other applications that lendthemselves to parallel operations. Consists of many of the same or similar type of machines. All machines share resources. They must trust each other so that does not require a password,otherwise you would need to do a manual start on each machine. 8. CONT In cluster computing each node within a cluster is an independentsystem, with its own operating system, private memory, and, in somecases, its own file system. Because the processors on one nodecannot directly access the memory on the other nodes, programs orsoftware run on clusters usually employ a procedure called "messagepassing" to get data and execution code from one node to another. Clusters have evolved to support applications ranging fromecommerce, to high performance database applications. 9. HISTORY Customers invented clusters, as soon as they could not fit all theirwork on one computer, or needed a backup. The first commercial clustering product was ARC net, developed byData point in 1977. Digital Equipment Corporation released their VAX cluster product in1984 for the VAX/VMS operating system. The ARCnet and VAX cluster products not only supported parallelcomputing, but also shared file systems and peripheral devices. 10. CONT.. Microsoft, Sun Microsystems, and other leading hardware andsoftware companies offer clustering packages. The idea was to provide the advantages of parallel processing,while maintaining data reliability and uniqueness. 11. Beowulf clusterStart from 1994Donald Becker of NASA assembled this cluster.Also called Beowulf clusterApplications like data mining, simulations, parallel processing,weather modeling, etc 12. A Beowulf Cluster is a computer design that uses parallel processing acrossmultiple computers to create cheap and powerful supercomputers. A BeowulfCluster in practice is usually a collection of generic computers connectedthrough an internal network. A cluster has two types of computers, a master computer, and nodecomputers. When a large problem or set of data is given to a Beowulf cluster, the mastercomputer first runs a program that breaks the problem into small discretepieces; it then sends a piece to each node to compute. As nodes finish theirtasks, the master computer continually sends more pieces to them until theentire problem has been computed.13 13. 14( Ethernet,Myrinet.)+ (MPI) Master: or service node or front node ( used to interact with users and manage thecluster ) Nodes : a group of computers (computing node s)( keyboard, mouse, floppy,video) Communications between nodes on an interconnect network platform ( Ethernet,Myrinet.) In order for the master and node computers to communicate, some sort messagepassing control structure is required. MPI,(Message Passing Interface) is the mostcommonly used such control. 14. ARCHITECTURE OF BEOWOLFCLUSTER 15. IBM hidro Clusters16 16. ARCHITECTURE OF CLUSTERA cluster consists of a collection of interconnected stand-alonecomputers cooperatively working together as single, integratedcomputing resource.A node:a single or multiprocessor system with memory, I/O facilities, &OSgenerally two or more computers (nodes) connected togetherin a single cabinet, or physically separated & connected via a LANappear as a single system to users and applicationsprovide a cost-effective way to gain features and benefits 17. ARCHITECTURE OF CLUSTER 18. SINGLE SYSTEM IMAGESingle Entry PointSingle File HierarchySingle Control PointSingle Job-Management SystemSingle User InterfaceSingle I/O SpaceSingle Memory spaceProcess migrationCheckpointing 19. DATA COMMUNICATIONParallel Virtual Machine (PVM)Message Passing Interface (MPI) 20. Parallel Virtual Machine(PVM) PVM was developed at the Oak RidgeNational Laboratory around 1989. It provides a set of software libraries that allowa computing node to act as a "parallel virtualmachine". It provides run-time environment for message-passing,task and resource management, andfault notification and must be directly installedon every cluster node. PVM can be used by user programs writtenin C, C++, or Fortran, etc. 21. Message Passing Interface(de facto standard)Message passing , an inherent element of all computer clusters,to coordinate the activities of the many nodes theyencompass. Unlike PVM,MPI is a specification rather than a specific set oflibraries. The specification emerged in the early 1990 with the initialeffort having been supported by ARPA and National ScienceFoundation. MPI implementations typically use TCP/IP and socketconnections. MPI is now a widely available communications model thatenables parallel programs to be written in languages suchas C, Fortran , Python, etc. 22. MIDDLEWARE Middleware allows programmers to build applications that notonly look and feel the same on various Pcs but thatapplications, that also use the same method to access dataregardless of the location of the data. This are certain standard programming interface and protocolsthat view that entire cluster as a single system image 23. Absolute Scalability:-A cluster can have dozens or evenhundreds of machines, each of which is a multi-processor.Incremental Scalability:-A cluster in configured in such a way thatit is possible to add new systems to the cluster in smallincrements. 24. High Availability:-Single point of failure can be eliminated, ifany one system component goes down, the system as awhole stay highly available.Performance @ Low Cost:-Reason for the growth in use of cluster isthat they have significantly reduced the cost ofprocessing power. 25. Cluster ManagementTask scheduling a challenge Done to share system resources effectivelyNode failure management1) failover2) failbackLoad Balancing 26. FAILURE MANAGEMENT : failoverNODE 1NODE 2NODE 3NODE 4CLUSTERNODEFAILEDREDUNDANTNODE 5NODE IN USE 27. FAILURE MANAGEMENT:failbackCLUSTERNODE 1NODE 2NODE 3NODE 4FIXED ANDRESTOREDNODEFAILURENODE 5NODE 28. LOAD BALANCINGNODE 1NODE 5 NODE 4NODE 3NODE 2OCVLEURLSTOEARD LOADBALANCED 29. TYPES OF CLUSTERSHigh Performance Computing (HPC) ClusterHigh Availability (HA) ClusterVisualization Cluster 30. HPC Clusters HPCC make up more than 80% of the 500 fastestcomputers in the world.HPC clusters pass on a massive cost savings benefit overSymmetric multi-processing.Example : Processing and analyses of massive datasets. 31. VISUALIZATION CLUSTERA visualization cluster is an HPC cluster with theaddition of powerful graphics cards, normallydesigned to work in sync with each other to tacklehigh-resolution and real-time simulations.Used in :1) weather forecasting,2) 3D modelling , etc. 32. HA clusters High-availability clusters (or failover clusters) - groups of computers thatsupport server applications that can be reliably utilized with a minimum ofdown-time. On failure of a node , a redundant node is used. without clustering ,what happens ? HA clustering remedies this situation by detecting hardware/software faults,and immediately restarting the application on another system without requiringadministrative intervention, a process known as failover. As part of thisprocess, clustering software may configure the node before starting theapplication on it. 33. Cluster-based systems can be used toexecute many Internet applications:Web servers;Search engines;Email; Security; Database servers , etc. 34. Other Applications:weather modeling, automobile crash simulations,Image processing,nuclear simulations,electromagnetics, etc. 35. AUTOMOBILE TESTING :SIMULATION 36. Any Questions??