33
Jürgen Knobloch/CERN Slide 1 Grids for Science in Europe by Jürgen Knobloch CERN IT-Department Presented at Gridforum.nl Annual Business Day Eindhoven April 2, 2008 planning for sustainability

Gridforum Juergen Knobloch Grids For Science 20080402

  • Upload
    vrij

  • View
    3.110

  • Download
    0

Embed Size (px)

DESCRIPTION

Gridforum.nl Annual Business Day 2008

Citation preview

Page 1: Gridforum Juergen Knobloch Grids For Science 20080402

Jürgen Knobloch/CERN Slide 1

Grids for Science in Europe

by Jürgen Knobloch

CERN IT-Department

Presented at

Gridforum.nl

Annual Business Day

Eindhoven

April 2, 2008

planning for sustainability

Page 2: Gridforum Juergen Knobloch Grids For Science 20080402

Jürgen Knobloch- cern-it Slide-2

CERNCERNAnnual budget: ~1000 MSFr (~600 M€)Staff members: 2650Member states: 20

+ 270 Fellows, + 440 Associates

+ 8000 CERN users

Basic researchFundamental questionsHigh E accelerator:

Giant microscope (p=h/λ)Generate new particles (E=mc2)

Create Big Bang conditions

Page 3: Gridforum Juergen Knobloch Grids For Science 20080402

Large Hadron Collider - LHC

• 27 km circumference• Cost ~ 3000 M€ (+ detectors)• Proton-proton (or lead ion)

collisions at 7+7 TeV• Bunches 1011 protons cross

every 25 nsec• 600 million collisions/sec• Physics questions

– Origin of mass (Higgs?)– Dark matter?– Symmetry matter-antimatter– Forces – supersymmetry– Early universe – quark-gluon plasma– …

Jürgen Knobloch/CERN Slide 3

16 micron squeeze

1.9 K

10-9 to 10-12

Page 4: Gridforum Juergen Knobloch Grids For Science 20080402

Jürgen Knobloch/CERN Slide 4

LHC gets ready …

Page 5: Gridforum Juergen Knobloch Grids For Science 20080402

LHC Computing Challenge

• Signal/Noise 10-9

• Data volume– High rate * large number of

channels * 4 experiments

15 PetaBytes of new data each year

• Compute power– Event complexity * Nb. events *

thousands users

100 k CPUs (cores)

• Worldwide analysis & funding– Computing funding locally in

major regions & countries

– Efficient analysis everywhere

GRID technology

Jürgen Knobloch/CERN Slide 5

Page 6: Gridforum Juergen Knobloch Grids For Science 20080402

Requirements Match?

Jürgen Knobloch/CERN Slide 6

Tape & diskrequirements:

>10 times CERNpossibility

Substantial computing

requiredoutside CERN

Page 7: Gridforum Juergen Knobloch Grids For Science 20080402

Jürgen Knobloch/CERN Slide 7

Timeline - Grids

Data Challenges

First physicsCosmics

GRID 3

EGEE 1

LCG 1

Service Challenges

EU DataGrid

GriPhyN, iVDGL, PPDG

EGEE 2

OSG

LCG 2

EGEE 3

1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008

WLCG

Partially decentralizedmodel– replicate the event data at

about five regional centres

– data transfer via network ormovable media

RC2

CERN

RC1

CCRC08Common Computing Readiness Challenge

Inter-operation

Page 8: Gridforum Juergen Knobloch Grids For Science 20080402

WLCG Collaboration

• The Collaboration– 4 LHC experiments– ~250 computing centres– 12 large centres

(Tier-0, Tier-1)– 38 federations of smaller

“Tier-2” centres– Growing to ~40 countries– Grids: EGEE, OSG, Nordugrid

• Technical Design Reports– WLCG, 4 Experiments:

June 2005• Memorandum of

Understanding– Agreed in October 2005

• Resources– 5-year forward look

Jürgen Knobloch/CERN Slide 8

Page 9: Gridforum Juergen Knobloch Grids For Science 20080402

Events at LHC

Jürgen Knobloch/CERN Slide 9

Luminosity :1034cm-2 s-1

40 MHz – every 25 ns

20 events overlaying

Page 10: Gridforum Juergen Knobloch Grids For Science 20080402

Trigger & Data Acquisition

Jürgen Knobloch/CERN Slide 10

Page 11: Gridforum Juergen Knobloch Grids For Science 20080402

Data Recording

Jürgen Knobloch/CERN Slide 11

1.25 GB/sec (Ions)

Page 12: Gridforum Juergen Knobloch Grids For Science 20080402

Tier-0, Tier-1, Tier-2

Jürgen Knobloch/CERN Slide 12

Tier-0 (CERN):•Data recording•First-pass reconstruction

•Data distributionTier-1 (11 centres):•Permanent storage•Re-processing•AnalysisTier-2 (>250 centres):• Simulation• End-user analysis

2 GB/secData Transfer out of Tier-0

Note: For a site to be reliable, many things have to work simultaneously!

Target

Page 13: Gridforum Juergen Knobloch Grids For Science 20080402

Middleware

• Security – Virtual Organization Management (VOMS) – MyProxy

• Data management – File catalogue (LFC)– File transfer service (FTS)– Storage Element (SE)– Storage Resource Management (SRM)

• Job management – Work Load Management System(WMS)– Logging and Bookeeping (LB)– Computing Element (CE)– Worker Nodes (WN)

• Information System– Monitoring: BDII (Berkeley Database Information Index), RGMA

(Relational Grid Monitoring Architecture) aggregate service information from multiple Grid sites, now moved to SAM (Site Availability Monitoring)

– Monitoring & visualization (Gridview, Dashboard, Gridmap etc.)

Jürgen Knobloch/CERN Slide 13

Page 14: Gridforum Juergen Knobloch Grids For Science 20080402

Increasing workloads

Factor 3 capacity ramp-up is in progress for LHC operation

>⅓non-LHC

Page 15: Gridforum Juergen Knobloch Grids For Science 20080402

Many Grid Applications

At present there are about 20 applications from more than 10 domains on the EGEE Grid infrastructure– Astronomy & Astrophysics - MAGIC, Planck – Computational Chemistry – Earth Sciences - Earth Observation, Solid Earth Physics,

Hydrology, Climate – Fusion – High Energy Physics - 4 LHC experiments (ALICE, ATLAS, CMS,

LHCb) BaBar, CDF, DØ, ZEUS – Life Sciences - Bioinformatics (Drug Discovery, GPS@,

Xmipp_MLrefine, etc.) – Condensed Matter Physics– Computational Fluid Dynamics – Computer Science/Tools – Civil Protection – Finance (through the Industry Task Force)

Jürgen Knobloch/CERN Slide 15

Page 16: Gridforum Juergen Knobloch Grids For Science 20080402

Grid Applications

Jürgen Knobloch/CERN Slide 16

Medical SeismologyChemistry

Astronomy

Fusion Particle Physics

Page 17: Gridforum Juergen Knobloch Grids For Science 20080402

Jürgen Knobloch/CERN Slide 17

Tier-1 Centers: TRIUMF (Canada); GridKA(Germany); IN2P3 (France); CNAF (Italy); SARA/NIKHEF (NL); Nordic Data Grid Facility (NDGF); ASCC (Taipei); RAL (UK); BNL (US); FNAL (US); PIC (Spain)

Computingat the

Terra-Scale

Page 18: Gridforum Juergen Knobloch Grids For Science 20080402

Evolution

Jürgen Knobloch/CERN Slide 18

National

Global

European e-Infrastructure

Page 19: Gridforum Juergen Knobloch Grids For Science 20080402

Sustainability

Jürgen Knobloch/CERN Slide 19

Page 20: Gridforum Juergen Knobloch Grids For Science 20080402

European Grid Initiative

Goal:• Long-term sustainability of grid infrastructures

in Europe

Approach:• Establishment of a new federated model

bringing together NGIs to build the EGI Organisation

EGI Organisation:• Coordination and operation of a common multi-

national, multi-disciplinary Grid infrastructure– To enable and support international Grid-based

collaboration– To provide support and added value to NGIs– To liaise with corresponding infrastructures outside Europe

www.eu-egi.org

20

Page 21: Gridforum Juergen Knobloch Grids For Science 20080402

EGI_DS Consortium members

• Johannes Kepler Universität Linz (GUP)• Greek Research and Technology Network S.A.

(GRNET)• Istituto Nazionale di Fisica Nucleare (INFN)• CSC - Scientific Computing Ltd. (CSC)• CESNET, z.s.p.o. (CESNET)• European Organization for Nuclear Research

(CERN)• Verein zur Förderung eines Deutschen

Forschungsnetzes - DFN-Verein (DFN)• Science & Technology Facilities Council (STFC)• Centre National de la Recherche Scientifique

(CNRS)

Jürgen Knobloch/CERN Slide 21

Page 22: Gridforum Juergen Knobloch Grids For Science 20080402

EGI – European Grid Initiative

• EGI Design Study started Sept 07 to establish a sustainable pan-European grid infrastructure after the end of EGEE-3 in 2010

• The main foundations of EGI are 37 National Grid Initiatives (NGIs)

• Project is funded by the European Commission's 7th Framework Program

Jürgen Knobloch/CERN Slide 22

www.eu-egi.orgwww.eu-egi.org

Page 23: Gridforum Juergen Knobloch Grids For Science 20080402

EGI_DS Schedule

Jürgen Knobloch/CERN Slide 23

Duration 27 months:

5 6 7 8 9 10 11 12 1 2 3 4 5 6 7 8 9 10 11 12 1 2 3 4 5 6 7 8 9 10 11 12 1 2 3 4 5 6 7 8 9 10

Develop EGI Proposal NGIs signing Proposal

Start of EGEE-III

Final Draft of EGI Blueprint Proposal

EGI Blueprint Proposal

EGEE-III transition to EGI-like structure

EGI Entity in place

EU Call Deadline for EGI Proposal

Submission of EGEE-III

Start of EGI Design Study

2008 2009 2010

EGEE-II (2YEARS) EGEE-III (2YEARS) EGI operational

Page 24: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Functionality Overview

• Management, Outreach & Dissemination - Representation of EU Grid Efforts

• Operations & Resource Provisioning & Security

• Application Support & Training• Middleware (Build&Test, Component

Selection/Validation/Deployment)• Standardisation & Policies• Industry take-up

Jürgen Knobloch/CERN Slide 24

Page 25: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Management Structure

Jürgen Knobloch/CERN Slide 25

www.eu-egi.org

EGI

EGI Organisation

Director + staff

EGI Council

CTO (Development)

COO (Operations)

Political Bodies such as:•eIRG,•EU,

•ESFRI,•Ntl. Bodies,

•...Advisory

Committees

Dev.Group

Oper.Group

Adm.+PRGroup

Strategy Committee

CAO(Admin.+ PR)

Page 26: Gridforum Juergen Knobloch Grids For Science 20080402

Characteristics of NGIs

Each NGI … should be a recognized national body with a

single point-of-contact … should mobilise national funding and

resources… should ensure an operational national grid

infrastructure… should support user communities

(application independent, and open to new user communities and resource providers)

… should contribute and adhere to international standards and policies

Responsibilities between NGIs and EGI are split to be federated and complementary

Jürgen Knobloch/CERN Slide 26

Page 27: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Operations

• NGIs perform operations of their national grid infrastructure, including monitoring and accounting, following a general SLA

• Each NGI has full responsibility towards other NGIs but acts independently

• EGI provides coordination level to guarantee effective, integrated operations through the adoption of standard services

Jürgen Knobloch/CERN Slide 27

Page 28: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Resource Strategy

• EGI owns very few hardware resources• Resource provisioning is responsibility

of NGIs Freedom to choose whatever resources to offer

• Development of a strategy for simple entry of new Virtual Organizations

Jürgen Knobloch/CERN Slide 28

Page 29: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Middleware

• Convergence of middleware stacks towards full interoperability (through general standards interfaces)

• Distinction between core and high-level functions EGI coordinates core services– Standards-compliance test for core services

provided by EGI– Common build&test system needed in Europe

Jürgen Knobloch/CERN Slide 29

Page 30: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Application Support

• Application support implemented and provided by NGIs

• EGI performs coordination role– Validation of user support by NGIs– Facilitating experts between NGIs– Organization of user forums

• EGI to coordinate training events

Jürgen Knobloch/CERN Slide 30

Page 31: Gridforum Juergen Knobloch Grids For Science 20080402

EGI Funding

• The EGI organisation has three major and separate cost centres:– general central services

funded through contributions of the NGIs– service provisioning

funded through service charges– Developments

funded through project grants

• There are in general no cross-subsidies possible between these cost centres.

Jürgen Knobloch/CERN Slide 31

Page 32: Gridforum Juergen Knobloch Grids For Science 20080402

Issues being discussed

• Small vs. large central EGI• Roles “core EGI” vs. NGI• Transition from a (EU) project-funded

grid to self-sustained operations• Maintaining the available expertise• Choice of legal structure and location

Jürgen Knobloch/CERN Slide 32

Page 33: Gridforum Juergen Knobloch Grids For Science 20080402

EGI – European Grid Initiative

• Future EGI Organisation = “Glue” between various grid communities in Europe and beyond

• EGI_DS defines required mechanisms and functionalities of the EGI Organisation

Towards a sustainable environment for the application communities utilizing grid infrastructures for their everyday work

Jürgen Knobloch/CERN Slide 33