Upload
basil-rodgers
View
217
Download
0
Tags:
Embed Size (px)
Citation preview
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Beyond GridPP2
Tony Doyle
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Outline
• Beyond GridPP2: what are the estimated resource requirements in the LHC exploitation era?
• Background:• PMB preliminary discussions in September• PPAP presentation in October:• Resources needed in medium-long term?
• (09/07-08/10) Exploitation medium• (09/10-08/14) Exploitation long-term• Focus on resources needed in 2008
• GridPP Oversight Committee outline on Monday• initial ideas for discussion here..
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Tier 0 and LCG: Foundation
Programme
• Aim: build upon Phase 1
• Ensure development programmes are linked
• Project management:GridPP LCG
• Shared expertise:
• LCG establishes the global computing infrastructure
• Allows all participating physicists to exploit LHC data
• Earmarked UK funding being reviewed
Required Foundation: LCG Deployment
F. LHC Computing Grid Project (LCG Phase 2) [review]
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Tier 0 and LCG: RRB meeting
(October)
• Jos Engelen proposal to RRB members (Richard Wade [UK]) on how a 20MCHF shortfall for LCG phase II can be funded
• Spain to fund ~2 staff. Others at this level? • Funding from UK (£1m), France, Germany
and Italy for 5 staff. Others?• Now vitally important that the LCG effort
established predominantly via UK funding (40%) is sustained at this level (~10%)
• Proposal to SC in preparation
Value to the UK? Required Foundation: LCG Deployment
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Grid and e-Science Support
in 2008What areas require support?IV Running the Tier-1 Data CentreIV Hardware annual upgrade
IV Contribution to Tier-2 Sysman effort (non-PPARC) hardware IV Frontend Tier-2 hardware
IV Contribution to Tier-0 support
III One M/S/N expert in each of 6 areasIII Production manager and four Tier-2 coordinators
II Application/Grid experts (UK support)
I ATLAS Computing MoU commitments and supportI CMS Computing MoU commitments and supportI LHCb Core Tasks and Computing SupportI ALICE Computing supportI Future experiments adopt e-Infrastructure methods
• No GridPP management: (assume production mode established + devolved management to Institutes)
III. Grid Middleware
I. Experiment Layer
II. Application Middleware
IV. Facilities and Fabrics
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
1/Jan/2005
£3.65m
£5.67m
£3.65m
£2.05m£1.87m
CERN
DataGrid
Tier - 1/A
ApplicationsOperations
PPARC Financial Input:
GridPP1 Outturn
LHC Computing Grid Project (LCG)Applications, Fabrics, Technology and Deployment
European DataGrid (EDG)Middleware Development
UK Tier-1/A Regional CentreHardware and Manpower
Grid Application DevelopmentLHC and US Experiments + Lattice QCD
Management Travel etc
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Jan 2005
£0.81m
£2.77m
£2.86m
£0.82m
£0.66m
£2.99m
£2.70m
£1.00m
£2.40m
Tier-1/AHardware
Tier-2Operations
Applications
M/S/N
LCG-2
MgrTravel
Ops
Tier-1/AOperations
C. Grid Application DevelopmentLHC and US Experiments + Lattice QCD + Phenomenology
B. Middleware Security NetworkDevelopment
F. LHC Computing Grid Project (LCG Phase 2) [review]
E. Tier-1/A Deployment:Hardware, System Management, Experiment Support
A. Management, Travel, Operations
D. Tier-2 Deployment: 4 Regional Centres - M/S/N support and System Management
PPARC Financial Input:
GridPP2 Components
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
IV. Hardware Support
UK Tier-1 2008
CPU Total (MSI2k) (7.8) 5.2
Disk Total (PB) (3.8) 1.6
Total Tape (PB) (2.3) 1.6
UK Tier-2 2008
CPU Total (MSI2k)
8.0
Disk Total (PB) 1.0
ALICE ATLAS CMS LHCb Total
CPU (MSI2K) 9.1 16.6 12.6 9.5 47.8 41.4 -13%
Disk (PBytes) 3.0 9.2 8.7 1.3 22.2 10.1 -55%
Tape (PBytes) 3.6 6 6.6 0.4 16.6 17.8 7%
Notes 1. 2.
Total resources required and planned in all Tier-1 Centres (except CERN)
Requirements will be reviewed by LHCC in January 2005Current planning includes estimates of resources for which funding has not yet been secured
Total resources planned at Tier-1 centres (note 2)
BalanceResource type
First full year of data taking (2008)
All data is preliminary
Estimated requirements (note 1)
1. Between (October) and December UK Tier-1 LHC estimates reduced (see Dave’s talk): now more realistic
2. Global shortfall of Tier-1 CPU was (-13%) and Disk (-55%) in October
3. UK Tier-1 estimated input in December now corresponds to ~20% (~7%) of global disk (CPU)
4. LCG MoU commitments required by April 2005
5. UK Tier-2 CPU and disk resources significant
6. Rapid physics analysis turnaround is a necessity
7. Priority is to ensure that ALL required software (experiment, middleware, OS) is routinely deployed on this hardware well before 2008
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
III. Middleware, Security and
Network
M/S/N builds upon UK strengths as part of International development
Configuration Management
Storage Interfaces
Network Monitoring
Security
Information Services
Grid Data Management
SecurityMiddleware
Networking
Require some support expertise in each of these areas in order to maintain the Grid
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
II. Application Middleware
Fabric
TapeStorage
Elements
RequestFormulator and
Planner
Client Applications
ComputeElements
Indicates component that w ill be replaced
DiskStorage
Elements
LANs andWANs
Resource andServices Catalog
Rep licaCatalog
Meta-dataCatalog
Authentication and SecurityGSISAM-specific user, group , node, st at ion regis tration B bftp ‘cookie’
Connectivity and Resource
CORBA UDP File transfer protocol s - ftp, b bftp, rcp GridFTP
Mass Storage s ystems protocol se.g. encp, hp ss
Collective Services
C atalogproto co ls
S igni fi cant Event Log ger Naming Servi ce Database ManagerC at alog Manager
SAM R es ource M an ag em entB at ch Sys tems - LSF, FB S, PB S,
C ondorData Mov erJob Services
Storage ManagerJob ManagerCache ManagerRequest Manager
“Dataset Editor” “File Storage Server”“Project Master” “Station Master” “Station Master”
Web Python codes, Java codesCommand line D0 Framework C++ codes
“Stager”“Optimiser”
CodeRepostory
Name in “quotes” is SAM-given software component name
or addedenhanced using PPDG and Grid tools
GANGA
SAMGridLattice QCD
AliEn
CMS
BaBar
Require some support expertise in each of these areas in order to maintain the Grid applications. Need to develop e-Infrastructure portals for new experiments starting up in exploitation era.
Pheomenology
ATLAS UK e-science forward look (Roger Jones)
Both will move from development to optimisation & maintenance
Current core and infrastructure activities:
Run Time Testing and Validation Framework, tracking and trigger instantiations
Provision of ATLAS Distributed Analysis & production tools
Production management GANGA development Metadata development ATLFast simulation ATLANTIS Event Display Physics Software Tools
~11 FTEs mainly ATLAS e-science with some GridPP & HEFCE
Current Tracking
and Trigger e-science:
Alignment effort ~6FTEs
Core software ~2.5FTEs
Tracking tools ~6FTEs
Trigger ~2FTEs
The current eScience funding will only take us (at best) to first data
Expertise required for the real-world problems and maintenance
Note for the HLT, the installation and commissioning will continue into the running period because of staging
Need ~15 FTE (beyond existing rolling grant) in 2007/9 - continued e-science/GridPP support
CMS UK e-science forward look (Dave Newbold)
NB: ‘First look’ estimates; well inevitably change as we approach running
Need ~9 FTE (beyond existing rolling grant) in 2007/9 - continued e-science/GridPP support
Work area Current FTEs FTEs 2007-9 FTEs 2009 -
Cmp sys / support (e-science WP1)
2.0 3.0 - ramp UP for running phase
3.0 - steady state
Monitoring / DQM
(e-science WP3)
2.5 2.0 - initial running 1.5 - support / maintenance
Tracker software
(e-science WP4)
2.0 1.5 - initial deployment running
1.0 - support / maintenance
ECAL software
(e-science WP5)
2.0 1.5 - initial running 1.0 - support / maintenance
Data management(GridPP2)
1.5 1.5 - final dplymnt / support
1.5 - support / maintenance
Analysis system(GridPP2)
1.5 1.0 - final dplymnt / support
1.0 - support / maintenance
11.5 10.5 9.0
Computing system / supportDevelopment / tuning of computing model + system; managementUser support for T1 / T2 centres (globally); liaison with LCG ops
Monitoring / DQMOnline data gathering/‘expert systems’ for CMS tracker, trigger
Tracker /ECAL softwareInstallation / calibration support; low-level reconstruction codes
Data managementPhedex system for bulk offline data movement and trackingSystem-level metadata; movement of HLT farm data online (new area)
Analysis systemCMS-specific parts of distributed analysis system on LCG
LHCb UK e-science forward look (Nick Brook)Current core activities: GANGA development Provision of DIRAC & production
tools Development of conditions DB The production bookkeeping DB Data management & metadata Tracking Data Challenge Production
Manager
~10 FTEs mainly GridPP, e-science, studentships with some HEFCE support
Will move from development to maintenance phase - UK pro rata share of LHCb core computing activities ~5 FTEs
Current RICH & VELO e-science:
RICH: UK provide bulk of the RICH s/w team including s/w coordinator ~7 FTEs about 50:50 e-science funding+rolling grant/HEFCE
VELO: UK provide bulk of the VELO s/w team including s/w coordinator ~4 FTEs about 50:50 e-science funding+rolling grant/HEFCE
ALL essential alignment activities for both detectors through e-science funding
Will move from development to maintenance and operational alignment
~3FTEs for alignment in 2007-9Need ~9 FTE (core+alignment+UK support) in 2007/9 - continued e-science support
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Grid and e-Sciencefunding
requirements
Category Sub-Cat FTEs FTE Cost FY08Exploitation
LCG Staff £500,000 Contribution to Tier-0 supportTier-1 Staff 15.00 £75,000 £1,125,000 Running the Tier-1 Data CentreTier-1 Hardware £1,000,000 Hardware annual upgradeTier-2 Staff 19.00 £65,000 £1,235,000 Contribution to Tier-2 Sysman effort in proportion to (non-PPARC) hardware Tier-2 Hardware £100,000 Frontend Tier-2 hardwareApplications Staff 8.00 £65,000 £520,000 Application/Grid experts for UK supportM/S/N Staff 6.00 £75,000 £450,000 One expert in each of 6 areasOperations Managers £0 Assume production mode established + devolved management to InstitutesOperations Staff 5.00 £75,000 £375,000 Production manager and four Tier-2 coordinatorsOperations Misc £100,000 Operations responsive budgetTravel Support 53.00 £3,000 £159,000 Supported staff travel GRID TOTAL 53.00 £5,564,000 Grid Computing for PPNew Expts. Staff 8.00 £65,000 £520,000 Future experiments adopt e-Infrastructure methodsATLAS Staff 15.00 £65,000 £975,000 Computing MoU commitments and supportCMS Staff 8.00 £65,000 £520,000 Computing MoU commitments and supportLHCb Staff 8.00 £65,000 £520,000 Core Tasks and Computing SupportALICE Staff 1.00 £65,000 £65,000 Computing supportTravel Support 40.00 £3,000 £120,000 Supported staff travel
E-SCIENCE TOTAL 93.00 £8,284,000 Computing support for experiments (currently assumed to be in experiments line)Road Map £5,900,000
• Simple model
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
LCG 9%
Tier-118%
Tier-222%
Applications9%
M/S/N8%
Operations7%
Operations2%
Travel3%
Tier-1 20%
Tier-22%
• Priorities in context of a financial snapshot in 2008• Grid (£5.6m p.a.) and e-Science (£2.7m p.a.)
• Assumes no GridPP project management• Savings?
– EGEE Phase 2 (2006-08) may contribute– UK e-Science context is
1. NGS (National Grid Service)2. OMII (Open Middleware Infrastructure Institute)3. DCC (Digital Curation Centre)
• Timeline?
Grid and e-Sciencefunding
requirements
New Expts.19%
ATLAS37%
CMS19%
LHCb19%
ALICE2%
Travel4%
To be compared To be compared withwith
Road Map: Not a Road Map: Not a BidBid
- Preliminary Input- Preliminary Input
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Grid and e-ScienceExploitation Timeline?
• PPAP initial input Oct 2004• Science Committee initial input• PPARC call assessment (2007-2010) 2005• Science Committee outcome Oct 2005• PPARC call Jan 2006• PPARC close of call May 2006• Assessment Jun-Dec 2006• PPARC outcome Dec 2006 • Institute Recruitment/Retention Jan-Aug 2007• Grid and e-Science Exploitation Sep 2007 -
….
• Note if the assessment from PPARC internal planning differs significantly from this preliminary advice from PPAP and SC, then earlier planning is required.
Tony Doyle - University of Glasgow
12 January 2005 Collaboration Board
Summary
Resources needed for Grid and e-Science in medium-long term? • Current Road Map
~£6m p.a. • Resources needed in 2008 estimated at
£8.3m p.a.• Timeline for decision-making outlined..
• PP community-supported strategy required