Upload
maude-rodgers
View
222
Download
0
Tags:
Embed Size (px)
Citation preview
TheParticle Physics Data Grid
Collaboratory Pilot
Richard P. MountFor the PPDG Collaboration
DOE SciDAC PI MeetingJanuary 15, 2002
www.ppdg.net
sin(2) = 0.59 ± 0.14 (statistical) ± 0.05 (systematic)
Observation of CP violation in the B0 meson system. (Announced July 5, 2001)
32 million B0 – anti-B0 pairs studied: these are the July 2001 plots after months of analysis
www.ppdg.net
The Top Quark Discovery (1995)
www.ppdg.net
Quarks Revealed: structure inside Protons & Neutrons
1990 Nobel Prize in Physics
Richard Taylor (SLAC)
www.ppdg.net
Scope and Goals
Who:• OASCR (Mary Anne Scott) and HENP (Vicky White)• Condor, Globus, SRM, SRB (PI, Miron Livny, U.Wisconsin)• High Energy and Nuclear Physics Experiments - ATLAS, BaBar, CMS, D0,
JLAB, STAR (PIs Richard Mount, SLAC and Harvey Newman, Caltech)• Project Coordinators: Ruth Pordes, Fermilab and Doug Olson, LBNL
Experiment data handling requirements today :• Petabytes of storage, Teraops/s of computing, Thousands of users, • Hundreds of institutions, 10+ years of analysis ahead
Focus of PPDG:• Vertical Integration of Grid middleware components into HENP
experiments’ ongoing work• Pragmatic development of common Grid services and standards –
data replication, storage and job management, monitoring and planning.
www.ppdg.net
End to end integration and deployment of experiment applications using existing and emerging Grid services.
Deployment of Grid technologies and services in production (24x7) environments with stressful performance needs.
Collaborative development of Grid middleware and extensions between application and middleware groups – leading to pragmatic and least risk solutions.
HENP experiments extend their adoption of common infrastructures to higher layers of their data analysis and processing applications.
Much attention paid to integration, coordination, interoperability and interworking with emphasis on incremental deployment of increasingly functional working systems.
The Novel Ideas
www.ppdg.net
Impact and ConnectionsIMPACT. Make Grids usable and useful for the real problems facing international physics
collaborations and for the average scientist in HENP. Improving the robustness, reliability and maintainability of Grid software
through early use in production application environments. Common software components that have general applicability and contributions
to standard Grid middleware.
Connections DOE Science Grid will deploy and support Certificate Authorities and develop
Policy documents. Security and Policy for Group Collaboration provides Community Authorization
Service. SDM/SRM working with PPDG on common storage interface APIs and
software components. Connections with other SciDAC projects (HENP and non-HENP).
www.ppdg.net
PPDGB
aBar
Dat
a M
anag
emen
t
BaBar
D0
STAR
Jefferson Lab
CMSAtlas
Globus Users
SRB Users
Condor Users
StacsUsers
CM
S D
ata Managem
ent
Jlab Data Management
D0 Data M
anagement
STAR Data Management
Atla
s D
ata
Man
agem
ent
Globus Team
Condor
SRB Team
HE
NP
GC
Challenge and Opportunity
www.ppdg.net
The Growth of “Computational Physics” in HENP
Det
ecto
r an
d C
ompu
ting
Har
dwar
eP
hysics Analysis and R
esults
Large Scale Data Management
WorldwideCollaboration
(Grids)
FeatureExtraction
andSimulation
1971
2001~500 people (BaBar)
~10 people
~7 Million Lines of Code (BaBar)
~100k LOC
www.ppdg.net
The Collaboratory Past
• 30 years ago an HEP “collaboratory” involved: Air freight of bubble chamber film (e.g. CERN to
Cambridge)• 20 years ago:
Tens of thousands of tapes 100 physicists from all over Europe (or US) Air freight of tapes, 300 baud modems
• 10 years ago: Tens of thousands of tapes 500 physicists from US, Europe, USSR, PRC … 64k bps leased lines and air freight
www.ppdg.net
The Collaboratory Present and Future
• Present: Tens of thousands of tapes 500 physicists from US, Europe, Japan, FSU, PRC … Dedicated intercontinental links at up to 155/622 Mbps Home brewed, experiment-specific, data/job distribution
software (if you’re lucky)• Future (~2006):
Tens of thousands of tapes 2000 physicists from, worldwide collaboration Many links at 2.5/10 Gbps The Grid
www.ppdg.net
End-to-End Applications& Integrated Production Systems
to allow thousands of physicists to share data & computing resources for scientific processing and analyses
Operators & Users
Resources: Computers, Storage, Networks
PPDG Focus:
- Robust Data Replication
- Intelligent Job Placement and Scheduling
- Management of Storage Resources
- Monitoring and Information of Global Services
Relies on Grid infrastructure:- Security & Policy- High Speed Data Transfer- Network management
the challenges!
Put to gooduse by theExperiments
www.ppdg.net
Project Activities to date – “One-to-one”Experiment – Computer Science developments
Replicated data sets for science analysisBaBar – SRB CMS – Globus, European Data GridSTAR – GlobusJLAB – SRB
http://www.jlab.org/hpc/WebServices/GGF3_WS-WG_Summary.pptDistributed Monte Carlo simulation job production and management
ATLAS – Globus, Condorhttp://atlassw1.phy.bnl.gov/magda/dyShowMain.pl
D0 – Condor CMS – Globus, Condor, EDG – SC2001 Demo
http://www-ed.fnal.gov/work/SC2001/mop-animate-2.htmlStorage management interfaces
STAR – SRMJLAB – SRB
www.ppdg.net
Cross-Cut – all collaborator - activities
Certificate Authority policy and authentication – working with the SciDAC Science Grid, SciDAC Security and Policy for Group Collaboration and ESNET to develop policies and procedures. PPDG experiments will act as early testers and adopters of the CA.
http://www.envisage.es.net/
Monitoring of networks, computers, storage and applications – collaboration with GriPhyN. Developing use cases and requirements; evaluating and analysing existing systems with many components – D0 SAM, Condor pools etc. SC2001 demo:
http://www-iepm.slac.stanford.edu/pinger/perfmap/iperf/anim.gif.
Architecture components and interfaces – collaboration with GriPhyN. Defining services and interfaces for analysis, comparison, and discussion with other architecture definitions such as the European Data Grid.
http://www.griphyn.org/mail_archive/all/doc00012.doc
International test beds – iVDGL and experiment applications.
www.ppdg.net
Common Middleware Services
Robust file transfer and replica services SRB Replication Services Globus replication services Globus robust file transfer GDMP application replication layer - common project between
European Data Grid Work Package 2 and PPDG.
Distributed Job Scheduling and Resource Management: Condor-G, DAGman, Gram; Sc2001 demo with GriPhyN
http://www-ed.fnal.gov/work/sc2001/griphyn-animate.html
Storage Resource Interface and Management Common API with EDG, SRM
Standards Committees Internet2 HENP Working Group Global Grid Forum
www.ppdg.net
Grid Realities
BaBar Offline Computing EquipmentBottom-up Cost Estimate (December 2001)
(Based only on costs we already expect)(To be revised annually)
0
1
2
3
4
5
6
7
8
9
10
1997 1998 1999 2000 2001 2002 2003 2004 2005
Eq
uip
men
t F
un
din
g $
M
TierA
Shared funding
SLAC baseline contribution
www.ppdg.net
Grid Realities
www.ppdg.net
PPDG World
An Experiment
PPDG
BaB
ar D
ata
Man
agem
ent
BaBar
D0
STAR
Jefferson Lab
CMSAtlas
Globus Users
SRB Users
Condor Users
StacsUsers
CM
S D
ata Managem
ent
Jlab Data Management
D0 Data M
anagement
STAR Data Management
Atla
s D
ata
Man
agem
ent
Globus Team
Condor
SRB Team
HE
NP
GC
PPDG
BaBarData Management
BaB
ar P
PD
G T
eam
PPDG
IN2P3 and
French Grid
IEPM
SLAC
GridPP
Netw
ork Monitoring
BaBar France
BaBar SLAC D
ata Management
BaBar UK
HICB
GriP
hyN
Man
agem
ent PPDGEuropean Data
Grid
iVDGLLHC
Computing Grid ProjectDataTAG
PParc (UK Grid)
French Grid
CrossGrid
SC
2/PE
B
iVDGL mgmt
PPDG managament
Data access Management
Dat
aTag
mgm
t
UK HENP m
gmt
EDG mgmt
IN2P3 mgmt
CrossG
rid Mgm
tExperiments
GriPhyNHENP Grid
PPDG
DOE SciDAC HENP
PPDG
DOE SciDAC MICS
SRM
IEPM
DOE Science
Grid
Middleware for science portals
Computing Infrastructure
for Lattice Gauge
Data Grid
Toolkit
Security and Grid
collaboration
SciDAC
SciDAC connections