18
The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

Embed Size (px)

Citation preview

Page 1: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

TheParticle Physics Data Grid

Collaboratory Pilot

Richard P. MountFor the PPDG Collaboration

DOE SciDAC PI MeetingJanuary 15, 2002

Page 2: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

sin(2) = 0.59 ± 0.14 (statistical) ± 0.05 (systematic)

Observation of CP violation in the B0 meson system. (Announced July 5, 2001)

32 million B0 – anti-B0 pairs studied: these are the July 2001 plots after months of analysis

Page 3: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

The Top Quark Discovery (1995)

Page 4: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Quarks Revealed: structure inside Protons & Neutrons

1990 Nobel Prize in Physics

Richard Taylor (SLAC)

Page 5: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Scope and Goals

Who:• OASCR (Mary Anne Scott) and HENP (Vicky White)• Condor, Globus, SRM, SRB (PI, Miron Livny, U.Wisconsin)• High Energy and Nuclear Physics Experiments - ATLAS, BaBar, CMS, D0,

JLAB, STAR (PIs Richard Mount, SLAC and Harvey Newman, Caltech)• Project Coordinators: Ruth Pordes, Fermilab and Doug Olson, LBNL

Experiment data handling requirements today :• Petabytes of storage, Teraops/s of computing, Thousands of users, • Hundreds of institutions, 10+ years of analysis ahead

Focus of PPDG:• Vertical Integration of Grid middleware components into HENP

experiments’ ongoing work• Pragmatic development of common Grid services and standards –

data replication, storage and job management, monitoring and planning.

Page 6: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

End to end integration and deployment of experiment applications using existing and emerging Grid services.

Deployment of Grid technologies and services in production (24x7) environments with stressful performance needs.

Collaborative development of Grid middleware and extensions between application and middleware groups – leading to pragmatic and least risk solutions.

HENP experiments extend their adoption of common infrastructures to higher layers of their data analysis and processing applications.

Much attention paid to integration, coordination, interoperability and interworking with emphasis on incremental deployment of increasingly functional working systems.

The Novel Ideas

Page 7: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Impact and ConnectionsIMPACT. Make Grids usable and useful for the real problems facing international physics

collaborations and for the average scientist in HENP. Improving the robustness, reliability and maintainability of Grid software

through early use in production application environments. Common software components that have general applicability and contributions

to standard Grid middleware.

Connections DOE Science Grid will deploy and support Certificate Authorities and develop

Policy documents. Security and Policy for Group Collaboration provides Community Authorization

Service. SDM/SRM working with PPDG on common storage interface APIs and

software components. Connections with other SciDAC projects (HENP and non-HENP).

Page 8: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

PPDGB

aBar

Dat

a M

anag

emen

t

BaBar

D0

STAR

Jefferson Lab

CMSAtlas

Globus Users

SRB Users

Condor Users

StacsUsers

CM

S D

ata Managem

ent

Jlab Data Management

D0 Data M

anagement

STAR Data Management

Atla

s D

ata

Man

agem

ent

Globus Team

Condor

SRB Team

HE

NP

GC

Challenge and Opportunity

Page 9: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

The Growth of “Computational Physics” in HENP

Det

ecto

r an

d C

ompu

ting

Har

dwar

eP

hysics Analysis and R

esults

Large Scale Data Management

WorldwideCollaboration

(Grids)

FeatureExtraction

andSimulation

1971

2001~500 people (BaBar)

~10 people

~7 Million Lines of Code (BaBar)

~100k LOC

Page 10: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

The Collaboratory Past

• 30 years ago an HEP “collaboratory” involved: Air freight of bubble chamber film (e.g. CERN to

Cambridge)• 20 years ago:

Tens of thousands of tapes 100 physicists from all over Europe (or US) Air freight of tapes, 300 baud modems

• 10 years ago: Tens of thousands of tapes 500 physicists from US, Europe, USSR, PRC … 64k bps leased lines and air freight

Page 11: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

The Collaboratory Present and Future

• Present: Tens of thousands of tapes 500 physicists from US, Europe, Japan, FSU, PRC … Dedicated intercontinental links at up to 155/622 Mbps Home brewed, experiment-specific, data/job distribution

software (if you’re lucky)• Future (~2006):

Tens of thousands of tapes 2000 physicists from, worldwide collaboration Many links at 2.5/10 Gbps The Grid

Page 12: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

End-to-End Applications& Integrated Production Systems

to allow thousands of physicists to share data & computing resources for scientific processing and analyses

Operators & Users

Resources: Computers, Storage, Networks

PPDG Focus:

- Robust Data Replication

- Intelligent Job Placement and Scheduling

- Management of Storage Resources

- Monitoring and Information of Global Services

Relies on Grid infrastructure:- Security & Policy- High Speed Data Transfer- Network management

the challenges!

Put to gooduse by theExperiments

Page 13: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Project Activities to date – “One-to-one”Experiment – Computer Science developments

Replicated data sets for science analysisBaBar – SRB CMS – Globus, European Data GridSTAR – GlobusJLAB – SRB

http://www.jlab.org/hpc/WebServices/GGF3_WS-WG_Summary.pptDistributed Monte Carlo simulation job production and management

ATLAS – Globus, Condorhttp://atlassw1.phy.bnl.gov/magda/dyShowMain.pl

D0 – Condor CMS – Globus, Condor, EDG – SC2001 Demo

http://www-ed.fnal.gov/work/SC2001/mop-animate-2.htmlStorage management interfaces

STAR – SRMJLAB – SRB

Page 14: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Cross-Cut – all collaborator - activities

Certificate Authority policy and authentication – working with the SciDAC Science Grid, SciDAC Security and Policy for Group Collaboration and ESNET to develop policies and procedures. PPDG experiments will act as early testers and adopters of the CA.

http://www.envisage.es.net/

Monitoring of networks, computers, storage and applications – collaboration with GriPhyN. Developing use cases and requirements; evaluating and analysing existing systems with many components – D0 SAM, Condor pools etc. SC2001 demo:

http://www-iepm.slac.stanford.edu/pinger/perfmap/iperf/anim.gif.

Architecture components and interfaces – collaboration with GriPhyN. Defining services and interfaces for analysis, comparison, and discussion with other architecture definitions such as the European Data Grid.

http://www.griphyn.org/mail_archive/all/doc00012.doc

International test beds – iVDGL and experiment applications.

Page 15: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Common Middleware Services

Robust file transfer and replica services SRB Replication Services Globus replication services Globus robust file transfer GDMP application replication layer - common project between

European Data Grid Work Package 2 and PPDG.

Distributed Job Scheduling and Resource Management: Condor-G, DAGman, Gram; Sc2001 demo with GriPhyN

http://www-ed.fnal.gov/work/sc2001/griphyn-animate.html

Storage Resource Interface and Management Common API with EDG, SRM

Standards Committees Internet2 HENP Working Group Global Grid Forum

Page 16: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Grid Realities

BaBar Offline Computing EquipmentBottom-up Cost Estimate (December 2001)

(Based only on costs we already expect)(To be revised annually)

0

1

2

3

4

5

6

7

8

9

10

1997 1998 1999 2000 2001 2002 2003 2004 2005

Eq

uip

men

t F

un

din

g $

M

TierA

Shared funding

SLAC baseline contribution

Page 17: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

Grid Realities

Page 18: The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002

www.ppdg.net

PPDG World

An Experiment

PPDG

BaB

ar D

ata

Man

agem

ent

BaBar

D0

STAR

Jefferson Lab

CMSAtlas

Globus Users

SRB Users

Condor Users

StacsUsers

CM

S D

ata Managem

ent

Jlab Data Management

D0 Data M

anagement

STAR Data Management

Atla

s D

ata

Man

agem

ent

Globus Team

Condor

SRB Team

HE

NP

GC

PPDG

BaBarData Management

BaB

ar P

PD

G T

eam

PPDG

IN2P3 and

French Grid

IEPM

SLAC

GridPP

Netw

ork Monitoring

BaBar France

BaBar SLAC D

ata Management

BaBar UK

HICB

GriP

hyN

Man

agem

ent PPDGEuropean Data

Grid

iVDGLLHC

Computing Grid ProjectDataTAG

PParc (UK Grid)

French Grid

CrossGrid

SC

2/PE

B

iVDGL mgmt

PPDG managament

Data access Management

Dat

aTag

mgm

t

UK HENP m

gmt

EDG mgmt

IN2P3 mgmt

CrossG

rid Mgm

tExperiments

GriPhyNHENP Grid

PPDG

DOE SciDAC HENP

PPDG

DOE SciDAC MICS

SRM

IEPM

DOE Science

Grid

Middleware for science portals

Computing Infrastructure

for Lattice Gauge

Data Grid

Toolkit

Security and Grid

collaboration

SciDAC

SciDAC connections