Upload
rudolf-benson
View
218
Download
1
Tags:
Embed Size (px)
Citation preview
Participation of JINR in the LCG and EGEE projects
V.V.Korenkov (JINR, Dubna)
NEC’2005, Varna
17 September, 2005
Russian distributed Tier2 Cluster
PNPI
IHEP
RRC KI
ITEP
JINR
SINP MSU
RRC-LHC
LCG Tier1/Tier2cloud
CERN
…
Gbits/s
FZK
Regional connectivity:
cloud backbone – Gbit’s/s
to labs – 100–1000 Mbit/s
Collaborative centers
Tier2cluster
GRID access
The protocol between CERN, Russia and JINR on a participation in LCG Project has been approved in 2003.
The tasks of the Russian institutes in the LCG:
LCG software testing;
evaluation of new Grid technologies (e.g. Globus toolkit 3) in a context of using in the LCG;
event generators repository, data base of physical events: support and development;
LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)
LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)
The tasks of the Russian institutes & JINR in the LCG (2004 and 2005 years):
• LCG Deployment and Operation• LCG Test suit• Castor• LCG AA- Genser&MCDB• ARDA
LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)
JINR in LCG (2004 and 2005 years)
• LCG2 infrastructure was created at JINR• The server for monitoring Russian LCG sites was installed; • LCG web-portal was created in Russia and now its development is in
progress:• Tests on data transferring by the GridFTP protocol (GlobusToookit 3) were
made.• the toolkit GoToGrid on the automatic installation and tuning of the LCG-2
package was developed;• development of the MCDB system;• software for installation and control of MonaLisa clients on the base of
RMS (Remote Maintenance Shell) was designed. • Works to develop CASTOR2 system was in progress: development of the
control process of the garbage collection module, communication to Oracle DB;
• participation in the work to create the TESTBED of the new gLite middleware;
• Testing of next components gLite: Metadata catalog, Fireman catalog• Monitoring of WMS (Workload Management System) gLite testbed in INFN
site gundam.cnaf.infn.it
MYRINETclusterCOMMON
PC-farm
INTERACTIVE PC-farm
Структура комплекса
130 CPU
18TB RAID-5
ATL~ 5 (15) TB
6 – Interactive
18 – Common PC-farm
30 – LHC
14 – MYRINET (Parallel)
20 – LCG
20 – File servers
8 – LCG-user interface
MCDB Web Interface Screen-shot
http://mcdb.cern.ch
Only Mozilla Browser Supported (for the time being)
Russian Data Intensive GRID
(RDIG) Consortium
EGEE Federation
Eight Institutes made up the consortium RDIG (Russian Data Intensive GRID) as a national federation in the EGEE project. They are: IHEP - Institute of High Energy Physics (Protvino), IMPB RAS - Institute of Mathematical Problems in Biology (Pushchino), ITEP - Institute of Theoretical and Experimental Physics (Moscow), JINR - Joint Institute of Nuclear Physics , KIAM RAS - Keldysh Institute of Applied Mathematics, PNPI - Petersburg Nuclear Physics Institute (Gatchina), RRC KI - Russian Research Center “Kurchatov Institute”, SINP-MSU - Skobeltsyn Institute of Nuclear Physics (MSU).
Russian Contribution to EGEE
RDIG as an operational and functional part of EGEE infrastructure (CIC, ROC, RC).
Activities:
SA1 - European Grid Operations, Support and Management
SA2 – Network Resource Provision
NA2 – Dissemination and Outreach
NA3 – User Training and Induction
NA4 - Application Identification and Support
JINR role and work in EGEE
• SA1 - European Grid Operations, Support and Management EGEE-RDIG monitoring and accounting. Middleware deployment and
resource induction. Participation in the OMII and GT4 evaluation and in the gLite testing.
LCG SC activity coordination in Russia.
• NA2 - Dissemination and Outreach Coordination of this activity in Russia, organization of EGEE RDIG
Conference, Creation and run the RDIG Web site (http://www.egee-rdig.ru), dissemination in JINR Member states.
• NA3 - User Training and Induction Organization of grid tutorials, induction courses and training courses for
administrators.
• NA4 - Application Identification and Support Coordination of this activity in Russia, organization of HEP applications
in Russia through the EGEE infrastructure.
Grid middleware evaluations
• The goal of the evaluations is to get a better understanding of the functionality, performance, solidity, interoperability, deployability, management and usability of components in different grid MW distributions
• Aid decision about possible usage of components for the EGEE MW and about provision of interoperability between these distributions and the EGEE MW
• Evaluation of OMII distribution by JINR and KIAM in February - April 2005
• Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU in May - October 2005
Evaluation of OMII distribution by JINR and KIAM
http://www.gridclub.ru/library/OMII-evaluaton-EGEE3.ppt • Installation and configuration, supported platforms • Performance, scalability and reliability studies of OMII
services: JobService, DataService, dummy services • Aspects of security, authorization, account management,
resource allocation, administration in regard to the operation of a grid with many users, big virtual organizations and many resource centers
• Interoperability with gLite Workload Management System (WMS)
Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU
http://theory.sinp.msu.ru/dokuwiki/doku.php?id=egee:gt4:gt4• Installation and configuration, supported platforms • Performance, reliability, functional characteristics,
interfaces of JAVA WS-Core, WS-GRAM, GridFTP, RLS, RFT , WS-MDS4, WS Delegation service
• Aspects of security, authorization, usability and administration
• comparison of corresponding GT4 and gLite components
Participation in EGEE MW testing
Development of test suites for gLite (EGEE JRA1 activity) by JINR, IHEP and PNPI from June 2005 (continuing)• WMS DAG tests• WMS MPI tests • WMS JDL tests • R-GMA tests
SC3 GOALS• Service Challenge 1 (end of 2004):Demonstrate the possibility of throughput of 500 MByte/s to Tier1 in LCG environment.
• Service Challenge 2 (spring 2005):Maintain the throughput 500 MByte/s cumulative on all Tier1s for prolonged time, and
evaluate the data transfer environment on Tier0 и Tier1s. • Service Challenge 3 (Summer-end of 2005)Show reliable and stable data transfer on each Tier1: to disk -150 MByte/s, to tape - 60
MByte/s. All Tier1s and some Tier2s involved.
• Service Challenge 4 (Spring 2006):Prove the GRID infrastructure performance to handle the LHC data in proposed rate
(from raw data transfer up to final analysis) with all Tier1s and majority of Tier2s. • Final Goal:Build the production GRID-infrastructure on all Tier0, Tier1 и Tier2 according to the LHC
experiments specifics.
Summary of Tier0/1/2 Roles• Tier0 (CERN): safe keeping of RAW data (first copy); first pass
reconstruction, distribution of RAW data and reconstruction output to Tier1; reprocessing of data during LHC down-times;
• Tier1: safe keeping of a proportional share of RAW and reconstructed data; large scale reprocessing and safe keeping of corresponding output; distribution of data products to Tier2s and safe keeping of a share of simulated data produced at these Tier2s;
• Tier2: Handling analysis requirements and proportional share of simulated event production and reconstruction. No long term data storage.
Tier2 Roles• Tier2 roles vary by experiment, but include:
– Production of simulated data;– Production of calibration constants;– Active role in [end-user] analysis
Must also consider services offered to T2s by T1s
– e.g. safe-guarding of simulation output;– Delivery of analysis input.
• No fixed dependency between a given T2 and T1
A Simple T2 Model • Each T2 is configured to upload MC data to
and download data via a given T1
• In case the T1 is logical unavailable, wait and retry
• For data download, retrieve via alternate route / T1
Tier2 in Russia
Institute Link CPUs Disk OS/Middleware
IHEP 100 Mb/s half-duplex
5+ 1.6 TB …?
…?
ITEP 60 Mb/s 20 ? 2 TB ? SL-3.0.4 (kernel 2.4.21)
…?
JINR 45Mb/s
(1Gbit/s)10 ? 5 TB SLC3.0.X
LCG-2_4_0, Castor, gridftp, gLite?
SINP 1Gbit/s 30 ? 2 TB ? SL-3.0.4 (kernel 2.4.21)
gridftp, Castor
Universal Grid infrastructure in University Center of JINR
• Grid infrastructure is set of virtual machines (VMs) running on physical ones (hosts)• Virtualisation was made using User Mode Linux• current number of VMs is 36 (6 VMs on each of 6 hosts)• all virtual resourses are grouped into independent testbeds which in turn can be used for different aims: system administrators and users training in grid field, debugging and testing custom grid services in desirable grid environment
● Course for system administrators using Nordugrid ARC middleware was successfully conducted on that infrastructure
● LCG2 or gLite installation and configuration course is in future plans