Transcript

Participation of JINR in the LCG and EGEE projects

V.V.Korenkov (JINR, Dubna)

NEC’2005, Varna

17 September, 2005

Federating Worldwide Resources for the LHC

Infrastructure LCG/EGEE

Russian distributed Tier2 Cluster

PNPI

IHEP

RRC KI

ITEP

JINR

SINP MSU

RRC-LHC

LCG Tier1/Tier2cloud

CERN

Gbits/s

FZK

Regional connectivity:

cloud backbone – Gbit’s/s

to labs – 100–1000 Mbit/s

Collaborative centers

Tier2cluster

GRID access

The protocol between CERN, Russia and JINR on a participation in LCG Project has been approved in 2003.

The tasks of the Russian institutes in the LCG:

LCG software testing;

evaluation of new Grid technologies (e.g. Globus toolkit 3) in a context of using in the LCG;

event generators repository, data base of physical events: support and development;

LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)

LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)

The tasks of the Russian institutes & JINR in the LCG (2004 and 2005 years):

• LCG Deployment and Operation• LCG Test suit• Castor• LCG AA- Genser&MCDB• ARDA

LHC Computing Grid Project (LCG)LHC Computing Grid Project (LCG)

JINR in LCG (2004 and 2005 years)

• LCG2 infrastructure was created at JINR• The server for monitoring Russian LCG sites was installed; • LCG web-portal was created in Russia and now its development is in

progress:• Tests on data transferring by the GridFTP protocol (GlobusToookit 3) were

made.• the toolkit GoToGrid on the automatic installation and tuning of the LCG-2

package was developed;• development of the MCDB system;• software for installation and control of MonaLisa clients on the base of

RMS (Remote Maintenance Shell) was designed. • Works to develop CASTOR2 system was in progress: development of the

control process of the garbage collection module, communication to Oracle DB;

• participation in the work to create the TESTBED of the new gLite middleware;

• Testing of next components gLite: Metadata catalog, Fireman catalog• Monitoring of WMS (Workload Management System) gLite testbed in INFN

site gundam.cnaf.infn.it

MYRINETclusterCOMMON

PC-farm

INTERACTIVE PC-farm

Структура комплекса

130 CPU

18TB RAID-5

ATL~ 5 (15) TB

6 – Interactive

18 – Common PC-farm

30 – LHC

14 – MYRINET (Parallel)

20 – LCG

20 – File servers

8 – LCG-user interface

MCDB Web Interface Screen-shot

http://mcdb.cern.ch

Only Mozilla Browser Supported (for the time being)

Russian Data Intensive GRID

(RDIG) Consortium

EGEE Federation

Eight Institutes made up the consortium RDIG (Russian Data Intensive GRID) as a national federation in the EGEE project. They are: IHEP - Institute of High Energy Physics (Protvino), IMPB RAS - Institute of Mathematical Problems in Biology (Pushchino), ITEP - Institute of Theoretical and Experimental Physics (Moscow), JINR - Joint Institute of Nuclear Physics , KIAM RAS - Keldysh Institute of Applied Mathematics, PNPI - Petersburg Nuclear Physics Institute (Gatchina), RRC KI - Russian Research Center “Kurchatov Institute”, SINP-MSU - Skobeltsyn Institute of Nuclear Physics (MSU).

Russian Contribution to EGEE

RDIG as an operational and functional part of EGEE infrastructure (CIC, ROC, RC).

Activities:

SA1 - European Grid Operations, Support and Management

SA2 – Network Resource Provision

NA2 – Dissemination and Outreach

NA3 – User Training and Induction

NA4 - Application Identification and Support

JINR role and work in EGEE

• SA1 - European Grid Operations, Support and Management EGEE-RDIG monitoring and accounting. Middleware deployment and

resource induction. Participation in the OMII and GT4 evaluation and in the gLite testing.

LCG SC activity coordination in Russia.

• NA2 - Dissemination and Outreach Coordination of this activity in Russia, organization of EGEE RDIG

Conference, Creation and run the RDIG Web site (http://www.egee-rdig.ru), dissemination in JINR Member states.

• NA3 - User Training and Induction Organization of grid tutorials, induction courses and training courses for

administrators.

• NA4 - Application Identification and Support Coordination of this activity in Russia, organization of HEP applications

in Russia through the EGEE infrastructure.

Grid middleware evaluations

• The goal of the evaluations is to get a better understanding of the functionality, performance, solidity, interoperability, deployability, management and usability of components in different grid MW distributions

• Aid decision about possible usage of components for the EGEE MW and about provision of interoperability between these distributions and the EGEE MW

• Evaluation of OMII distribution by JINR and KIAM in February - April 2005

• Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU in May - October 2005

Evaluation of OMII distribution by JINR and KIAM

http://www.gridclub.ru/library/OMII-evaluaton-EGEE3.ppt • Installation and configuration, supported platforms • Performance, scalability and reliability studies of OMII

services: JobService, DataService, dummy services • Aspects of security, authorization, account management,

resource allocation, administration in regard to the operation of a grid with many users, big virtual organizations and many resource centers

• Interoperability with gLite Workload Management System (WMS)

Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU

http://theory.sinp.msu.ru/dokuwiki/doku.php?id=egee:gt4:gt4• Installation and configuration, supported platforms • Performance, reliability, functional characteristics,

interfaces of JAVA WS-Core, WS-GRAM, GridFTP, RLS, RFT , WS-MDS4, WS Delegation service

• Aspects of security, authorization, usability and administration

• comparison of corresponding GT4 and gLite components

Participation in EGEE MW testing

Development of test suites for gLite (EGEE JRA1 activity) by JINR, IHEP and PNPI from June 2005 (continuing)• WMS DAG tests• WMS MPI tests • WMS JDL tests • R-GMA tests

RDIG мониторинг в ОИЯИ http://rocmon.jinr.ru:8080

SC3 GOALS• Service Challenge 1 (end of 2004):Demonstrate the possibility of throughput of 500 MByte/s to Tier1 in LCG environment.

• Service Challenge 2 (spring 2005):Maintain the throughput 500 MByte/s cumulative on all Tier1s for prolonged time, and

evaluate the data transfer environment on Tier0 и Tier1s. • Service Challenge 3 (Summer-end of 2005)Show reliable and stable data transfer on each Tier1: to disk -150 MByte/s, to tape - 60

MByte/s. All Tier1s and some Tier2s involved.

• Service Challenge 4 (Spring 2006):Prove the GRID infrastructure performance to handle the LHC data in proposed rate

(from raw data transfer up to final analysis) with all Tier1s and majority of Tier2s. • Final Goal:Build the production GRID-infrastructure on all Tier0, Tier1 и Tier2 according to the LHC

experiments specifics.

Summary of Tier0/1/2 Roles• Tier0 (CERN): safe keeping of RAW data (first copy); first pass

reconstruction, distribution of RAW data and reconstruction output to Tier1; reprocessing of data during LHC down-times;

• Tier1: safe keeping of a proportional share of RAW and reconstructed data; large scale reprocessing and safe keeping of corresponding output; distribution of data products to Tier2s and safe keeping of a share of simulated data produced at these Tier2s;

• Tier2: Handling analysis requirements and proportional share of simulated event production and reconstruction. No long term data storage.

Tier2 Roles• Tier2 roles vary by experiment, but include:

– Production of simulated data;– Production of calibration constants;– Active role in [end-user] analysis

Must also consider services offered to T2s by T1s

– e.g. safe-guarding of simulation output;– Delivery of analysis input.

• No fixed dependency between a given T2 and T1

A Simple T2 Model • Each T2 is configured to upload MC data to

and download data via a given T1

• In case the T1 is logical unavailable, wait and retry

• For data download, retrieve via alternate route / T1

Tier1/2 Network Topology

Tier2 in Russia

Institute Link CPUs Disk OS/Middleware

IHEP 100 Mb/s half-duplex

5+ 1.6 TB …?

…?

ITEP 60 Mb/s 20 ? 2 TB ? SL-3.0.4 (kernel 2.4.21)

…?

JINR 45Mb/s

(1Gbit/s)10 ? 5 TB SLC3.0.X

LCG-2_4_0, Castor, gridftp, gLite?

SINP 1Gbit/s 30 ? 2 TB ? SL-3.0.4 (kernel 2.4.21)

gridftp, Castor

Universal Grid infrastructure in University Center of JINR

• Grid infrastructure is set of virtual machines (VMs) running on physical ones (hosts)• Virtualisation was made using User Mode Linux• current number of VMs is 36 (6 VMs on each of 6 hosts)• all virtual resourses are grouped into independent testbeds which in turn can be used for different aims: system administrators and users training in grid field, debugging and testing custom grid services in desirable grid environment

● Course for system administrators using Nordugrid ARC middleware was successfully conducted on that infrastructure

● LCG2 or gLite installation and configuration course is in future plans

EGEE NA3 Coursesin Dubna

28.06.2004NA3. Introduction Courses

29-3-.03.2005LCG2 Administrator`s Course,

06.09.2005LCG-2 Induction Courses for CMS Users


Recommended