16
March03 CHEP'03 1 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

Embed Size (px)

Citation preview

Page 1: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 1

Databases for data management in PHENIX

Irina Sourikova

Brookhaven National Laboratory

for the

PHENIX collaboration

Page 2: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 2

PHENIX experiment

• PHENIX is a large RHIC experiment producing about 100TB of data a year

• 400+ collaborators in 57 institutions• Widely distributed – 12 countries, 4 continents

Page 3: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 3

BNLCC-F

Lund CC-JVUUNM SUNYSB

Major PHENIX Computing Sites

LLNL

Page 4: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 4

PHENIX data

• PHENIX is in the third year of data taking and has accumulated about 200,000 data files so far. At the end of current run (May 2003) this number will increase up to 400,000 files

• Files are produced at multiple computing centers • Large amount of data is moved from/to BNL and

between PHENIX institutions• More on PHENIX data management in Andrey

Shevel’s talk in Category 2, 5:30p.m.

Page 5: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 5

analysis, simulation,

and reconstruction

of simulated data

CC-J

UNMWI

CC-F

BNL

simulation farm off-site analysis

analysis andreconstruction

50 (300) TB/yr 10 TB/yr

5 TB/yr10 TB/yr

Page 6: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 6

File Catalog requirements

• Only 10% of the data can be kept on disk for analysis• Set of disk resident files is constantly changing • File Catalog should

– Contain up-to-date data for both master and replica file location

– Provide fast and reliable access at all PHENIX sites– Provide write permissions to the sites that store portions of

the data set

• Last requirement was not satisfied by existing catalogs(Objectivity based PHENIX catalog, MAGDA, Globus Replica Catalog and others)

Page 7: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 7

Objy: difficulty of PDN negotiation, exposes primary DB to world

WAN latency, single-point-of-failure

Page 8: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 8

What we need: hybrid solution, extra software involved

Page 9: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 9

Database technology choice

• Objectivity – problems with peer-to-peer replication

• Oracle was an obvious candidate(but expensive)

• MySQL didn’t have ACID properties and referential integrity a year ago when we were considering our options. Had only master-slave replication

• postgreSQL seemed a very attractive DBMS with several existing projects on peer-to-peer replication

• SOLUTION: to have central Objy based metadata catalog and distributed file replica catalog

Page 10: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 10

PostgreSQL

• ACID compliant• Multi-Version Concurrency Control for

concurrent applications• Easy to use• Free• LISTEN and NOTIFY support message passing

and client notification of an event in the database. Important for automating data replication

Page 11: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 11

PostgreSQL Replicator

• http://pgreplicator.sourceforge.net• Partial, peer-to-peer, async replication• Table level data ownership model• Table level replicated database set• Master/Slave, Update Anywhere, Workload

Partitioning data ownership models are supported

• Table level conflict resolution

Page 12: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 12

Distributed PHENIX Catalog

• Distribution includes BNL and Stony Brook University

• Two sites have common tables that are synchronized during replica session

• Common tables have Workload Partitioning data ownership model, that means the table entries can be updated only by the site-originator

Page 13: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 13

SB BNL SB BNL

ARGO ARGO

SYNC

Stony Brook BNL

DBadmin DBadminClients Clients

Production

Staging

Page 14: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 14

File Catalog replication

• Synchronization is two-way

• DB replication is partial, only latest updates are transferred over the network

• That solves scalability problems

• Replication of 20 000 new updates takes 1min

Page 15: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 15

Page 16: March03CHEP'031 Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration

March03 CHEP'03 16

What we gained by distributing File Catalog

• Reliability(no single point of failure)• Accessibility(reduced workload, no WAN latency)• Solved security problems (firewall conduits only

between servers)• Catalog data is more up-to-date(all sites update the

catalog online)• Less manpower needed to maintain the catalog