Upload
lucinda-simon
View
213
Download
1
Embed Size (px)
Citation preview
Grid-based Information Architecture for iSERVO International Solid Earth Research
Virtual Organization
Western Pacific Geophysics Meeting (WPGM)
Beijing Convention Center
July 26 2006
Geoffrey FoxComputer Science, Informatics, Physics
Pervasive Technology Laboratories
Indiana University Bloomington IN 47401
http://grids.ucs.indiana.edu/ptliupages/presentations/[email protected] http://www.infomall.org
APEC Cooperation for Earthquake Simulation ACES is a seven year-long collaboration among scientists
interested in earthquake and tsunami predication• iSERVO is Infrastructure to support
work of ACES
• SERVOGrid is (completed) US Grid that is a prototype of iSERVO
• http://www.quakes.uq.edu.au/ACES/
Chartered under APEC – the Asia Pacific Economic Cooperation of 21 economies
Participating Institutions CSIRO Australia Monash University Australia University of Western Australia, Perth,
Australia University of Queensland Australia
University of Western Ontario Canada University of British Columbia Canada
China National Grid Chinese Academy of Sciences China Earthquake Administration China Earthquake Network Center
Brown University Boston University Jet Propulsion Laboratory Cal State Fullerton San Diego State University
UC Davis UC Irvine UC San Diego University of Southern California University of Minnesota Florida State University US Geological Survey Pacific Tsunami Warning Center PTWC
Hawaii
National Central University, Taiwan (Taiwan Chelungpu-fault Drilling Project)
University of Tokyo Tokyo Institute of Technology (Titech) Sophia University National Research Institute for Earth
Science and Disaster Prevention (NIED) Japan
Geographical Survey Institute, Japan
Role of Information Technology and Grids in ACES
Numerical simulations of physical, biological and social systems
Engineering designEconomic analysis and planningSensor networks and sensor websHigh performance computingData mining and pattern analysisDistance collaborationDistance learningPublic outreach and educationEmergency response communication and planningGeographic Information SystemsResource allocation and management
Grids and Cyberinfrastructure Grids are the technology based on Web services that implement
Cyberinfrastructure i.e. support eScience or science as a team sport
• Internet scale managed services that link computers data repositories sensors instruments and people
There is a portal and services in SERVOGrid for• Applications such as GeoFEST, RDAHMM, Pattern
Informatics, Virtual California (VC), Simplex, mesh generating programs …..
• Job management and monitoring web services for running the above codes.
• File management web services for moving files between various machines.
• Geographical Information System services • Quaketables earthquake specific database• Sensors as well as databases• Context (dynamic metadata) and UDDI system long term
metadata services• Services support streaming real-time data
Database Database
Analysis and VisualizationPortal
RepositoriesFederated Databases
Data Filter
Services
Field Trip DataStreaming Data
Sensors
?DiscoveryServices
SERVOGrid
ResearchSimulations
Research Education
CustomizationServices
From Research
to Education
EducationGrid ComputerFarmGrid of Grids: Research Grid and Education Grid
GISGrid
Sensor GridDatabase Grid
Compute Grid
SERVOGrid has a portalThe Portal is built from portlets
– providing user interface fragments for each service that are composed into the full interface – uses OGCE technology as does planetary science VLAB portal with University of Minnesota
Semantically Rich Services with a Semantically Rich Distributed Operating Environment
Database
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
SS
FS
FS
FS
FS
FS
FS
FS
FS
FS FS
FS
FS
FS
FS
FS
FS
FS
FS FS
FS
FS PortalFS
OS
OS
OS
OS
OS
OS
OS
OS
OS
OS
OS
OS
MD
MD
MD
MD
MD
MD
MD
MD
MD
MetaData
Filter Service
Sensor Service
OtherService
SOAP Message Streams
SOAP Message Streams
Raw Data Raw Data
Raw Data
Raw Data
Data
Data Data
Data
Information
Information
Knowledge
Knowledge
Wisdom
Decisions
Information
AnotherService
AnotherService
AnotherGrid
AnotherGrid
Grids of Grids Architecture
is same as outwardfacing application
service
Linking Grids and Services Linkage of Services and Grids requires that messages sent by one
Grid/Service can be understood by another Inside SERVOGrid all messages use
• Web service system standards we like (UDDI, WS-Context, WSDL, SOAP) and
• GML as extended by WFS so that data sources and simulations all use same syntax
All other Web service based Grids use their favorite Web service system standards but these differ from Grid to Grid• Further there is no agreement on application specific standards –
not all Earth Science Grids use OGC standards• OGC standards include some capabilities overlapping general Web
Services• Use of WSDL and SOAP is agreed although there are versioning
issues So there is essentially there is no service level interoperability between
Grids but rather interoperation is at diverse levels with shared technology• SQL for databases, PBS for Job scheduling, Condor for job
management, GT4 or Unicore for Grids
Grids in Babylon Presumptuous Tower of Babel (from the web)
• In the Bible, a city (now thought to be Babylon) in Shinar where God confounded a presumptuous attempt to build a tower into heaven by confusing the language of its builders into many mutually incomprehensible languages.
For Grids, everybody likes to do their own thing and Grids are complex multi-level entities where no obvious points of interoperation• so one does not need divine intervention to create multiple Grid
specifications• But data in China, Tsunami sensors in Indian ocean and simulations in
USA etc. will not be linked for better warning and forecasting unless the national efforts can interoperate
Two interoperation strategies:• Make all Grids use the same specifications (divine harmony)• Build translation services (filters!) using say OGF standards as a common
target language (more practical) Don’t need computers (jobs) to be interoperable (although this
would be good) as each country does its own computing• Rather need data and some metadata on each Grid to be accessible from
all Grids
Interoperability Summary Need to define common infrastructure and domain specific
standards• Build Interoperable Infrastructure gatewayed to existing legacy
applications and Grids Generic Middleware
• Grid software including workflow• Portals/Problem Solving environments incl. visualization• We need to ensure that we can make security, job submission,
portal, data access (sharing) mechanisms in different economies interoperate
Geographic Information Systems GIS• Use services as defined by Open Geospatial Consortium (Web
Map and Feature Services) http://www.crisisgrid.net/ Earthquake/Tsunami Science Specific
• Satellites, sensors (GPS, Seismic)• Fault, Tsunami … Characteristics stored in databases need
GML extensions - Schema for QuakeTables developed by SERVOGrid can be used Internationally
Country
and/or
Economies
Data (shared
as part of acollaboration)
Earthquake
Forecast/Model
Wave
Motion
Infrastructure
Institutions
Australia Seismic data, fault database, GPS
Finley, LSM
PANDAS
prototype Access
Canada Polaris Radarsat Pattern Informatics
P.R. China Seismic GPS LURR CAS China National Grid
Japan GPS
Seismic
Daichi (InSAR)
GeoFEM JST-CREST Earth Simulator
Naregi
ChineseTaipei
FORMOSAT-3/COSMIC (F/C)
U.S.A. QuakeTables
Sesismic
InSAR
PBO (GPS)
Pattern Informatics
ALLCAL
GeoFEST, PARK,
VirtualCalifornia
TeraShake SERVOGrid
GEON
SCECGrid
Vlab
International IMS Pacific Rim Universities
(APRU ) PRAGMA
ACES Components
National Earthquake Grids of Relevance APAC –GT2 GT4 gLite ACcESS – Some link to SERVOGrid China National Grid – GOS GT3 GT4 ChinaGrid – CGSP built on GT4 CNGI – China’s Next Generation Internet has significant
earthquake data component Naregi – Uses GT4 and Unicore with much enhancements Japanese Earthquake Simulation Grid – unclear K*Grid Korea Enhanced SRB, GT2 to GT4 TIGER Taiwan Integrated Grid for Education and Research
unclear technology and unclear earthquake relevance SERVOGrid – Uses WS-I+ simple Web Services TeraGrid – Uses GT4 but not a clear model except for core job
submittal
TeraGrid: Integrating NSF Cyberinfrastructure
TeraGrid is a facility that integrates computational, information, and analysis resources at the San Diego Supercomputer Center, the Texas Advanced Computing Center, the University of Chicago / Argonne National Laboratory, the National Center for Supercomputing Applications, Purdue University, Indiana University, Oak Ridge National Laboratory, the Pittsburgh Supercomputing Center, and the National Center for Atmospheric Research.Today 100 Teraflop; tomorrow a petaflop; Indiana 20 teraflop today.
SDSCTACC
UC/ANL
NCSA
ORNL
PU
IU
PSCNCAR
Caltech
USC-ISI
UtahIowa
Cornell
Buffalo
UNC-RENCI
Wisc
QPSF
ANU
VPAC
ac3
TPAC
CSIRO
Network:GrangeNet / AARNetAPAC Private Network (AARNet)
Security: APAC CAMyProxyVOMRS
APAC National GridCore Grid Services
Portal Tools:GridSphere
Info Services:APAC RegistryINCA2?
IVEC
SAPAC
APACNational Facility
Systems:GatewaysPartners’ systems
QPSF(JCU)
ACcESS at UQ (ACES Partner) outside APAC
National “Grid Projects” in China
Net-based Res. Env.
Plan
Research
Develop
Production Procure Deploy Operate Manage
CAS e-Science
Science and Technology R &D Assets Foundation Platform
Next-Generation Network Initiative
Edu. & Res. Grid
ChinaNational
Grid
StateCouncil
NSFC
CAS
MoE
MoST
NationalPlanningCommission
Semantic Grid
China e-Nation Strategy (2006-2020)
Virtual Comp. Env.
€M’s €10M’s Grid activities still growing
CNGrid (2006-2010)• HPC Systems
– 100 Tflop/s by 2008, Pflop/s by 2010?
• Grid Software Suite: CNGrid GOS– Merge with
international efforts– Emphasize production
• CNGrid Environment– Nodes, Centers, Policies
• Applications– Science– Resource & Environment– Manufacturing– Services
– Domain Grids
Cyber Science Infrastructure toward Petascale Computing (planned 2006-
2011)
Cyber-Science Infrastructure (CSI)
( IT Infra. for Academic Research and Education)
Operation/Maintenance
( Middleware )Networking Contents
NII Collaborative Operation Center
Delivyer
Delivery
Networking Infrastructure (Super-SINET )
Univ./National Supercomputing
VODomain Specific
VO(e.g ITBL)
Feedback
Feedback
R&D Collaboration
Operaontional Collaborati
MiddlewareCA
NAREGI Site
Research
Dev. ( )βver.V1.0V2.0
International Collaboration- EGEE- UNIGRIDS-Teragrid -GGF etc.
Feedback
Delivery
Project-Oriented
VO
Delivery
Domain SpecificVOs
CustomizationOperation/Maintenance
ナノ分野実証・評価
分子研
ナノ分野実証・評価
分子研
NanoProof of al.Concept
Eval.IMS
NanoProof, Eval.
IMSJoint Project
( Bio )Osaka-U
Joint Project
AIST
R&D Collaboration
IndustrialProjects
Project-orientedVO
Note: names of VO are tentative )
Peta-scaleSystem
VOCore Site
R&D Collaboration
Operation/Maintenance( UPKI,CA )
IMS,AIST,KEK,NAO,etc.
Japanese Earthquake Simulation Grid
Integrated Integrated Observation-SimulationObservation-Simulation
Data GridData Grid
PC ClusterERI,
64xOpteronparaAVS
Data-ServerGSI
8xOpteron20TB
Data-ServerNIED
48xG5, 15TB
PC ClusterEPS,
64xOpteronparaAVS
Super SINET (10Gbps)
Earth Simulator5,120xSX6
GONET Hi-net K-NET
Database for ModelConstruction
Plate Motion
Platform for Integrated SimulationData Processing, Visualization, Linear Solvers
Simulation Output
PC clusters for small-intermediate problemsEarth Simulator for large-scale problems
GIS UrbanInformation
Tectonic Loading EarthquakeRupture
Structure OscillationWave Propagation
Tsunami Generation
Earthquake Generation
Strong Motion and Tsunami Generation
JST-CREST Integrated Predictive Simulation System
Artificial Structure OscillationCrustal Movement
Data AnalysisSeismic Activity
Data AnalysisStrong MotionData Analysis
Current PTWC Network of Seismic Stations
(from GSN & USNSN & Other Contributing Networks)
The NCES/WS-*/GS-* Features/Service Areas IService or Feature WS-* GS-* NCES
(DoD)Comments
A: Broad Principles
FS1: Use SOA: Service Oriented Arch.
WS1 Core Service Architecture, Build Grids on Web Services. Industry best practice
FS2: Grid of Grids Strategy for legacy subsystems: modular architecture
B: Core Services (Mainly Service Infrastructure and W3C/OASIS focus)
FS3: Service Internet, Messaging
WS2 NCES3 Core Infrastructure including reliability, publish-subscribe messaging cf. FS13C
FS4: Notification WS3 NCES3 JMS, MQSeries, WS-Eventing, Notification
FS5: Workflow WS4 NCES5 Grid Programming
FS6: Security WS5 GS7 NCES2 Grid-Shib, Permis Liberty Alliance ...
FS7: Discovery WS6 NCES4 UDDI and extensions
FS8: System Metadata & State
WS7 Globus MDSSemantic Grid, WS-Context
FS9: Management WS8 GS6 NCES1 CIM
FS10: Policy WS9 ECS
FS11: Portals and Users WS10 NCES7 Portlets JSR168, NCES Capability Interfaces
The NCES/WS-*/GS-* Features/Service Areas IIService or Feature WS-* GS-* NCES Comments
B: Core Services (Mainly Higher level and OGF focus)
FS12: Computing GS3 Job Management major Grid focus
FS13A: Data as Repositories: Files and Databases
GS4 NCES8 Distributed Files, OGSA-DAIManaged Data is FS14B
FS13B: Data as Sensors and Instruments
OGC SensorML
FS13C: Data Transport WS 2,3 GS4 NCES3,8 GridFTP or WS Interface to non SOAP transport
FS14A: Information as Monitoring GS4 Major Grid effort for job status etc.
FS14B: Information, Knowledge, Wisdom part of D(ata)IKW
GS4 NCES8 VOSpace for IVOA, JBI for DoD, WFS for OGCFederation at this layer major research areaNCOW Data Strategy
FS15: Applications and User Services GS2 NCES9 Standalone ServicesProxies for jobs
FS16: Resources and Infrastructure GS5 Ad-hoc networks; Network Monitoring
FS17: Collaboration and Virtual Organizations
GS7 NCES6 XGSP, Shared Web Service ports
FS18: Scheduling and matching of Services and Resources
GS3 Current work only addresses scheduling “batch jobs”. Need networks and services