Upload
sophie-alexander
View
218
Download
0
Tags:
Embed Size (px)
Citation preview
PL-Grid: Polish Infrastructure for Supporting Computational Science in the European Research Space
1
Distributed Computing Instrastructure as a Tool for e-Science
ACK Cyfronet AGHPL-Grid Consortium
Jacek Kitowski, Kazimierz Wiatr, Łukasz Dutka, Maciej Twardy, Tomasz Szepieniec, Mariusz Sterzel, Renata Słota and Robert Pająk
PPAM 2015, 7-9.09.2015, Kraków
2
Outline
National e-Infrastructure
Assumptions and foundations
Tool for e-Science
e-Infrastructure creation – motivation, background and issues
Conceptualization and implementation
PLGrid case study
Enhancement of Achievements
Platforms and Environments – Selected Examples
Conclusions
3e-Infrastructure CreationMotivation and Background
Experiments in silico:Computing and Data for Open Science
International collaboration
User/platform driven e-infrastructure innovation (e-Science and e-Infrastructure interaction)
Increasing importance of Computational Science and Big Data Analysis
Data intensive computing
Numerically intensive computing
Computational Science problems:Algorithms, environments and deployment
Future and emerging technologies
4th paradigm, distributed, grid and cloud computing,Data Farming
Needs:Preventing users from technical problems
Expert support for making science
Increase of resources
Openess for future paradigms
Activity initiated by Cyfronet
4
Organizational:
horizontal perspective - federation of computer centres supporting the e-infrastructure with different kinds of resources and competences
vertical perspective - involvement of computer, computational and domain-specific experts into e-infrastructure operations
Meeting user demands in the field of grand challenges applications
Activity supported by users with scientific achievements and by well-defined requirements
Technological:
different computing hardware and software
various middleware environments
Energy awareness:
optimal scheduling strategies of computing jobs among federation resources to minimize energy consumption as a whole
Synergistic effort in several dimensions:
e-Infrastructure CreationIssues
5
PL-Grid Consortium
Consortium creation – 2007a response to requirements from Polish scientists
due to ongoing eScience activities in Europe and in the World
Aim: significant extension of computing resources and solutions provided to the scientific community
PL-Grid Programme
Development based on (SWOT analysis): projects funded by the European Regional Development Fund as part of the Innovative Economy Program
close international collaboration (EGI, ….)
previous projects (5FP, 6FP, 7FP, EDA…)
National Network Infrastructure available: Pionier National Project
computing resources: Top500 list
Polish scientific communities: ~75% highly rated Polish publications in 5 Communities
PL-Grid Consortium members: 5 High Performance Computing Polish Centres,
representing the Communities
coordinated by ACC Cyfronet AGH
ACK Cyfronet AGH 42 years of expertise
High Performace Computing
High Performance Networking
Centre of Competence
Human Resources
Infrastructure Resources
Network Resources
RankTOP500
Site System CoresRmax
TflopsRpeak
Tflops
49VII.2015
CyfronetPoland
PrometheusHP Apollo 8000
41,472 1262.4 1658.9
269VII.2015
CyfronetPoland
ZeusCluster Platform Hewlett-Packard
25,468 266.9 373.9
Social Networking
7
The most powerful HPC Asset (in Poland)Prometheus Cluster (2014/2015)
Rpeak = 1658.9 TFlops
1728 servers
41,472 Haswell cores
216 TB RAM (DDR4)
10 PB disks, 180 GB/s
HP Apollo 8000
In operation April 2015
49th position on the July 2015 edition of the TOP500 list
Q4 2015 Extensions
Rpeak = 483.8 TFlops
504 servers
12,096 Haswell cores
RpeakNVIDIA= 256.3 TFlops
144 Nvidia K40 XL
In SUMMARY: 2,4 PFlops (with GPU)
88TOP500, July 2015 Polish Sites
Rank Site System CoresRmax
(TFlop/s)Rpeak
(TFlop/s)
49
269
Cyfronet,Krakow
Prometheus - HP Apollo 8000, Xeon E5-2680v3 12C 2.5GHz, Infiniband FDRHewlett-Packard
Zeus - Cluster Platform SL390/BL2x220, Xeon X5650 6C, 2.660GHz, Infiniband QDR, NVIDIA 2090Hewlett-Packard
41,472
25,468
1,262.4
266.9
1,658.9
373.9
126TASK, Gdańsk
Tryton - HP ProLiant XL230a Gen9, Xeon E5-2670v3 12C 2.3GHz, InfinibandMegatel/Action
17,280 530.5 635.9
135WCSS,Wrocław
BEM - Actina Solar 820 S6, Xeon E5-2670v3 12C 2.3GHz, Infiniband FDRACTION
17,280 480.1 635.9
155NCNR, Świerk
Świerk Computing Centre - Supermicro TwinBlade SBI-7227R/Bull DLC B720, Intel Xeon E5-2680v2/E5-2650 v3 10C 2.8GHz, Infiniband QDR/FDRFormat, Bull, Atos Group
17,960 423.2 490.4
380
418
NGSC &ICM,
University of Warsaw
ORION - Dell PowerEdge R730, Xeon E5-2680v3 12C 2.5GHz, Infiniband FDR, AMD FirePro S9150Dell
BlueGene/Q, Power BQC 16C 1.600GHz, Custom InterconnectIBM
16,800
16,384
198.8
189.0
903.0
209.7
10
Family of PL-Grid Projects coordinated by Cyfronet
PLGrid PLUS (2011–2015)
Outcome:
Focus on users (training, helpdesk…)
Domain specific solutions: 13
PL-Grid (2009–2012)
Outcome: Common base infrastructure
PLGrid NG (2014–2015)
Outcome:
Optimization of resources usage, training
Extension of domain specific by 14
PLGrid CORE (2014–2015)
Outcome: Competence Center
Open Science paradigm(large workflow app., data farming mass comp., ……)
End-user services
230 Tflops
+1500 Tflops
+500 Tflops
+8 Tflops
Assumed Performance
Real Users
12
Summary of Projects Results(up-to-date)
Close collaboration between Partners and research communities
Development of tools, environments and middleware services, Clouds
Integration, HPC, Data intensive, Instruments
Development of 27 domain specific solutions
Development of IT PL-Grid Infrastructure and ecosystem
13
Summary of Projects Results(up-to-date)
Facilitation of community participation in international collaboration
EGI Council, EGI Executive Board
FP7 (VPH-Share, VirtROLL….)
EGI-InSPIRE, FedSM, …
EGI-Engage, Indico DataCloud, EPOS, CTA, PRACE, H2020….
Publications
26 papers on PL-Grid Project results
36 paperson PLGrid Plus Project results147 authors, 76 reviewers
14
Journal Publications (subjective selection)
Journal IF
J.Chem.Theor.Phys.Appl. 5.31
Phys.Lett. B 6,019
J.High Energy Phys. 6,22
Astonomy &Astrophys. 4,479
Inorganic Chem. 4,794
J.Org.Chem. 4,638
Optic Lett. 3,179
Appl.Phys.Lett. 3.515
J.Comput.Chem. 3,601
J.Phys.Chem. B 3,377
Soft Matter 4,151
Int.J.Hydrogen Energy 2,93
Physica B 1,133
Journal IF
J.Chem.Phys. 3,122
J.Phys.Chem.Lett. 6,687
Phys.Chem.Chem.Phys. 4,638
Fuel Processing Techn. 3,019
J.Magn. & Magn. Mat. 2,002
Eur.J.Inorg.Chem. 2,965
Chem.Phys.Lett. 1,991
Phys.Rev.B 3,664
Eur.Phys.J. 2,421
Future Gen.Comp.Syst. 2,639
J.Phys.Chem. C 4,835
Crystal Growth & Desing 4,558
Journal IF
Macromolecules 5,927
Astrophys.J.Lett. 5,602
Phys.Rev.Letters 7,728
J.Chem.Theor.Appl. 5,31
Astrophys.J 6,28
Chem.Physics 2,028
Molec.Pharmaceutics 4,787
Eur.J.Pharmacology 2,684
Energy 4,159
Carbon 6,16
J.Biogeography 4,969
Electrochem.Comm. 4,287
J.Magn.&Magn.Mat. 1,892Conferences:
•Cracow Grid Workshop (since 2001)
•KU KDM (since 2008)
15
Summary of Projects Results(up-to-date)
# users
# users’ grants(active)
16
Summary of Projects Results(up-to-date)
Examples of active grants
PROTMD (18.9.2015-18.9.2016) – Cyfronet Research on proteins using MD25 mln hours (2,800 cores)
PCJ2015GA (26.8.2015-31.12.2015) – ICMResearch on connectome of nematodes using GA15 mln hours (6,000 cores)
PSB (1.3.2015-1.3.2016) – TASK, Cyfronet, ICMM, WCSSNew characteristics of DNA in the context of tumor therapy11 mln hours (1,200 cores)
17
Summary of Projects Results(up-to-date)
18
Deployed PLGrid IT Platforms and Tools– selected examples (by Cyfronet)
19
GridSpace A platform for e-Science applications
Experiment: an e-science application composed of code fragments (snippets), expressed in either general-purpose scripting programming languages, domain-specific languages or purpose-specific notations. Each snippet is evaluated by a corresponding interpreter.
GridSpace2 Experiment Workbench: a web application - an entry point to GridSpace2. It facilitates exploratory development, execution and management of e-science experiments.
Embedded Experiment: a published experiment embedded in a web site.
GridSpace2 Core: a Java library providing an API for development, storage, management and execution of experiments. Records all available interpreters and their installations on the underlying computational resources.
Computational Resources: servers, clusters, grids, clouds and e-infrastructures where the experiments are computed.
Contact: E. Ciepiela, D. Harężlak, M. Bubak
20
InSilicoLab science gateway framework
Goals
Complex computations done in non-complex way
Separating users from the concept of jobs and the infrastructure
Modelling the computation scenarios in an intuitive way
Different granularity of the computations
Interactive nature of applications
Dependencies between applications
Summary
The framework proved to be an easy way to integrate new domain-specific scenarios
Even if done by external teams
Natively supports multiple types of computational resources
Including private resources – e.g. private clouds
Supports various types of computations
Architecture of the InSilicoLab framework: Domain Layer
Mediation Layer with its Core ServicesResource Layer with different kinds of workers
Different kinds of users different kinds of resources
Contact: J. Kocot, M. Sterzel, T. Szepieniec
21
DataNetcollaborative metadata management
Objectives
Provide means for ad-hoc metadata model creation and deployment of corresponding storage facilities
Create a research space for metadata model exchange and discovery with associated data repositories with access restrictions in place
Support different types of storage sites and data transfer protocols
Support the exploratory paradigm by making the models evolve together with data
Architecture
Web Interface is used by users to create, extend and discover metadata models
Model repositories are deployed in the PaaS Cloud layer for scalable and reliable access from computing nodes through REST interfaces
Data items from Storage Sites are linked from the model repositories
Contact: E. Ciepiela, D. Harężlak, M. Bubak
22
Onedatatransparent access to data
A system that provides a unified and efficient access to data stored in organizationally distributed environments.
Provides a uniform and coherent view on all data stored on the storage systems distributed across the infrastructure
Supports working in groups by creation of an easy-to-use shared workspace for each group.
Serves data efficiently
Onedata Global Registry
Contact: Ł. Dutka
23
Scalarmdata farming experiments
Self-scalable platform for parametric studies
Adapting to experiment size and simulation type
Exploratory approach for conducting experiments
Supporting online analysis of experiment partial results
Integrates with clusters, Grids, Clouds
Data farming experiments with an exploratory approach
What problems are addressed with Scalarm ?Scalarm overview
Scalarm Graphical User Interface
Parameter space generation with support of design of experiment methods
Accessing heterogeneous computational infrastructure
75% all submitted tasks
Self-scalability of the management/execution partsContact: R. Słota
24
Rimrockaccess to resources
A service which simplifies the management of processes and tasks executed in the PLGrid infrastructure.
Contact: D. Harężlak
Rimrock architecture
simplicity – non-complicated integration with other applications, scripts and services
interactivity – a user can modify working processes based on indirect results
universalism – supported by many programming languages
versatility – it allows to execute an application in a batch mode or start an interactive application
user friendliness – it does not require advanced knowledge (basic information about Bash shell and curl command are sufficient to start using it)
Rimrock features
25
Cloud Computing
The Cloud increases elasticity of research, as scientists can tune the virtual machines to their specific needs.
The catalogue of VMs offered by PL-Grid contains many OSs.
Cloud platform is also the best and in many cases the only solution for running jobs with legacy software packages.
Open Nebula migration to Open Stack, ….
Cloud Platform for VPH-Share applications (Atmoshere env.)
IaaS, PaaS, STaaS….
Contact: J. Meizner, T. Szepieniec, M. Radecki
26
Cloud environment for VPH-Share app. Portal and Atmosphere
27
Applications Catalog service
Objective:to present in one place and in a uniform manner the current offer of the software available in the PLGrid infrastructure, broken down into supercomputing centers, clusters as well as categories and areas of application.
Applications Catalog is a system collecting and providing information on the applications, development tools and libraries offered in the PLGrid infrastructure. It allows to search for applications, check the status of their operation, obtain information about changes and updates, as well as it provides documentation and examples of usage. It is
designed for all those interested in the use of the applications available in the PLGrid infrastructure.
28
Map-Reduce service
Apache Spark 1.5.0 functionality:
API, RDD, DataFrame, SQL
Backend Execution: DataFrame and SQL
Integrations: Data Sources, Hive, Hadoop, Mesos and Cluster Management
R Language
Machine Learning and Advanced Analytics
Spark Streaming
29
Summary and Conclusions
Three dimensions of development:HPC/GRID/CLOUDs
Data & Knowledge layer
Network & Future Internet
Deployments have the national scope; however with close European links
Development oriented on end-users & research projects
Achieving synergy between research projects and e-infrastructures by close cooperation and offering relevant services
Durability at least 5 years after finishing the projects - confirmed in contracts
Future plans: continuation of development
Center of Excellence
CGW, KUKDM as places to exchange experience and for collaboration between eScience centers in Europe
30
More information
Please visit our Web pages:
http://www.plgrid.pl/en
http://www.plgrid.pl
CREDITS!
31
Credits
ACC Cyfronet AGHMichał TurałaMarian BubakKrzysztof ZielińskiKarol KrawentekAgnieszka SzymańskaMaciej TwardyAngelika Zaleska-WalterbachAndrzej OziębłoZofia MosurskaMarcin RadeckiRenata Słota Tomasz GubałaDarin NikolowAleksandra PałukPatryk LasońMarek MagryśŁukasz Flis
ICMMarek NiezgódkaPiotr BałaMaciej Filocha
PCSSMaciej StroińskiNorbert MeyerKrzysztof Kurowski Tomasz Piontek Paweł Wolniewicz
WCSSJacek OkoJózef JanyszekMateusz TykierkoPaweł DziekońskiBartłomiej Balcerek
TASKRafał TylmanMścislaw NakoniecznyJarosław Rybicki
… and many others…..
Special thanks to manydomain experts !