48
1 Enabling immediate access to Earth science models through cloud 1 computing: application to the GEOS-Chem model 2 3 Jiawei Zhuang 1 , Daniel J. Jacob 1 , Judit Flo Gaya 1 , Robert M. Yantosca 1 , Elizabeth W. Lundgren 1 , Melissa 4 P. Sulprizio 1 , Sebastian D. Eastham 2 5 6 1. John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA 7 2. Laboratory for Aviation and the Environment, Massachusetts Institute of Technology, Cambridge, MA 8 9 Corresponding Author Information: Jiawei Zhuang ([email protected]) 10 Abstract 11 Cloud computing platforms can provide fast and easy access to complex Earth science models and large 12 datasets. This article presents a mature capability for running the GEOS-Chem global 3-D model of 13 atmospheric chemistry (http://www.geos-chem.org) on the Amazon Web Services (AWS) cloud. GEOS- 14 Chem users at any experience level can get immediate access to the latest, standard version of the model 15 in a pre-configured software environment with all needed meteorological and other input data, and they 16 can analyze model output data easily within the cloud using Python tools in Jupyter notebooks. Users 17 with no prior knowledge of cloud computing are provided with easy-to-follow, step-by-step instructions 18 (http://cloud.geos-chem.org). They can learn how to complete a demo project in less than one hour, and 19

Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

1

Enabling immediate access to Earth science models through cloud 1

computing: application to the GEOS-Chem model 2

3

Jiawei Zhuang1, Daniel J. Jacob1, Judit Flo Gaya1, Robert M. Yantosca1, Elizabeth W. Lundgren1, Melissa 4

P. Sulprizio1, Sebastian D. Eastham2 5

6

1. John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA 7

2. Laboratory for Aviation and the Environment, Massachusetts Institute of Technology, Cambridge, MA 8

9

Corresponding Author Information: Jiawei Zhuang ([email protected]) 10

Abstract 11

Cloud computing platforms can provide fast and easy access to complex Earth science models and large 12

datasets. This article presents a mature capability for running the GEOS-Chem global 3-D model of 13

atmospheric chemistry (http://www.geos-chem.org) on the Amazon Web Services (AWS) cloud. GEOS-14

Chem users at any experience level can get immediate access to the latest, standard version of the model 15

in a pre-configured software environment with all needed meteorological and other input data, and they 16

can analyze model output data easily within the cloud using Python tools in Jupyter notebooks. Users 17

with no prior knowledge of cloud computing are provided with easy-to-follow, step-by-step instructions 18

(http://cloud.geos-chem.org). They can learn how to complete a demo project in less than one hour, and 19

Page 2: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

2

from there they can configure and submit their own simulations. The cloud is particularly attractive for 20

beginning and occasional users who otherwise may need to spend substantial time configuring a local 21

computing environment. Heavy users with their own local clusters can also benefit from the cloud to 22

access the latest standard model and datasets, share simulation configurations and results, benchmark 23

local simulations, and respond to surges in computing demand. Software containers allow GEOS-Chem 24

and its software environment to be moved smoothly between cloud platforms and local clusters, so that 25

the exact same simulation can be reproduced everywhere. Because the software requirements and 26

workflows tend to be similar across Earth science models, the work presented here provides general 27

guidance for porting models to cloud computing platforms in a user-accessible way. 28

Capsule 29

Cloud computing provides fast and easy access to the comprehensive GEOS-Chem atmospheric 30

chemistry model and its large input datasets for the international user community. 31

Introduction 32

Cloud computing involves on-demand access to a large remote pool of computing and data resources, 33

typically through a commercial vendor. It has considerable potential for Earth science modeling (Vance 34

et al. 2016). Cloud computing addresses three common problems researchers face when performing 35

complex computational tasks: compute, software, and data. Public cloud computing platforms like 36

Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform allow users to request 37

computing resources on demand and only pay for the computing time they consume, without having to 38

Page 3: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

3

invest in local computing infrastructure (compute problem). Due to the use of virtual machines (VMs) on 39

cloud platforms, it is very easy to replicate an existing software environment, so researchers can avoid 40

configuring software from scratch, which can often be difficult and time-consuming (software problem). 41

Large volumes of data can be quickly shared and processed in the cloud, saving researchers the time of 42

downloading data to local machines and the cost of storing redundant copies of data (data problem). Yet 43

cloud computing has made little penetration in Earth science modeling so far because of several 44

roadblocks. Here we show how these roadblocks can be removed, and we demonstrate practical user-45

oriented application with the GEOS-Chem atmospheric chemistry model which is now fully functional 46

and user-accessible on the AWS cloud. 47

48

A number of Earth science models have been tested in the cloud environment, including the MITgcm 49

(Evangelinos and Hill 2008), the Weather Research and Forecast Model (WRF) (Withana et al. 2011; 50

Molthan et al. 2015; Duran-Limon et al. 2016; Siuta et al. 2016), the Community Earth System Model 51

(CESM) (Chen et al. 2017), the NASA GISS ModelE (Li et al. 2017), the Regional Ocean Modelling 52

System (ROMS) (Jung et al. 2017), and the Model for Prediction Across Scales (MPAS) (Coffrin et al. 53

2019). Extensive studies have benchmarked the computational performance of cloud platforms against 54

traditional clusters (Walker 2008; Hill and Humphrey 2009; Jackson et al. 2010; Gupta and Milojicic 55

2011; Yelick et al. 2011; Zhai et al. 2011; Roloff et al. 2012, 2017; Strazdins et al. 2012; Gupta et al. 56

2013; Mehrotra et al. 2016; Salaria et al. 2017). Cloud platforms are found to be efficient for small-to-57

medium sized simulations with less than 100 CPU cores, but the typically slower inter-node 58

communication on the cloud can affect the parallel efficiency of larger simulations. Cost comparisons 59

Page 4: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

4

between cloud platforms and traditional clusters show inconsistent results, either in favor of the cloud 60

(Roloff et al. 2012; Huang et al. 2013; Oesterle et al. 2015; Thackston and Fortenberry 2015; Dodson et 61

al. 2016) or local clusters (Carlyle et al. 2010; Freniere et al. 2016; Emeras et al. 2017; Chang et al. 2018), 62

depending on assumptions regarding resource utilization, parallelization efficiency, storage requirement, 63

and billing model. The cloud is particularly cost-effective for occasional or intermittent workloads. 64

65

For complex model code, cloud platforms can considerably simplify the software configuration process. 66

On traditional machines, researchers need to properly configure library dependencies like HDF5, 67

NetCDF, and MPI, and this configuring process is becoming more and more difficult due to the growing 68

use of complicated software frameworks like ESMF (Hill et al. 2004) and NUOPC (Carman et al. 2017). 69

On cloud platforms, users can simply copy the exact software environment from an existing system 70

(virtual machine). Once a model is built, configured, and made available on the cloud by the developer, 71

it can be immediately shared with everyone. An example is the OpenFOAM Computational Fluid 72

Dynamics software officially distributed through the AWS cloud (https://cfd.direct/cloud/). 73

74

Cloud computing also greatly enhances the accessibility of Earth science datasets. Earth observing 75

systems and model simulations can produce Terabytes (TBs) or Petabytes (PBs) of data, and downloading 76

these data to local machines is often impractical. Instead of “moving data to compute”, the new paradigm 77

should be “moving compute to data”, i.e. perform data analysis in the cloud computing environment 78

where the data are already available (Yang et al. 2017). For example, NOAA’s Next Generation Weather 79

Radar (NEXRAD) product is shared through the AWS cloud (Ansari et al. 2017), and “data access that 80

Page 5: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

5

previously took 3+ years to complete now requires only a few days” (NAS 2018). NASA’s Earth 81

Observing System Data and Information System (EOSDIS) plans to move PBs of Earth observation data 82

to the AWS cloud to enhance data accessibility (Lynnes et al. 2017). Other Earth science datasets such as 83

NASA’s Earth Exchange (NEX) data, NOAA’s GOES-16 data, and ESA’s Sentinel-2 data are publicly 84

available on the AWS cloud (https://aws.amazon.com/earth/). The Google Earth Engine (Gorelick et al. 85

2017) and Climate Engine (Huntington et al. 2017) are another examples of cloud platforms that provide 86

easy access to various Earth science data collections as well as the computing power to process the data. 87

88

Computing on the cloud further facilitates reproducibility in research (Howe 2012; de Oliveira et al. 89

2017). Scientific journals increasingly require that model source code and data be available online (Irving 90

2016). However, due to complicated software dependencies of Earth science models, the configuration 91

scripts for one system would usually require significant modifications to work on other systems, and 92

sometimes the platform differences can lead to differences in simulation results (Hong et al. 2013; Li et 93

al. 2016). It is difficult to reproduce a model simulation even if the source code is published online. Cloud 94

platforms can solve this problem by guaranteeing a consistent system environment for different research 95

groups, providing massive computing power to rerun a model, and sharing large volumes of input/output 96

data. 97

98

Our guiding example in this article is the GEOS-Chem global 3-D model of atmospheric chemistry 99

(http://www.geos-chem.org), which we have made available for users on the AWS cloud. GEOS-Chem 100

was originally described by Bey et al. (2001) and is presently used by over 150 registered research groups 101

Page 6: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

6

worldwide (GEOS-Chem 2018a) for a wide range of applications in air quality, aerosol microphysics and 102

radiative forcing, global tropospheric-stratospheric chemistry, biogeochemical cycling of persistent 103

pollutants, and budgets of greenhouse gases. For some recent applications see, e.g., Christian et al., 2018; 104

Jeong and Park, 2018; Jing et al., 2018; Tian et al., 2018; Yu et al., 2018; Zhu et al., 2018. GEOS-Chem 105

is driven by assimilated meteorological data from the Goddard Earth Observing System (GEOS) of the 106

NASA Global Modeling and Assimilation Office (GMAO). It can operate in global or nested mode at 107

resolutions as fine as 25 km in latitude-longitude or cubed-sphere grids (Eastham et al. 2018). The GEOS-108

Chem code (https://github.com/geoschem) is open-access and undergoes continual development by its 109

users, with updates to the standard model overseen by an international GEOS-Chem Steering Committee 110

(GEOS-Chem 2018b). New versions are released and benchmarked every few months by the GEOS-111

Chem Support Team of scientific programmers based at Harvard University. 112

113

Our porting of GEOS-Chem to the cloud was motivated by the need to serve a diverse and growing base 114

of model users, many with little access to high-performance computing (HPC) resources and software 115

engineering support. This includes novice users needing easy access to the model for occasional 116

computations such as interpreting data from a field campaign, determining the atmospheric implications 117

of laboratory measurements, or specifying boundary conditions for urban/regional models. Despite intent 118

for GEOS-Chem to be an easy-to-use facility, the model is becoming increasingly complicated because 119

of more comprehensive scientific schemes, higher grid resolution, larger input datasets, and more 120

extensive software infrastructure to support these advances. Data transfer is an increasing problem as the 121

model resolution increases. GEOS-Chem users presently have access to 30 TB of model input data on 122

Page 7: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

7

FTP servers at Harvard University. With a typical bandwidth of 1 MB/s, it takes two weeks to download 123

a 1-TB subset and a year to download the full 30 TB. This is a significant bottleneck for research progress. 124

125

To solve this problem, we have made GEOS-Chem available through the AWS cloud, with the exact same 126

code and software environment as the standard benchmarked version managed at Harvard University. All 127

GEOS-Chem input data are now hosted on the AWS cloud under the AWS public dataset program 128

(https://registry.opendata.aws/geoschem-input-data/), and can be quickly accessed by users in the cloud 129

with no additional charge. The data repository includes the GEOS-Forward Processing (GEOS-FP) 130

meteorological product at global 0.25° × 0.3125° resolution (2012-present), the Modern-Era 131

Retrospective analysis for Research and Applications version 2 (MERRA-2) meteorological product at 132

global 0.5° × 0.625° resolution (2000-present), and coarser-resolution and nested domain versions. The 133

repository also contains a large collection of emission inventories managed through the Harvard–NASA 134

Emission Component (Keller et al. 2014), and other files needed for driving GEOS-Chem simulations 135

such as the model initial conditions. Fig. 1 shows examples of these datasets. With already-available 136

datasets as well as a pre-configured software environment on the AWS cloud, a beginning user can start 137

GEOS-Chem simulations immediately and with confidence that the results will replicate those of the 138

latest standard benchmarked version. To facilitate cloud adoption by users, we provide a detailed user 139

manual with step-by-step instructions for a complete research workflow (http://cloud.geos-chem.org/), 140

assuming no prior knowledge of cloud computing. Because the software requirements and workflows 141

tend to be similar between Earth science models, our work provides general guidance beyond GEOS-142

Chem for porting models to cloud computing platforms in a user-accessible way. 143

Page 8: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

8

Cloud computing for research: removing the roadblocks 144

Three practical roadblocks have hindered scientists from exploiting cloud computing: licensing for 145

commercial software, concerns about potential vendor lock-in, and lack of science-oriented 146

documentation and tooling. Here we show how these first two issues have been effectively addressed by 147

the availability of open-source software and HPC containers. We address the third issue by our own 148

development of documentation targeted at Earth scientists. 149

Open-source software 150

The licensing of proprietary software has been a major roadblock to cloud adoption (Section 3.2.2 of 151

Netto et al., 2017). Commercial software programs such as Intel compilers and MATLAB are often pre-152

installed on supercomputing centers and university clusters, but using the same software on cloud 153

platforms requires researchers to bring their own licenses, whose cost can be prohibitive. Further, 154

although an advantage of the cloud is the ability to share the entire software environment with users, it is 155

harder to share a system containing proprietary software1. 156

157

The increasing availability and capability of open-source software is rapidly obviating the need for 158

proprietary software in Earth science. In particular, GEOS-Chem has recently freed itself of the need for 159

proprietary software by adopting Python libraries for data analysis, and by making code changes to ensure 160

1 Strictly speaking, the Intel compiler license is only required for compiling source code, not for running pre-compiled applications. However, it is common for GEOS-Chem users to modify the model source code, for purposes like implementing new schemes, saving custom data fields, and debugging, so the ability to recompile source code is important.

Page 9: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

9

compatibility with the GNU Fortran compiler. Data analysis in GEOS-Chem had historically relied on 161

GAMAP (http://acmg.seas.harvard.edu/gamap/), a data analysis package written in the proprietary 162

Interactive Data Language (IDL) that cannot be easily used and distributed on the cloud. With the maturity 163

of the open-source scientific Python stack in recent years (VanderPlas 2016), GEOS-Chem data analysis 164

has migrated to Python. Existing Python libraries can easily replicate, and often surpass, the 165

functionalities in MATLAB and IDL. Commonly used Python libraries for GEOS-Chem users include 166

Jupyter notebooks (Shen 2014; Perkel 2018) for user interface, Xarray (Hoyer and Hamman 2017) for 167

conveniently manipulating NetCDF files, Dask (Rocklin 2015) for parallel computation, Matplotlib 168

(Hunter 2007) and Cartopy (Met Office 2016) for data visualization, and xESMF (Zhuang 2018) for 169

transforming data between different grids. We provide a Python tutorial for the GEOS-Chem user 170

community at https://github.com/geoschem/GEOSChem-python-tutorial; the tutorial code can be 171

executed in a pre-configured Python environment on the cloud platform provided freely by the Binder 172

project (https://mybinder.org). Most contents in the tutorial are general enough to be applied to other 173

Earth data analysis problems. 174

175

Atmospheric models are typically written in Fortran that can in principle accommodate different Fortran 176

compilers. In practice, compilers have different syntax requirements. Earlier versions of GEOS-Chem 177

were intended for the proprietary Intel compiler, but failed to compile with the open-source GNU 178

compiler due to invalid syntaxes in legacy modules. With a recent refactor of legacy code, GEOS-Chem 179

now compiles with all major versions of the GNU Fortran compiler (GEOS-Chem 2018c). Although some 180

models like WRF are found to be significantly slower with GNU compilers than with Intel compilers 181

Page 10: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

10

(Langkamp 2011; Siuta et al. 2016), for GEOS-Chem we find that switching to the GNU compiler 182

decreases performance by only 5% (GEOS-Chem 2018d), in part due to a sustained effort to only bring 183

hardware-independent and compiler-agonistic optimizations into the GEOS-Chem main code branch. 184

185

HPC containers 186

Being locked-in by a particular cloud vendor is a major concern for researchers (Section 5 of Bottum et 187

al., 2017), who may then be hostage to rising costs and unable to take advantage of cheap computing 188

elsewhere. It is indeed highly desirable for researchers to be able to switch smoothly between different 189

cloud platforms, supercomputing centers, and their own clusters, depending on the costs, the available 190

funding, the location of data repositories, and their collaborators (ECAR 2015a). The container 191

technology (Boettiger 2014) enables this by allowing immediate replication of the same software 192

environment on different systems. 193

194

From a user’s perspective, containers behave like virtual machines (VMs) that can encapsulate software 195

libraries, model code, and small data files into a single “image”. VMs and containers both allow platform-196

independent deployment of software. While VMs often incur performance penalties due to running an 197

additional guest operating system (OS) inside the original host OS, containers run on the native OS and 198

can achieve near-native performance (Hale et al. 2017). Docker (https://www.docker.com) is the most 199

widely-used container and is readily available on the cloud, but it cannot be used on shared HPC clusters 200

due to security risks (Jacobsen and Canon 2015). To address Docker’s limitations, HPC containers such 201

Page 11: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

11

as Shifter (Canon and Jacobsen 2016), CharlieCloud (Priedhorsky et al. 2017), and Singularity (Kurtzer 202

et al. 2017) have been recently developed to allow secure execution of Docker images on shared HPC 203

clusters. It is now increasingly standard for large HPC clusters to be equipped with software containers. 204

For example, Harvard’s Odyssey cluster supports the Singularity container (Harvard RC 2018) and 205

NASA’s Pleaides cluster supports the CharlieCloud container (NASA HECC 2018). All these different 206

containers are compatible as they can all execute Docker images. 207

208

A GEOS-Chem user might want to perform initial pilot simulations on the AWS cloud, and then switch 209

to their own clusters for more intensive simulations. We provide container images on Docker Hub 210

(https://hub.docker.com) with pre-configured GEOS-Chem for users to download to their own clusters. 211

The container provides exactly the same software environment as used on the cloud, so the model is 212

guaranteed to compile and execute correctly when ported to the local cluster, and even achieve bit-wise 213

reproducibility (Hacker et al. 2017). 214

215

Science-oriented documentation 216

Cloud-computing was initially developed for business IT applications, and interest in the cloud for 217

scientific computing is more recent (Fox 2011). According to the US National Academy of Sciences 218

(NAS 2018), “the majority of U.S. Earth science students and researchers do not have the training that 219

they need to use cloud computing and big data”. Standard cloud platform documentations (Amazon 220

2018a) are largely written for system administrators and web developers, and can be difficult for scientists 221

Page 12: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

12

to understand. Some research-oriented cloud materials (Amazon 2018b; Microsoft 2018a) and textbooks 222

(Foster and Gannon 2017) have eased the learning curve but are not specific enough to guide Earth science 223

applications. The Serverless Computing paradigm (Jonas et al. 2019) has the potential to greatly simplify 224

the use of cloud computing and reduce the system administration burden on users, but its use for scientific 225

computing is still at infancy. 226

227

To make cloud computing accessible to scientists, several high-level, black-box services have been 228

proposed to hide the technical details of cloud platforms (Tsaftaris, 2014; Hossain et al., 2017; Li et al., 229

2017; Section 2.3 of Netto et al., 2017). However, they tend to make strict assumptions on the workflow 230

and lack customizability and generalizability. We take the opposite tack -- to teach low-level AWS cloud 231

concepts to users and make them manage their own cloud infrastructures including servers, storage, and 232

network. In cloud computing terminology, we stick to the Infrastructure as a Service (IaaS) framework, 233

not the higher-level Software as a Service (SaaS) framework. 234

235

Our research-oriented tutorial and documentation (http://cloud.geos-chem.org/) provide the necessary 236

training for GEOS-Chem users, with practical focus on model simulation and data management. The 237

documentation includes “beginner tutorials” for beginning and occasional users, “advanced tutorials” for 238

heavy users, and a “developer guide” for model developers to install software libraries and configure 239

environment from scratch. By following the documentation, GEOS-Chem users can successfully finish a 240

demo project in half an hour for their first experience on the AWS cloud, and within minutes for 241

subsequent uses. This is detailed in the next section. 242

Page 13: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

13

243

Exposing low-level infrastructures makes our work highly generalizable to other Earth science 244

applications and cloud platforms. First, users have the freedom to install and run any model by using our 245

GEOS-Chem workflow as a reference. Second, once users get familiar with AWS concepts, it becomes 246

relatively easy for them to learn other cloud platforms like Microsoft Azure and Google cloud, because 247

these different platforms have similar services and technical concepts (Google 2018a; Microsoft 2018b). 248

Third, the same set of cloud computing skills can be applied to a wide range of problems beyond model 249

simulations, such as analyzing large public Earth science data or using GPUs on the cloud to accelerate 250

machine learning workloads (Appendix B of Chollet, 2017). 251

Research workflow using basic AWS functionalities 252

Here we describe the workflow for GEOS-Chem users in two scenarios: 253

1. A demo project for initiation to the cloud. The users run our pre-configured “demo simulation” for a 254

short period. The demo is configured with GEOS-Chem’s standard troposphere-stratosphere oxidant-255

aerosol chemical mechanism and a global horizontal resolution of 4° × 5°. This same configuration 256

is used for the official GEOS-Chem benchmark (GEOS-Chem 2018e). It is fast to simulate due to the 257

coarse spatial resolution and is a good starting point for a new user. The users perform quick analysis 258

and visualization of the model output data, and then exit the system discarding all data files. 259

2. An actual research project for scientific analysis. The users set up their desired custom configuration 260

of the model, conduct their simulation, archive their output, and preserve the model configuration for 261

future runs. For illustrative and cost evaluation purposes, we assume here a 1-year global simulation 262

Page 14: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

14

at 2° × 2.5° resolution using the standard troposphere-stratosphere oxidant-aerosol chemical 263

mechanism. For output, the simulation stores 3-D daily averaged fields for 168 transported chemical 264

species, resulting in 150 GB of data. This is typical of a research project using GEOS-Chem. The 265

same workflow applies to any other configuration of the model. 266

267

The workflow uses the most basic AWS functionalities to keep the learning curve minimum. AWS offers 268

over a hundred services (Amazon 2018c), leading to a complicated web portal that can be daunting for 269

new users. However, Earth science modeling tasks can be achieved with only two core services: Elastic 270

Compute Cloud (EC2) for computation and Simple Storage Service (S3) for data storage. Many other 271

services target various IT/business applications and are not necessary for most scientific users. The steps 272

for a complete modeling research workflow are illustrated in Fig. 2 and are explained in what follows. 273

274

Step 1: Launch virtual server 275

Under the IaaS framework, users request their own servers (“EC2 instances” in AWS terminology) with 276

customizable hardware capacity and software environment. Requests are done through the AWS console 277

in the user’s web browser. The software environment is determined by a virtual machine image (the 278

Amazon Machine Image, or AMI), which defines the operating system, pre-installed software libraries, 279

and data files that a newly-launched EC2 instance will contain. Here, users can select a publicly available 280

AMI with pre-configured GEOS-Chem environment. The hardware capacity is defined by the choice of 281

EC2 instance type (Amazon 2018d) with different capacities in CPUs, memory, disk storage, and network 282

bandwidth. Numerical models run most efficiently on “Compute-Optimized” types, which prioritize CPU 283

Page 15: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

15

performance over other aspects (Montes et al.,2017). Launching a new EC2 instance only takes seconds. 284

Repeated launches of EC2 instances can be automated by the AWS Command Line Interface (AWSCLI, 285

https://aws.amazon.com/cli/), to avoid having to browse through the web console. 286

287

Step 2: Log into server and perform computation. 288

Once the EC2 instance is created, it can be used as a normal server, i.e. via the Secure Shell (SSH) in the 289

command line terminal. Since most Earth scientists have experience with local servers and the command 290

line, using EC2 instances is straightforward for them; if not, they can learn the command line basics easily 291

from online materials like Software Carpentry (http://swcarpentry.github.io/shell-novice/). There are no 292

cloud-computing-specific skills required at this stage. 293

294

Following standard practice for Earth science models, a GEOS-Chem simulation is controlled by a “run 295

directory” containing run-time configurations and the model executable. A pre-configured run directory 296

with a pre-compiled executable and a sample of meteorological and emission input data are provided by 297

our GEOS-Chem AMI for the demo project, so users can execute a GEOS-Chem simulation immediately 298

after log-in, without having to set up the run directory and compile model code on their own. This short 299

demo simulation finishes in minutes. 300

301

For an actual research project, users should re-compile the model with the desired model version, grid 302

resolution, and chemical mechanism. Recompilation takes about 2 minutes. Because the software libraries 303

and environment variables are already properly configured, the model is guaranteed to compile without 304

Page 16: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

16

error. Additional operations such as modifying model source code or installing new software libraries can 305

be done as usual, just like on local servers. 306

307

GEOS-Chem users also need to access additional meteorological and emission input data for their desired 308

simulation configurations, and they can do so from the public GEOS-Chem input data repository residing 309

in the AWS Simple Storage Service (S3). The data transfer from S3 to EC2 is simply done by AWSCLI 310

command “aws s3 cp”, analogous to the “cp” command for copying data on local filesystems. For the 311

actual research project scenario described here, the user should retrieve 1 year of global 2°×2.5° 312

meteorological input data with size of 112 GB. With a typical ~250 MB/s network bandwidth between 313

S3 and EC2, the retrieval will finish in 8 minutes. 314

315

Users of shared HPC clusters are accustomed to using job schedulers to manage long model simulations. 316

However, an EC2 instance completely belongs to a single user, so a scheduler for queuing jobs is not 317

necessary. Instead, users just execute the program interactively. A simple way to keep the program 318

running after logging out of the server is to execute the model inside terminal multiplexers such as GNU 319

Screen (https://www.gnu.org/software/screen/) or tmux (https://github.com/tmux/tmux), which are 320

standard tools for managing long-running programs on the cloud (Shaikh 2018). 321

322

GEOS-Chem simulations generate potentially large amounts of output data in NetCDF format (GEOS-323

Chem 2018f). Analysis of output data can be done within the cloud using Python in web-based Jupyter 324

notebooks. Jupyter always displays the user interface, code, and graphics in the web browser, no matter 325

Page 17: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

17

whether the program is running on the local computer or a remote cloud platform. Thus, users have a 326

convenient data analysis environment on the cloud, as if they are using Jupyter locally. Alternatively, the 327

data can be downloaded to local computers for analysis, although this involves data transfer time and a 328

data egress fee (see next section). 329

330

For a demo project, the user can simply terminate the EC2 instance after the model simulation and output 331

data analysis are finished. The output data are deleted upon termination. For an actual research project, 332

the output data will generally need to be archived before EC2 termination, as described next. 333

334

Step 3: Working with persistent storage 335

The lack of persistent disk storage is the major difference between cloud platforms and local computers. 336

The Elastic Block Store (EBS) volume is the temporary disk that backs up an EC2 instance. However, 337

when the EC2 instance is terminated, its EBS volume containing the user’s data will typically be deleted. 338

Instead of leaving files on EBS, important files such as output data and customized run directories should 339

be transferred to the S3 persistent storage service. S3 is independent of EC2 and guarantees the persistence 340

of data. 341

342

S3 offers many other advantages over EBS besides data persistence. While an EBS volume can only be 343

attached to single EC2, data in S3 can be accessed simultaneously by any number of EC2 instances, as 344

well as by computers outside of the cloud. Files in an EBS volume can only be seen from an EC2 instance, 345

but files in S3 can be viewed directly in the AWS web portal. While an EBS volume has a limited storage 346

Page 18: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

18

capacity just like traditional disks, there is effectively no size limit on S3. Furthermore, S3 can be 347

integrated with advanced AWS services (e.g., Amazon Elastic MapReduce) and community big-data 348

platforms (e.g., Pangeo, https://pangeo.io) for large-scale data processing. 349

350

Despite the advantages of S3, new users might still feel uncomfortable with managing two types of storage 351

(S3 and EBS) at the same time. This can be simplified by mounting S3 via Filesystem in Userspace 352

(FUSE, https://github.com/s3fs-fuse/s3fs-fuse), so that data in S3 appear like normal files on the server’s 353

disk. The data transfer between EC2 and S3 happens automatically under the hood when a file is accessed, 354

without requiring the user to explicitly transfer the data. However, this approach is currently very 355

inefficient for NetCDF data, due to incompatibilities between the HDF5 library and the object-based 356

storage model that S3 uses (Rocklin 2018). This problem might be improved by developing a more cloud-357

friendly backend, such as Zarr (https://github.com/zarr-developers/zarr), for the NetCDF library. For now, 358

we recommend explicitly transferring data between EC2 and S3. 359

360

Once important files are transferred to S3 (or downloaded to local storage), users can safely terminate the 361

EC2 instance. Modification to the software environment can also be saved, by snapshotting the EC2 362

instance into an AMI and then using the new AMI to launch subsequent EC2 instances. 363

364

Advantages and limitations of the workflow 365

The workflow presented here depends on very few AWS functionalities, has a light learning curve, and 366

is flexible enough to match different usage patterns. For example, it is common to repeatedly analyze 367

Page 19: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

19

output data after a model simulation. Whenever users need to access the data in S3, they can simply launch 368

a new EC2 instance, pull data from S3 to perform analysis, and terminate the EC2 instance after the 369

analysis is done. Simultaneous model runs can be done in parallel by launching multiple EC2 instances, 370

with input data replicated in each instance’s EBS volume. 371

372

We tested the demo project on 15 members of the Atmospheric Chemistry Modeling Group at Harvard 373

University, and recorded the time they spent on each step (Fig. 3). All were GEOS-Chem users but none 374

had any prior experience with cloud computing. Most finished the demo in less than 30 minutes, and all 375

in less than 60 minutes. The time represents the initial learning curve for first-time AWS users; an 376

experienced AWS user could finish the demo in one minute. An actual research project requires additional 377

steps such as configuring model details and working with S3, but the workflow is not fundamentally 378

different. Overall, it is very fast to get the model running on the cloud, in sharp contrast to the often slow 379

and complicated procedure of setting up the model on a local server. 380

381

The workflow as described above has several limitations for heavy users. Although GEOS-Chem is able 382

to parallelize across multiple compute nodes using MPI (Eastham et al. 2018), here we only execute the 383

model on a single EC2 instance (with up to 64 physical cores on the “x1.32large” instance type), 384

equivalent to a single compute node on an HPC cluster. Also, users need to manually manage the start 385

and termination of each EC2 instance, which can be an administrative burden if there are a large number 386

of simultaneous simulations. These issues may be addressed by creating an HPC cluster environment on 387

the cloud, using software tools such as AWS ParallelCluster (https://aws-parallelcluster.readthedocs.io,388

Page 20: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

20

supersedes the previous CfnCluster), StarCluster (http://star.mit.edu/cluster, unmaintained), ElastiCluster 389

(http://elasticluster.readthedocs.io), AlcesFlight (http://docs.alces-flight.com), and EnginFrame 390

(https://www.nice-software.com/products/enginframe). A cluster consists of a “master node” (typically a 391

small EC2 instance) and multiple “compute nodes” (typically high-performance EC2 instances). Such a 392

cluster allows inter-node MPI communication, provides a shared disk (EBS volume) for all nodes via the 393

Network File System (NFS), and often supports “auto-scaling” capability (Amazon 2018e) that 394

automatically launches or terminates compute nodes according to the number of pending jobs. However, 395

we find that cluster tools have a much steeper learning curve for scientists and are generally an overkill 396

for moderate computing workloads. While managing individual EC2 instances is straightforward, 397

configuring and customizing a cluster involve heavier system administration tasks. We plan to address 398

this issue in future work. 399

Performance and cost 400

Computational performance and cost compared to local clusters 401

EC2 instances incur hourly cost to users, with different pricing models. Commonly used are the standard 402

“on-demand” pricing and the discounted “spot” pricing. Spot instances are cheaper than on-demand 403

instances, usually by 60%~70%, with the caveat that they may be reclaimed by AWS to serve the demand 404

from standard EC2 users. Although such spot interruption can cause model simulations to crash, a newly 405

added “Spot Hibernation” option (Amazon 2017a) allows the instance to recover the previous state so that 406

previous simulations can continue when capacity becomes available again. A recent update to the spot 407

Page 21: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

21

pricing model further reduces the chance of interruption, so an instance is rarely reclaimed and can 408

generally keep uninterrupted for a month (Amazon 2018f). We recommend using spot pricing for 409

computationally expensive tasks. 410

411

It is challenging to compare the pay-as-you-go pricing model on the cloud with the cost of local HPC 412

clusters that vary in billing model (ECAR 2015b). To simplify such estimation, we use the NASA Pleaides 413

cluster that provides a simple, convenient billing model called Standard Billing Unit (SBU). NASA’s 414

High-End Computing Capability (HECC) Project uses this billing model to evaluate the cost of AWS 415

against the Pleaides cluster (Chang et al. 2018). Jobs on Pleaides are charged by CPU hours, with the cost 416

rate “calculated as the total annual HECC costs divided by the total number of [CPU hours used]” and 417

thus is able to represent “the costs of operating the HECC facility including hardware and software costs, 418

maintenance, support staff, facility maintenance, and electrical power costs”. Pleiades costs are not borne 419

directly by the users in that allocation of SBUs is through NASA research grants, but the costs are still 420

borne by NASA. We consider several types of AWS EC2 instances and compute nodes on the Pleiades 421

cluster with comparable performance, as summarized in Table 1. 422

423

For performance and cost comparisons we used GEOS-Chem version 12.1.1 to conduct a 7-day 424

simulation with tropospheric and stratospheric chemistry (Eastham et al. 2014) at global 4° × 5° resolution, 425

using OpenMP parallelization. We tested the model performance both on the native machine and inside 426

the software container, on both the AWS cloud (using the Singularity container) and Pleiades (using the 427

Page 22: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

22

CharlieCloud container). The performance difference introduced by running the model inside the 428

container is less than 1%, so there is no significant penalty to using containers. 429

430

The timing results and total costs are shown in Fig.4. The newer “c5” generation has better performance 431

and lower cost than the older “c4” generation. The smaller instance “c5.4xlarge” is more cost-efficient 432

than the bigger “c5.9xlarge” instance, due to the sublinear scaling of GEOS-Chem OpenMP 433

parallelization and a fixed I/O time. With spot pricing, the cost of EC2 is close to Pleiades. It is important 434

to point out that Pleaides has a very high utilization rate of over 80% (Chang et al. 2018); for other local 435

clusters with lower utilization rates, the pay-as-you-go pricing on the cloud becomes even more attractive. 436

From a user standpoint, any comparative cost decision is complicated by the varying levels of research 437

subsidy and grant-based allocation for time on local or supercomputing clusters. A better decision basis 438

is the cost of an actual research project on the cloud, and we discuss this below. 439

440

Total cost of an example project 441

Here we calculate the total cost of the actual research simulation example in the previous section (1-year 442

2 ° × 2.5 ° global GEOS-Chem simulation of tropospheric and stratospheric chemistry with daily chemical 443

fields for 168 species archived, representing 150 GB of output data). Besides the major charge with EC2, 444

minor charges include the storage cost with EBS and S3. Users also incur a “data egress fee” if they 445

download data to their local clusters. While AWS and the other major cloud vendors usually do not charge 446

for transferring data into the cloud or between cloud services, they charge for transferring data out of the 447

Page 23: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

23

cloud. AWS allows academic users to waive the data egress fee, up to 15% of the total charge to the user 448

(Amazon 2016). 449

450

Table 2 summarizes the major AWS services used in the GEOS-Chem simulation and their unit costs. 451

The same services would be used for any other Earth science model. Table 3 summarizes the total cost of 452

our example GEOS-Chem research simulation conducted on the most cost-efficient “c5.4xlarge” EC2 453

instance type. The compute cost for the 1-year simulation with EC2 is $224 with on-demand pricing, and 454

can be reduced to $90 with spot pricing (which we recommend). The temporary storage cost with EBS 455

($14) is relatively small. The long-term storage cost with S3 or the data egress fee greatly depends on 456

users’ needs. 1-year daily concentration fields for 168 chemical species on the global 3-D domain 457

represent 150 GB of data, only contributing a small amount to the total cost ($3.50 per month). However, 458

if the user needs to produce TBs of data (e.g. hourly fields for a year), the storage cost can become 459

important. The cost of a long-term archive can be greatly reduced by using Amazon S3 Glacier 460

(https://aws.amazon.com/glacier/), a cheaper version of S3 with 80% price reduction but longer data 461

retrieval time. Glacier can be a good option for long-term archiving of data after the research project has 462

been completed. 463

464

In summary, the total cost of a 1-year global GEOS-Chem simulation with 2° × 2.5° resolution and full 465

chemistry on the AWS cloud is about $120 using EC2 spot pricing. A 1-year 4° × 5° simulation costs four 466

times less ($30). The cost is mostly from the computing itself. Downloading or archiving output data is 467

Page 24: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

24

generally cheap by comparison. Transferring GEOS-Chem input data from S3 to EBS is free and fast as 468

these data are already resident on the cloud. 469

470

The cost of computing on the cloud should make it highly attractive to occasional users, particularly 471

because of the time savings in model configuration and data transfer, as well as the assurance that the 472

simulation replicates the current standard benchmarked version of the model. AWS currently offers a 473

$100 credit per year to university students (Amazon 2018g) to encourage experimentation on the cloud. 474

Heavy users may still prefer local or supercomputing clusters, particularly if grant-funded. The cloud will 475

become more attractive with a funding model comparable to the research subsidies and grant allocations 476

for local clusters. Despite special funding programs such as NSF BIGDATA (NSF 2018a), E-CAS (NSF 477

2018b), AWS Cloud Credits for Research (Amazon 2018h), Azure AI for Earth (Microsoft 2018c), and 478

Google Cloud Platform research credits (Google 2018b) that directly give credits on commercial clouds, 479

the funding is still far lagging that of traditional supercomputing centers (Section 12 of Bottum et al., 480

2017). Chapter 6.3.3 of NAS (2016) suggests several ways for NSF to better fund commercial cloud 481

resources. 482

483

Heavy users running GEOS-Chem on local or supercomputing clusters can still benefit from the cloud 484

for downloading model input data, because AWS S3 has a very high outgoing bandwidth which can often 485

fully utilize the ingoing bandwidth of a local cluster. We achieve a data transfer rate of 100 MB/s from 486

S3 to the Pleiades cluster, an order of magnitude faster than from the FTP server at Harvard. Transferring 487

1 year of global 2° × 2.5° meteorological input data (112 GB) for input to GEOS-Chem finishes in 20 488

Page 25: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

25

minutes, which would have taken 3 hours from the Harvard FTP server. The egress fee is only $10. The 489

cloud, together with software containers, can thus accelerate the deployment of GEOS-Chem on local 490

clusters as well. 491

Conclusions and future work 492

Cloud computing is becoming increasingly attractive for Earth science modeling. It provides easy access 493

to complex models and large datasets and allows straightforward and error-free computing. We have 494

described how the GEOS-Chem global model of atmospheric chemistry is now fully accessible for users 495

on the AWS cloud. The cloud gives users immediate access to the computational resources, software 496

environment, and large input data needed to perform GEOS-Chem simulations for any year and for any 497

grid resolution supported by the standard model. Single-node GEOS-Chem simulations on the AWS cloud 498

compare favorably in performance and true cost with local clusters. Scientists can learn the single-node 499

workflow on the cloud very quickly by following our project documentation and tutorial. Our specific 500

application is for GEOS-Chem but the general principles can be adapted to other Earth science models, 501

which tend to follow the same structure and requirements. 502

503

We find the cloud to be particularly attractive for beginning or occasional users, who otherwise may need 504

to spend significant personnel time configuring a local computing environment. Heavy users with their 505

own local clusters should still find the cloud useful for getting the latest model and data updates, 506

benchmarking their local simulations against the standard model resident on the cloud, carrying out 507

Page 26: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

26

collaborative research in a reproducible environment, and temporarily expanding their computing 508

capacity. 509

510

The cloud is also a promising vehicle for massively parallel simulations emulating local HPC clusters, 511

but several issues need to be addressed. A proper research funding model becomes particularly important 512

for compute-intensive simulations, as the costs can be significant. Although it is widely perceived that 513

cloud platforms are not efficient for large-scale MPI programs due to slow communication across nodes, 514

the network performance of cloud platforms has been improving rapidly over the past several years. The 515

Microsoft Azure cloud now offers InfiniBand connection and achieves similar scalability as traditional 516

supercomputers (Mohammadi and Bazhirov 2017). Recent updates on the AWS cloud, including a new 517

instance type with 100 Gb/s bandwidth (which exceeds the typical ~50 Gb/s bandwidth on local HPC 518

clusters) (Amazon 2018j), a low-latency network interface (Amazon 2018j) to improve the scaling of MPI 519

programs, and a parallel file system service for efficient I/O (Amazon 2018k), have greatly increased the 520

potential of cloud computing for HPC applications. There remain several challenges. Although many 521

HPC cluster management tools have been developed for the cloud, the learning curve of these tools is 522

steeper than for the single-node applications described here. Also, although containers are highly portable 523

in a single-node environment, they become less portable when inter-node MPI communication is needed. 524

Docker is found to have significant overhead for multi-node MPI applications (Younge et al. 2017; Zhang 525

et al. 2017); Singularity achieves better performance by directly using the MPI installed on the host 526

machine, but requires compatibility between the MPI libraries inside and outside of the container. These 527

issues will be addressed in future work. 528

Page 27: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

27

Code availability 529

The source code for our project documentation, the scripts for building AWS images, the Python 530

notebooks for generating all plots in this article, are available at our project repository 531

https://github.com/geoschem/geos-chem-cloud. Scripts for building GEOS-Chem Docker images are at 532

https://github.com/geoschem/geos-chem-docker. The survey results producing Fig. 3 are available at 533

https://github.com/geoschem/geos-chem-cloud/issues/15. 534

535

Acknowledgments. This work was funded by the NASA Earth Science Division. JZ was supported by 536

the NASA Earth and Space Science Fellowship. The authors would like to thank Kevin Jorissen, Joe 537

Flasher, and Jed Sundwall at Amazon Web Services for supporting this work. 538

References 539

Amazon, 2016: AWS Offers Data Egress Discount to Researchers. Accessed December 20, 2018, 540

https://aws.amazon.com/blogs/publicsector/aws-offers-data-egress-discount-to-researchers/. 541

——, 2017a: Amazon EC2 Spot Lets you Pause and Resume Your Workloads. Accessed December 20, 542

2018, https://aws.amazon.com/about-aws/whats-new/2017/11/amazon-ec2-spot-lets-you-pause-543

and-resume-your-workloads/. 544

——, 2017b: Announcing Amazon EC2 per second billing. Accessed December 20, 2018, 545

https://aws.amazon.com/about-aws/whats-new/2017/10/announcing-amazon-ec2-per-second-546

billing/. 547

Page 28: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

28

——, 2018a: AWS official documentation. Accessed December 20, 2018, https://docs.aws.amazon.com/. 548

——, 2018b: AWS Research Cloud Program: Researcher’s Handbook. Accessed December 20, 2018, 549

https://aws.amazon.com/government-education/research-and-technical-computing/research-cloud-550

program/. 551

——, 2018c: AWS Cloud Products. Accessed December 20, 2018, https://aws.amazon.com/products/. 552

——, 2018d: Amazon EC2 Instance Types. Accessed December 20, 2018, 553

https://aws.amazon.com/ec2/instance-types/. 554

——, 2018e: AWS Auto Scaling. Accessed December 20, 2018, https://aws.amazon.com/autoscaling/. 555

——, 2018f: New Amazon EC2 Spot pricing model: Simplified purchasing without bidding and fewer 556

interruptions. Accessed December 20, 2018, https://aws.amazon.com/blogs/compute/new-amazon-557

ec2-spot-pricing/. 558

——, 2018g: AWS Educate Program. Accessed December 20, 2018, 559

https://aws.amazon.com/education/awseducate/. 560

——, 2018h: AWS Cloud Credits for Research. Accessed December 20, 2018, 561

https://aws.amazon.com/research-credits/. 562

——, 2018i: New C5n Instances with 100 Gbps Networking. Accessed December 20, 2018, 563

https://aws.amazon.com/blogs/aws/new-c5n-instances-with-100-gbps-networking/. 564

——, 2018j: Introducing Elastic Fabric Adapter. Accessed December 20, 2018, 565

https://aws.amazon.com/about-aws/whats-new/2018/11/introducing-elastic-fabric-adapter/. 566

——, 2018k: Amazon FSx for Lustre. Accessed December 20, 2018, https://aws.amazon.com/fsx/lustre/. 567

——, 2018l: Physical Cores by Amazon EC2 and RDS DB Instance Type. Accessed December 20, 2018, 568

Page 29: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

29

https://aws.amazon.com/ec2/physicalcores/. 569

Ansari, S., and Coauthors, 2017: Unlocking the potential of NEXRAD data through NOAA’s Big Data 570

Partnership. Bull. Am. Meteorol. Soc., BAMS-D-16-0021.1, doi:10.1175/BAMS-D-16-0021.1. 571

http://journals.ametsoc.org/doi/10.1175/BAMS-D-16-0021.1. 572

Bey, I., and Coauthors, 2001: Global modeling of tropospheric chemistry with assimilated meteorology: 573

Model description and evaluation. J. Geophys. Res. Atmos., 106, 23073–23095. 574

Boettiger, C., 2014: An introduction to Docker for reproducible research, with examples from the R 575

environment. ACM SIGOPS Oper. Syst. Rev., 49, 71–79, doi:10.1145/2723872.2723882. 576

http://arxiv.org/abs/1410.0846%0Ahttp://dx.doi.org/10.1145/2723872.2723882. 577

Bottum, J., and Coauthors, 2017: The Future of Cloud for Academic Research Computing. 1-21 pp. 578

Canon, R. S., and D. Jacobsen, 2016: Shifter : Containers for HPC. Cray User Gr. 2016,. 579

Carlyle, A. G., S. L. Harrell, and P. M. Smith, 2010: Cost-effective HPC: The community or the cloud? 580

Proc. - 2nd IEEE Int. Conf. Cloud Comput. Technol. Sci. CloudCom 2010, 169–176, 581

doi:10.1109/CloudCom.2010.115. 582

Carman, J. C., and Coauthors, 2017: The national earth system prediction capability: Coordinating the 583

giant. Bull. Am. Meteorol. Soc., 98, 239–252, doi:10.1175/BAMS-D-16-0002.1. 584

Chang, S., and Coauthors, 2018: Evaluating the Suitability of Commercial Clouds for NASA ’ s High 585

Performance Computing Applications : A Trade Study. 1–46. 586

Chen, X., X. Huang, C. Jiao, M. G. Flanner, T. Raeker, and B. Palen, 2017: Running climate model on a 587

commercial cloud computing environment: A case study using Community Earth System Model 588

(CESM) on Amazon AWS. Comput. Geosci., 98, 21–25, doi:10.1016/j.cageo.2016.09.014. 589

Page 30: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

30

Chollet, F., 2017: Deep Learning with Python. Manning,. 590

Christian, K. E., W. H. Brune, J. Mao, and X. Ren, 2018: Global sensitivity analysis of GEOS-Chem 591

modeled ozone and hydrogen oxides during the INTEX campaigns. Atmos. Chem. Phys., 18, 2443–592

2460, doi:10.5194/acp-18-2443-2018. 593

Coffrin, C., and Coauthors, 2019: The ISTI Rapid Response on Exploring Cloud Computing 2018. 1–72. 594

http://arxiv.org/abs/1901.01331. 595

Dodson, R., and Coauthors, 2016: Imaging SKA-scale data in three different computing environments. 596

Astron. Comput., 14, 8–22, doi:10.1016/j.ascom.2015.10.007. 597

Duran-Limon, H. A., J. Flores-Contreras, N. Parlavantzas, M. Zhao, and A. Meulenert-Peña, 2016: 598

Efficient execution of the WRF model and other HPC applications in the cloud. Earth Sci. 599

Informatics, 9, 365–382, doi:10.1007/s12145-016-0253-7. http://link.springer.com/10.1007/s12145-600

016-0253-7. 601

Eastham, S. D., D. K. Weisenstein, and S. R. H. Barrett, 2014: Development and evaluation of the unified 602

tropospheric-stratospheric chemistry extension (UCX) for the global chemistry-transport model 603

GEOS-Chem. Atmos. Environ., 89, 52–63, doi:10.1016/j.atmosenv.2014.02.001. 604

http://dx.doi.org/10.1016/j.atmosenv.2014.02.001. 605

——, and Coauthors, 2018: GEOS-Chem high performance (GCHP v11-02c): A next-generation 606

implementation of the GEOS-Chem chemical transport model for massively parallel applications. 607

Geosci. Model Dev., 11, 2941–2953, doi:10.5194/gmd-11-2941-2018. 608

ECAR, 2015a: Research Computing in the Cloud -- Functional Considerations for Research. 609

——, 2015b: TCO for Cloud Services. 610

Page 31: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

31

Emeras, J., S. Varrette, and P. Bouvry, 2017: Amazon elastic compute cloud (EC2) vs. in-house HPC 611

platform: A cost analysis. IEEE International Conference on Cloud Computing, CLOUD, 284–293. 612

Evangelinos, C., and C. N. Hill, 2008: Cloud Computing for parallel Scientific HPC Applications: 613

Feasibility of Running Coupled Atmosphere-Ocean Climate Models on Amazon’s EC2. Cca’08, 2, 614

2–34. 615

Foster, I., and D. B. Gannon, 2017: Cloud computing for science and engineering. 616

https://mitpress.mit.edu/books/cloud-computing-science-and-engineering. 617

Fox, A., 2011: Cloud computing - What’s in it for me as a scientist? Science (80-. )., 331, 406–407, 618

doi:10.1126/science.1198981. 619

Freniere, C., A. Pathak, M. Raessi, and G. Khanna, 2016: The feasibility of amazon’s cloud computing 620

platform for parallel, GPU-accelerated, multiphase-flow simulations. Comput. Sci. Eng., 18, 68–77, 621

doi:10.1109/MCSE.2016.94. 622

GEOS-Chem, 2018a: GEOS-Chem People and Projects. Accessed December 20, 2018, 623

http://acmg.seas.harvard.edu/geos/geos_people.html. 624

——, 2018b: GEOS-Chem Steering Committee. Accessed December 20, 2018, 625

http://acmg.seas.harvard.edu/geos/geos_steering_cmte.html. 626

——, 2018c: GNU Fortran compiler. Accessed December 20, 2018, http://wiki.seas.harvard.edu/geos-627

chem/index.php/GNU_Fortran_compiler. 628

——, 2018d: Timing tests with GEOS-Chem 12.0.0. Accessed December 20, 2018, 629

http://wiki.seas.harvard.edu/geos-chem/index.php/Timing_tests_with_GEOS-Chem_12.0.0. 630

——, 2018e: GEOS-Chem benchmarking. Accessed December 20, 2018, 631

Page 32: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

32

http://wiki.seas.harvard.edu/geos-chem/index.php/GEOS-Chem_benchmarking. 632

——, 2018f: List of diagnostics archived to netCDF format. Accessed December 20, 2018, 633

http://wiki.seas.harvard.edu/geos-634

chem/index.php/List_of_diagnostics_archived_to_netCDF_format. 635

Google, 2018a: Google Cloud Platform for AWS Professionals. Accessed December 20, 2018, 636

https://cloud.google.com/docs/compare/aws/. 637

——, 2018b: Google Cloud Platform announces new credits program for researchers. Accessed 638

December 20, 2018, https://www.blog.google/products/google-cloud/google-cloud-platform-639

announces-new-credits-program-researchers/. 640

Gorelick, N., M. Hancher, M. Dixon, S. Ilyushchenko, D. Thau, and R. Moore, 2017: Google Earth 641

Engine: Planetary-scale geospatial analysis for everyone. Remote Sens. Environ., 642

doi:10.1016/j.rse.2017.06.031. http://linkinghub.elsevier.com/retrieve/pii/S0034425717302900. 643

Gupta, A., and D. Milojicic, 2011: Evaluation of hpc applications on cloud. Open Cirrus Summit (OCS), 644

2011 Sixth, 22–26. 645

——, and Coauthors, 2013: The Who , What , Why and How of High Performance Computing 646

Applications in the Cloud. HP Lab., doi:10.1109/CloudCom.2013.47. 647

Hacker, J. P., J. Exby, D. Gill, I. Jimenez, C. Maltzahn, T. See, G. Mullendore, and K. Fossell, 2017: A 648

containerized mesoscale model and analysis toolkit to accelerate classroom learning, collaborative 649

research, and uncertainty quantification. Bull. Am. Meteorol. Soc., 98, 1129–1138, 650

doi:10.1175/BAMS-D-15-00255.1. 651

Hale, J. S., L. Li, C. N. Richardson, and G. N. Wells, 2017: Containers for Portable, Productive, and 652

Page 33: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

33

Performant Scientific Computing. Comput. Sci. Eng., 19, 40–50, doi:10.1109/MCSE.2017.2421459. 653

Harvard RC, 2018: Singularity on Odyssey. Accessed December 20, 2018, 654

https://www.rc.fas.harvard.edu/resources/documentation/software/singularity-on-odyssey/. 655

Hill, C., C. DeLuca, Balaji, M. Suarez, and A. Da Silva, 2004: The architecture of the Earth system 656

modeling framework. Comput. Sci. Eng., 6, 18–28, doi:10.1109/MCISE.2004.1255817. 657

Hill, Z., and M. Humphrey, 2009: A quantitative analysis of high performance computing with Amazon’s 658

EC2 infrastructure: The death of the local cluster? Proc. - IEEE/ACM Int. Work. Grid Comput., 26–659

33, doi:10.1109/GRID.2009.5353067. 660

Hoesly, R. M., and Coauthors, 2018: Historical (1750-2014) anthropogenic emissions of reactive gases 661

and aerosols from the Community Emissions Data System (CEDS). Geosci. Model Dev., 11, 369–662

408, doi:10.5194/gmd-11-369-2018. 663

Hong, S.-Y., M.-S. Koo, J. Jang, J.-E. Esther Kim, H. Park, M.-S. Joh, J.-H. Kang, and T.-J. Oh, 2013: 664

An Evaluation of the Software System Dependency of a Global Atmospheric Model. Mon. Weather 665

Rev., 141, 4165–4172, doi:10.1175/MWR-D-12-00352.1. 666

http://journals.ametsoc.org/doi/abs/10.1175/MWR-D-12-00352.1. 667

Hossain, M. M., R. Wu, J. T. Painumkal, M. Kettouch, C. Luca, S. M. Dascalu, and F. C. Harris, 2017: 668

Web-service framework for environmental models. 2017 Internet Technol. Appl., 104–109, 669

doi:10.1109/ITECHA.2017.8101919. http://ieeexplore.ieee.org/document/8101919/. 670

Howe, B., 2012: Virtual appliances, cloud computing, and reproducible research. Comput. Sci. Eng., 14, 671

36–41, doi:10.1109/MCSE.2012.62. 672

Hoyer, S., and J. J. Hamman, 2017: xarray: N-D labeled Arrays and Datasets in Python. J. Open Res. 673

Page 34: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

34

Softw., 5, 1–6, doi:10.5334/jors.148. 674

http://openresearchsoftware.metajnl.com/articles/10.5334/jors.148/. 675

Huang, Q., C. Yang, K. Benedict, S. Chen, A. Rezgui, and J. Xie, 2013: Utilize cloud computing to 676

support dust storm forecasting. Int. J. Digit. Earth, 6, 338–355, doi:10.1080/17538947.2012.749949. 677

Hunter, J. D., 2007: Matplotlib: A 2D graphics environment. Comput. Sci. Eng., 9, 99–104, 678

doi:10.1109/MCSE.2007.55. 679

Huntington, J. L., and Coauthors, 2017: Climate Engine: Cloud Computing and Visualization of Climate 680

and Remote Sensing Data for Advanced Natural Resource Monitoring and Process Understanding. 681

Bull. Am. Meteorol. Soc., BAMS-D-15-00324.1, doi:10.1175/BAMS-D-15-00324.1. 682

http://journals.ametsoc.org/doi/10.1175/BAMS-D-15-00324.1. 683

Irving, D., 2016: A minimum standard for publishing computational results in the weather and climate 684

sciences. Bull. Am. Meteorol. Soc., 97, 1149–1158, doi:10.1175/BAMS-D-15-00010.1. 685

Jackson, K. R., L. Ramakrishnan, K. Muriki, S. Canon, S. Cholia, J. Shalf, H. J. Wasserman, and N. J. 686

Wright, 2010: Performance Analysis of High Performance Computing Applications on the Amazon 687

Web Services Cloud. 2010 IEEE Second Int. Conf. Cloud Comput. Technol. Sci., 159–168, 688

doi:10.1109/CloudCom.2010.69. http://ieeexplore.ieee.org/document/5708447/. 689

Jacobsen, D. M., and R. S. Canon, 2015: Contain This, Unleashing Docker for HPC. Cray User Gr. 2015, 690

14. https://www.nersc.gov/assets/Uploads/cug2015udi.pdf. 691

Jeong, J. I., and R. J. Park, 2018: Efficacy of dust aerosol forecasts for East Asia using the adjoint of 692

GEOS-Chem with ground-based observations. Environ. Pollut., 234, 885–893, 693

doi:10.1016/j.envpol.2017.12.025. 694

Page 35: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

35

Jing, Y., T. Wang, P. Zhang, L. Chen, N. Xu, and Y. Ma, 2018: Global atmospheric CO2 concentrations 695

simulated by GEOS-Chem: Comparison with GOSAT, carbon tracker and ground-based 696

measurements. Atmosphere (Basel)., 9, doi:10.3390/atmos9050175. 697

Jonas, E., and Coauthors, 2019: Cloud Programming Simplified: A Berkeley View on Serverless 698

Computing. 1–33. http://arxiv.org/abs/1902.03383. 699

Jung, K., Y. Cho, and Y. Tak, 2017: Performance evaluation of ROMS v3 . 6 on a commercial cloud 700

system. 701

Keller, C. A., M. S. Long, R. M. Yantosca, A. M. Da Silva, S. Pawson, and D. J. Jacob, 2014: HEMCO 702

v1.0: A versatile, ESMF-compliant component for calculating emissions in atmospheric models. 703

Geosci. Model Dev., 7, 1409–1417, doi:10.5194/gmd-7-1409-2014. 704

Kurtzer, G. M., V. Sochat, and M. W. Bauer, 2017: Singularity: Scientific containers for mobility of 705

compute. PLoS One, 12, e0177459, doi:10.1371/journal.pone.0177459. 706

http://www.ncbi.nlm.nih.gov/pubmed/28494014. 707

Langkamp, T., 2011: Influence of the compiler on multi-CPU performance of WRFv3. Geosci. Model 708

Dev. Discuss., 4, 547–573, doi:10.5194/gmdd-4-547-2011. http://www.geosci-model-dev-709

discuss.net/4/547/2011/. 710

Li, R., L. Liu, G. Yang, C. Zhang, and B. Wang, 2016: Bitwise identical compiling setup: Prospective for 711

reproducibility and reliability of Earth system modeling. Geosci. Model Dev., 9, 731–748, 712

doi:10.5194/gmd-9-731-2016. 713

Li, Z., C. Yang, Q. Huang, K. Liu, M. Sun, and J. Xia, 2017: Building Model as a Service to support 714

geosciences. Comput. Environ. Urban Syst., 61, 141–152, 715

Page 36: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

36

doi:10.1016/j.compenvurbsys.2014.06.004. 716

Lynnes, C., K. Baynes, and M. A. McInerney., 2017: Archive Management of NASA Earth Observation 717

Data to Support Cloud Analysis. 0–3. 718

https://ntrs.nasa.gov/archive/nasa/casi.ntrs.nasa.gov/20170011455.pdf. 719

Mehrotra, P., J. Djomehri, S. Heistand, R. Hood, H. Jin, A. Lazanoff, S. Saini, and R. Biswas, 2016: 720

Performance evaluation of Amazon Elastic Compute Cloud for NASA high-performance computing 721

applications. Concurr. Comput. , 28, 1041–1055, doi:10.1002/cpe.3029. 722

http://doi.wiley.com/10.1002/cpe.3029. 723

Met Office, 2016: Cartopy: a cartographic python library with a matplotlib interface. 724

http://scitools.org.uk/cartopy. 725

Microsoft, 2018a: Microsoft Azure for Research Online. Accessed December 20, 2018, 726

https://a4ronline.azurewebsites.net. 727

——, 2018b: Azure for AWS Professionals. Accessed December 20, 2018, 728

https://docs.microsoft.com/en-us/azure/architecture/aws-professional/. 729

——, 2018c: Azure AI for Earth grants. Accessed December 20, 2018, https://www.microsoft.com/en-730

us/ai-for-earth/grants. 731

Mohammadi, M., and T. Bazhirov, 2017: Comparative benchmarking of cloud computing vendors with 732

High Performance Linpack. 1–6, doi:10.1145/3195612.3195613. http://arxiv.org/abs/1702.02968. 733

Molthan, A., J. Case, and J. Venner, 2015: Clouds in the cloud: weather forecasts and applications within 734

cloud computing environments. Bull., 96, 1369–1379. 735

http://journals.ametsoc.org/doi/abs/10.1175/BAMS-D-14-00013.1. 736

Page 37: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

37

Montes, D., J. A. Añel, T. F. Pena, P. Uhe, and D. C. H. Wallom, 2017: Enabling BOINC in infrastructure 737

as a service cloud system. Geosci. Model Dev., 10, 811–826, doi:10.5194/gmd-10-811-2017. 738

NAS, 2018: Thriving on Our Changing Planet. National Academies Press, Washington, D.C., 739

https://www.nap.edu/catalog/24938. 740

NASA HECC, 2018: Enabling User-Defined Software Stacks with Charliecloud. Accessed December 20, 741

2018, https://www.nas.nasa.gov/hecc/support/kb/enabling-user-defined-software-stacks-with-742

charliecloud_558.html. 743

National Academies of Sciences, E., 2016: Future Directions for NSF Advanced Computing 744

Infrastructure to Support U.S. Science and Engineering in 2017-2020. 745

Netto, M. A. S., R. N. Calheiros, E. R. Rodrigues, R. L. F. Cunha, and R. Buyya, 2017: HPC Cloud for 746

Scientific and Business Applications: Taxonomy, Vision, and Research Challenges. 1, 1–29. 747

http://arxiv.org/abs/1710.08731. 748

NSF, 2018a: Critical Techniques, Technologies and Methodologies for Advancing Foundations and 749

Applications of Big Data Sciences and Engineering (BIGDATA). Accessed December 20, 2018, 750

https://www.nsf.gov/funding/pgm_summ.jsp?pims_id=504767. 751

——, 2018b: NSF and Internet2 to explore cloud computing to accelerate science frontiers. Accessed 752

December 20, 2018, https://nsf.gov/news/news_summ.jsp?cntn_id=297193. 753

Oesterle, F., S. Ostermann, R. Prodan, and G. J. Mayr, 2015: Experiences with distributed computing for 754

meteorological applications: Grid computing and cloud computing. Geosci. Model Dev., 8, 2067–755

2078, doi:10.5194/gmd-8-2067-2015. 756

de Oliveira, A. H. M., D. de Oliveira, and M. Mattoso, 2017: Clouds and Reproducibility: A Way to Go 757

Page 38: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

38

to Scientific Experiments? Cloud Computing, Springer International Publishing, 127–151 758

http://link.springer.com/10.1007/978-3-319-54645-2_5. 759

Perkel, J. M., 2018: Why Jupyter is data scientists’ computational notebook of choice. Nature, 563, 145–760

146, doi:10.1038/d41586-018-07196-1. http://www.nature.com/articles/d41586-018-07196-1. 761

Priedhorsky, R., T. C. Randles, and T. Randles, 2017: Charliecloud: Unprivileged containers for user-762

defined software stacks in HPC Charliecloud: Unprivileged containers for user-defined software 763

stacks in HPC. SC17 Int. Conf. High Perform. Comput. Networking, Storage Anal., 17, p1-10, 764

doi:10.1145/3126908.3126925. http://permalink.lanl.gov/object/tr?what=info:lanl-765

repo/lareport/LA-UR-16-22370. 766

Rocklin, M., 2015: Dask : Parallel Computation with Blocked algorithms and Task Scheduling. Proc. 767

14th Python Sci. Conf., 130–136. 768

——, 2018: HDF in the Cloud: challenges and solutions for scientific data. Accessed December 20, 2018, 769

http://matthewrocklin.com/blog/work/2018/02/06/hdf-in-the-cloud. 770

Roloff, E., M. Diener, A. Carissimi, and P. O. A. Navaux, 2012: High Performance Computing in the 771

cloud: Deployment, performance and cost efficiency. 4th IEEE International Conference on Cloud 772

Computing Technology and Science Proceedings, IEEE, 371–378 773

http://ieeexplore.ieee.org/document/6427549/. 774

——, ——, L. P. Gaspary, and P. O. A. Navaux, 2017: HPC Application Performance and Cost Efficiency 775

in the Cloud. Proc. - 2017 25th Euromicro Int. Conf. Parallel, Distrib. Network-Based Process. PDP 776

2017, 473–477, doi:10.1109/PDP.2017.59. 777

Salaria, S., K. Brown, H. Jitsumoto, and S. Matsuoka, 2017: Evaluation of HPC-Big Data Applications 778

Page 39: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

39

Using Cloud Platforms. Proc. 17th IEEE/ACM Int. Symp. Clust. Cloud Grid Comput., 1053–1061, 779

doi:10.1109/CCGRID.2017.143. https://doi.org/10.1109/CCGRID.2017.143. 780

Shaikh, R., 2018: tmux on AWS. Accessed December 20, 2018, 781

https://github.com/reshamas/fastai_deeplearn_part1/blob/master/tools/tmux.md. 782

Shen, H., 2014: Interactive notebooks: Sharing the code. Nature, 515, 151–152, doi:10.1038/515151a. 783

http://www.nature.com/doifinder/10.1038/515151a. 784

Siuta, D., G. West, H. Modzelewski, R. Schigas, and R. Stull, 2016: Viability of Cloud Computing for 785

Real-Time Numerical Weather Prediction. Weather Forecast., 31, 1985–1996, doi:10.1175/WAF-786

D-16-0075.1. http://journals.ametsoc.org/doi/10.1175/WAF-D-16-0075.1. 787

Strazdins, P. E., J. Cai, M. Atif, and J. Antony, 2012: Scientific application performance on HPC, private 788

and public cloud resources: A case study using climate, cardiac model codes and the NPB benchmark 789

suite. Proc. 2012 IEEE 26th Int. Parallel Distrib. Process. Symp. Work. IPDPSW 2012, 1416–1424, 790

doi:10.1109/IPDPSW.2012.186. 791

Thackston, R., and R. C. Fortenberry, 2015: The performance of low-cost commercial cloud computing 792

as an alternative in computational chemistry. J. Comput. Chem., 36, 926–933, 793

doi:10.1002/jcc.23882. 794

Tian, Y., Y. Sun, C. Liu, P. Xie, K. Chan, J. Xu, W. Wang, and J. Liu, 2018: Characterization of urban 795

CO2 column abundance with a portable low resolution spectrometer (PLRS): Comparisons with 796

GOSAT and GEOS-Chem model data. Sci. Total Environ., 612, 1593–1609. 797

Tsaftaris, S. A., 2014: A scientist’s guide to cloud computing. Comput. Sci. Eng., 16, 70–76, 798

doi:10.1109/MCSE.2014.12. 799

Page 40: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

40

Vance, T. C., N. Merati, C. P. Yang, and M. Yuan, 2016: Cloud Computing in Ocean and Atmospheric 800

Sciences. 1-415 pp. 801

VanderPlas, J., 2016: Python Data Science Handbook. O’Reilly Media, Inc. 802

Walker, E., 2008: Benchmarking Amazon EC2 for high-performance scientific computing. login Mag. 803

USENIX SAGE, 33, 18–23. 804

Withana, E. C., B. Plale, and C. Mattocks, 2011: Towards Enabling Mid-Scale Geoscience Experiments 805

Through Microsoft Trident and Windows Azure. 806

Yang, C., Q. Huang, Z. Li, K. Liu, and F. Hu, 2017: Big Data and cloud computing: innovation 807

opportunities and challenges. Int. J. Digit. Earth, 10, 13–53, doi:10.1080/17538947.2016.1239771. 808

Yelick, K., and Coauthors, 2011: The Magellan Report on Cloud Computing for Science. 170. 809

Younge, A. J., K. Pedretti, R. E. Grant, and R. Brightwell, 2017: A Tale of Two Systems: Using 810

Containers to Deploy HPC Applications on Supercomputers and Clouds. Proc. Int. Conf. Cloud 811

Comput. Technol. Sci. CloudCom, 2017–Decem, 74–81, doi:10.1109/CloudCom.2017.40. 812

Yu, K., C. A. Keller, D. J. Jacob, A. M. Molod, S. D. Eastham, and M. S. Long, 2018: Errors and 813

improvements in the use of archived meteorological data for chemical transport modeling: An 814

analysis using GEOS-Chem v11-01 driven by GEOS-5 meteorology. Geosci. Model Dev., 11, 305–815

319, doi:10.5194/gmd-11-305-2018. 816

Zhai, Y., M. Liu, J. Zhai, X. Ma, and W. Chen, 2011: Cloud Versus In-house Cluster: Evaluating Amazon 817

Cluster Compute Instances for Running MPI Applications. State Pract. Reports - SC ’11, 1, 818

doi:10.1145/2063348.2063363. 819

Zhang, J., X. Lu, and D. K. Panda, 2017: Is Singularity-based Container Technology Ready for Running 820

Page 41: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

41

MPI Applications on HPC Clouds? Proc. the10th Int. Conf. Util. Cloud Comput. - UCC ’17, 151–821

160, doi:10.1145/3147213.3147231. http://dl.acm.org/citation.cfm?doid=3147213.3147231. 822

Zhu, L., M. Val Martin, A. Hecobian, L. V. Gatti, R. Kahn, and E. V. Fischer, 2018: Development and 823

implementation of a new biomass burning emissions injection height scheme for the GEOSChem 824

model. Geosci. Model Dev. Discuss., 1–30, doi:10.5194/gmd-2018-93. https://www.geosci-model-825

dev-discuss.net/gmd-2018-93/. 826

Zhuang, J., 2018: JiaweiZhuang/xESMF: v0.1.1. doi:10.5281/ZENODO.1134366. 827

https://zenodo.org/record/1134366. 828

829

830

Page 42: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

42

Table 1. Hardware specification and cost as of December 2018a. 831

Instance/node typeb Processor informationc Hourly costd

On-demand Spote

AWS

EC2 c4.8xlarge Intel Xeon CPU E5-2666v3, 2.9GHz, 32 vCPUs $1.59 $0.57

EC2 c5.9xlarge Intel Xeon Platinum 8124M, 3.0GHz, 32 vCPUs $1.53 $0.58

EC2 c4.4xlarge Intel Xeon CPU E5-2666v3, 2.9GHz, 16 vCPUs $0.80 $0.25

EC2 c5.4xlarge Intel Xeon Platinum 8124M, 3.0GHz, 16 vCPUs $0.68 $0.27

NASA HECC

Pleiades Sandy Bridge Intel Xeon E5-2680v2, 2.8 GHz, 16 CPU cores $0.29 --

Pleiades Haswell Intel Xeon E5-2680v3, 2.5 GHz, 24 CPU cores $0.53 --

a. AWS costs are from https://aws.amazon.com/ec2/pricing/. NASA High-End Computing Capability 832

(HECC) costs are from https://www.hec.nasa.gov/user/policies/sbus.html. Costs are in USD. The price 833

can vary between countries and regions. Shown here are for the US East (N. Virginia) region. Same for 834

Tables 2 and 3. 835

b. The naming of an EC2 instance follows “family, generation, size”. For example, “c4” refers to the 836

“compute-optimized” family, generation four; “8xlarge” indicates the instance size, which has twice as 837

many cores and memory as “4xlarge”. 838

c. EC2 instances are virtual machines and the processors are described by “virtual CPUs” (“vCPUs”). A 839

vCPU is a hyperthread and corresponds to half of a physical core (Amazon 2018l). 840

Page 43: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

43

d. EC2 actually uses per-second billing, so that short computations are very cheap (Amazon 2017b). 841

e. The spot price can fluctuate with time. We show here the price when the simulations where conducted. 842

Price fluctuation is typically within 20% in a month. 843

f. The AWS cloud and container environments used GNU Fortran compiler 7.3.0; the NASA Pleiades 844

native environment used GNU Fortran compiler 6.2.0 as it is the latest available version. 845

846

Table 2. Description and cost of major AWS services as of December 2018a 847

Service Description and Purpose Cost

EC2 Server used for computing on AWS. Users

request an EC2 instance with a chosen number

of virtual CPUs (vCPUs) for their computing

need.

~$0.05 per vCPU hour, depending on

instance type. 60%~70% cheaper with

spot pricing. See Table 1 for examples.

EBS Disk storage for temporary data. It hosts input

and output data during model simulations.

$0.1 per GB per month, for the

standard solid-state drive (SSD).

Cheaper options with different I/O

characteristics are available.

S3 Major storage service. User transfer data from

EC2 to S3 for persist storage, and later retrieve

data from S3 to EC2 for continued work.

$0.023 per GB per month. Cheaper

options for infrequent access patterns

are available.

Data egress Downloading data to local machines. $0.09 per GB

a. From https://aws.amazon.com/pricing/services/ 848

Page 44: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

44

849

Table 3. AWS cost of a 1-year 2ox2.5o global GEOS-Chem simulation with tropospheric and stratospheric 850

chemistry 851

Service Amount of resources needed Total cost (USD)

EC2 330 hours on c5.4xlarge $224 with on-demand pricing,

$90 with spot pricing

EBS 300 GB disk to host input and output

data files during simulation

$14 with standard solid-state drive

(SSD)

S3 or download 150 GB output data files $3.50 per month on S3,

or $13.50 to download

852

Page 45: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

45

. 853

Figure 1. Examples of GEOS-Chem input data hosted on the AWS cloud. Top: Global 25 km-resolution 854

air temperature data at 4 km altitude on July 1, 2016 (0 UTC), from the GEOS-FP meteorological product. 855

Middle: Global 50 km-resolution anthropogenic nitrogen oxides (NOx) emission data averaged over Jan 856

2014,c from the Community Emissions Data System (CEDS, Hoesly et al., 2018). Bottom: Global 2° × 857

2.5° resolution ozone concentration data at 4 km altitude on July 1, 2016 (0 UTC), from model initial 858

condition files. 859

Page 46: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

46

860

Figure 2. A complete research workflow for using the GEOS-Chem model on the AWS cloud. From the 861

AWS console accessed through web browsers, users select the GEOS-Chem Amazon Machine Image 862

(AMI) which includes the fully configured model, and launch a Linux server (EC2 instance) from that 863

AMI. Users then login to that EC2 instance through their local terminal, customize the model 864

configuration as desired, and execute the model simulation. Model output can be analyzed on the cloud 865

using Python in Jupyter notebooks, or can be downloaded to users’ local computers. Local I/O is managed 866

Page 47: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

47

through the Elastic Block Storage (EBS) volume on the EC2 instance. The complete meteorological and 867

emission data needed to drive GEOS-Chem can be retrieved from a public bucket residing permanently 868

in the Simple Storage Service (S3) of the AWS cloud. Users can also create their own S3 buckets to 869

archive model output data and customized run directories. After finishing the computation, users 870

terminate the EC2 instance; data in the EBS volume will then be deleted but data in the S3 buckets will 871

persist. 872

873

874

Figure 3. Wall-clock time required for 15 GEOS-Chem users with no prior cloud experience to complete 875

a demo GEOS-Chem simulation on the AWS cloud and visualize the output. The actual simulation 876

runtime is not included and is a few minutes for this demo case. 877

Page 48: Enabling immediate access to Earth science models through ...acmg.seas.harvard.edu/publications/2019/zhuang2019.pdf · 47 and user-accessible on the AWS cloud. 48 49 A number of Earth

48

878

879

Figure 4. Performance and cost comparisons between the AWS EC2 instances and the NASA Pleiades 880

supercomputing cluster for a 7-day GEOS-Chem simulation with tropospheric and stratospheric 881

chemistry at global 4° × 5° resolution, running on a single compute node. See Table 1 for description of 882

the different hardware types. The SBU model is used by NASA to estimate the true cost of Pleiades. Cost 883

is in USD. 884