Upload
lyduong
View
217
Download
2
Embed Size (px)
Citation preview
National Energy Research Scientific Computing Center
2015 Annual Report
NERSC 2015 Annual Report
Ernest Orlando Lawrence Berkeley National Laboratory
1 Cyclotron Road, Berkeley, CA 94720-8148
This work was supported by the Director, Office of Science, Office of Advanced Scientific Computing Research of the
U.S. Department of Energy under Contract No. DE-AC02-05CH11231
Image Credits:
FRONT COVER: Roy Kaltschmidt, Lawrence Berkeley National Laboratory
BACK COVER (TOP TO BOTTOM): Cornford et al., The Cryosphere, 2015 (story p. 35); Scott French, Lawrence Berkeley National
Laboratory (story p. 31); Journal of Materials Research (story p. 34); Ben Bowen, Lawrence Berkeley National Laboratory
(story p. 44); Nicholas P. Brawand, University of Chicago (story p. 32)
THIS PAGE: © Emily Hagopian/Courtesy Perkins + Will
National Energy Research Scientific Computing Center
2015 Annual Report
NERSC ANNUAL REPORT 20152
Table of Contents
Director’s Note: A Year of Transitions 4
NERSC User/Usage Demographics 6
Innovations 12PAVING THE WAY TO EXASCALE 13
Extreme-Scale and Data-Intensive Platforms 14New Benchmarks for Data-Intensive Computing 15NERSC Introduces Power Measurement Library 16
DATA-INTENSIVE SCIENCE 17 Shifter: Enabling Docker-Like Containers for HPC 17New 400Gbps Production Network Speeds Science 19Cori Gets a Real-Time Queue 20iPython/JupyterHub Portal Debuts 21Cray-AMPLab-NERSC Partnership Targets Data Analytics 22Deep Learning Comes to HPC 23
OPERATIONAL EXCELLENCE 25Thruk: A New Monitoring Interface 25Environmental Data Collection Projects 27Centralizing Sensor Data Collection 27Compact Cori: An Educational Tool for HPC 28
Science Highlights 30Simulations Link Mantle Plumes with Volcanic Hotspots (BES) 31What the Blank Makes Quantum Dots Blink? (BES) 32Stabilizing a Tokamak Plasma (FES) 33 Experiments + Simulations = Better Nuclear Power Research (BES) 34BISICLES Ice Sheet Model Quantifies Climate Change (ASCR) 35Tall Clouds from Tiny Raindrops Grow (BER) 36Supernova Hunting with Supercomputers (HEP) 37Petascale Pattern Recognition Enhances Climate Science (BER) 38Celeste: A New Model for Cataloging the Universe (ASCR) 39Supercomputers Speed Search for New Subatomic Particles (HEP) 40What Causes Electron Heat Loss in Fusion Plasma? (FES) 41Documenting CO2’s Increasing Greenhouse Effect at Earth’s Surface (BER) 42What Ignites a Neutron Star? (NP) 43‘Data Deluge’ Pushes MSI Analysis to New Heights (ASCR) 44Could Material Defects Actually Improve Solar Cells? (BES) 45
Table of Contents 3
User Support and Outreach 46NESAP Begins to Bloom 47NERSC Sponsors Trainings, Workshops and Hackathons 48IXPUG Annual Meeting a Hit 49Application Portability Efforts Progress 50
DATA AND ANALYTICS SUPPORT 51OpenMSI + iPython/Jupyter = Interactive Computing 51Supporting LUX with SciDB 53Burst Buffer Early User Program Moves Forward 53
OPERATIONAL EXCELLENCE 56HPSS Disk Cache Gets an Upgrade 56Streamlining Data and Analytics Software Support 56The Big Move: Minimizing Disruption, Improving Operations 57
Center News 58NERSC’S New Home: Shyh Wang Hall 59Changing of the Guard: Hopper and Carver Retire, Cori Moves In 60APEX: A Joint Effort for Designing Next-Generation HPC Systems 61Industry Partnership Program Expands 62‘Superfacilities’ Concept Showcases Interfacility Collaborations 64NERSC and the 2015 Nobel Prize in Physics 64OpenMSI Wins 2015 R&D 100 Award 65NERSC Staff Honored with Director’s Awards for Exceptional Achievement 664th Annual HPC Achievement Awards Honor Innovative Research 66Organizational Changes Reflect Evolving HPC Data Environment 68Personnel Transitions 68
Director’s Reserve Allocations 70
Publications 73
Appendix A: NERSC Users Group Executive Committee 74
Appendix B: Office of Advanced Scientific Computing Research 75
Appendix C: Acronyms and Abbreviations 76
Director’s Note: A Year of Transitions
NERSC ANNUAL REPORT 20154
2015 was a year of transition and change for NERSC on many levels. The most pronounced change was the center’s move from the Oakland Scientific Facility in downtown Oakland to the new Shyh Wang Hall (also known internally as the Computational Research and Theory facility, or CRT) on the main Berkeley Lab campus. A significant amount of staff time was spent planning and implementing the move of computational systems, file systems and auxiliary systems while minimizing the impact on NERSC users and their ongoing projects.
Indeed, one of NERSC’s greatest challenges in 2015 was running multiple systems across two different facilities with a constant number of staff. Despite having to physically move NERSC’s largest system, Edison, and all file systems in November 2015, NERSC was able to deliver more than the committed 3 billion hours to its users—the same amount committed to in 2014 without the move. This was accomplished by keeping the Hopper system in production until Cori Phase 1 was stable.
Wang Hall itself brings many new changes. The building was designed to be extraordinarily energy efficient and expandable to meet the needs of next-generation HPC systems. It features 12.5 megawatts of power, upgradable to over 40 megawatts, a PUE (power usage effectiveness) of less than 1.1, ambient “free” cooling and the ability to reclaim heat from the computers to heat the building. The new machine room currently measures 20,000 square feet and features a novel seismic isolation floor to protect NERSC’s many valuable resources.
Another big change in 2015 was the retirement of two of NERSC’s legacy systems, Carver and Hopper. This was done in conjunction with the deployment of the first phase of Cori and the Edison move. As of early 2016, the transfer of systems into the new building was substantially complete and was accomplished with a minimum of downtime. The process was facilitated by a first-of-a-kind 400 Gbs link, provided by ESnet between Wang Hall and the Oakland Scientific Facility, that enabled moves of data to be effectively transparent to users.
Another major undertaking for NERSC in 2015 was the NERSC Exascale Science Applications Program (NESAP). We kicked off the project with 20 applications teams, plus an additional 20 teams participating in training and general optimization efforts. Through NESAP, Cray and Intel staff are able to work directly with code teams to optimize applications and learn about using Cray and Intel performance tools, which was unique because in most cases no such relationship previously existed.
2015 was also the year that NERSC decided there was enough momentum around HPC and data-intensive science that we should form a department focused on data. The new department has four groups, two of them new: a Data Science Engagement group that engages with users of experimental and observational facilities, and an Infrastructure Services group that focuses on infrastructure to support NERSC and our data initiative. The existing Storage Systems Group and Data and Analytics Group complete the new department.
Director's Note 5
In parallel with these changes, NERSC expanded its workforce in 2015, hiring 18 additional staff members. Our new staff have brought new perspectives to NERSC, and they are playing a critical role in our strategic initiatives. New positions became available due to a large number of retirements in 2014 and 2015, hiring of postdocs into NESAP and normal turnover and churn from competition with Bay Area technology companies.
Also in 2015, NERSC reinvigorated its Private Sector Partnership program, bringing on nine new industry partners in energy-specific fields such as semiconductor manufacturing and geothermal modeling. NERSC has been working with industry partners through the DOE’s Small Business Innovation Research (SBIR) grant program for many years. Since 2011, industry researchers have been awarded 40 million hours at NERSC via the SBIR program, and NERSC currently has about 130 industry users from 50 companies.
All of these changes allowed us to continue to address our primary mission: to accelerate scientific discovery for the DOE Office of Science through HPC and data analysis. In 2015, NERSC users reported more than 2,000 refereed papers based on work performed at the center and NERSC staff contributed some 120 papers to scientific journals and conferences, showcasing the center’s increasing involvement in technology and software development designed to enhance the utilization of HPC across a broad range of science and data-management applications.
Looking back on an exciting year, we are especially proud of what we were able to accomplish given the many transitions we experienced. NERSC is regarded as a well-run scientific computing facility providing some of the largest computing and storage systems available anywhere, but what really distinguishes the center is its success in creating an environment that makes these resources effective for scientific research and productivity. We couldn’t achieve this without our dedicated staff, our great user community and the continued support of our DOE Office of Science sponsors.
Sudip Dosanjh NERSC Division Director
NERSC ANNUAL REPORT 20156
NERSC User/Usage DemographicsIn 2015, the NERSC Program supported 6,000 active users from universities, national laboratories and industry who used approximately 3 billion supercomputing hours. Our users came from across the United States, with 48 states represented, as well as from around the globe, with 47 countries represented.
NERSC User/Usage Demographics 7
2015 NERSC Usage by Institution Type (MPP HOURS IN MILLIONS) 2015 NERSC Usage by Institution Type
(MPP Hours in Millions)
2015 NERSC Usage by DOE Program Office(MPP Hours in Millions)
2015 NERSC Usage By Discipline (MPP Hours in Millions)
Fusion Energy585
Materials Science616
Climate333
Lattice QCD341
Chemistry311
Geosciences132
Nuclear Physics208
Astrophysics158
Computer Science
88
Life Sciences
129
High Energy Physics
83Mathematics
55
Combustion45
Accelerator Physics
56
EnvironmentalScience
24
Engineering24
Nuclear Energy
4
Other<1
BES 1,121
BER 551
FES 528
HEP 441
NP 282
ASCR 267
Materials Science 616
Fusion Energy 585
Lattice QCD 341
Climate 333
Chemistry 311
Nuclear Physics 208
AstroPhysics 158
Geosciences 132
Life Sciences 129
Computer Science 88
High Energy Physics 83
Accelerator Physics 56
Mathematics 55
Combustion 45
Environmental Science 24
Engineering 24
Nuclear Energy 4
Other <1
Universities 1,570DOE Labs 1,478Other Government Labs 115Industry 23Nonprofits 4
0
500
1000
1500
2000
2500
3000
3500
2015 NERSC Usage by DOE Program Office(MPP HOURS IN MILLIONS)
2015 NERSC Usage by Institution Type (MPP Hours in Millions)
2015 NERSC Usage by DOE Program Office(MPP Hours in Millions)
2015 NERSC Usage By Discipline (MPP Hours in Millions)
Fusion Energy585
Materials Science616
Climate333
Lattice QCD341
Chemistry311
Geosciences132
Nuclear Physics208
Astrophysics158
Computer Science
88
Life Sciences
129
High Energy Physics
83Mathematics
55
Combustion45
Accelerator Physics
56
EnvironmentalScience
24
Engineering24
Nuclear Energy
4
Other<1
BES 1,121
BER 551
FES 528
HEP 441
NP 282
ASCR 267
Materials Science 616
Fusion Energy 585
Lattice QCD 341
Climate 333
Chemistry 311
Nuclear Physics 208
AstroPhysics 158
Geosciences 132
Life Sciences 129
Computer Science 88
High Energy Physics 83
Accelerator Physics 56
Mathematics 55
Combustion 45
Environmental Science 24
Engineering 24
Nuclear Energy 4
Other <1
Universities 1,570DOE Labs 1,478Other Government Labs 115Industry 23Nonprofits 4
0
500
1000
1500
2000
2500
3000
3500
Universities 1,570
DOE Labs 1,478
Other Government 115 Labs
Industry 23
Nonprofits 4
BES 1,121
BER 551
FES 528
HEP 441
NP 282
ASCR 267
NERSC ANNUAL REPORT 20158
2015 DOE & Other Lab Usage at NERSC (MPP HOURS IN MILLIONS)
2015 Academic Usage at NERSC (MPP HOURS IN MILLIONS)
Materials ScienceMaterials Science
NASA GSFC 1
PPPL 214Sandia Lab CA 17
Argonne Lab 126
Oak Ridge 136
Sandia Lab NM 14
Livermore Lab 33 NCAR 91
Los Alamos Lab 87
Ames Lab 22
Fermilab 19 Brookhaven Lab 35
IGES-COLA 2
US EPA AMAD 5
SLAC 27
Naval Research Lab 8
JGI 2
Natl Energy Tech Lab 23
Other (11) <1
Berkeley Lab 619PPPL 214Oak Ridge 136Argonne Lab 126NCAR 91PNNL 89Los Alamos Lab 87Brookhaven Lab 35Livermore Lab 33SLAC 27Natl Energy Tech Lab 23Ames Lab 22Fermilab 19Sandia Lab CA 17Sandia Lab NM 14Jefferson Lab 10Naval Research Lab 8Smithsonian Astro Observ 6US EPA AMAD 5NERSC 4NREL 3IGES - COLA 2JGI 2NOAA 2DOE Germantown 1NASA GSFC 1Other (11) <1
Berkeley Lab 619NREL 3
PNNL 89
Jefferson Lab 10
NERSC 4 NOAA 2DOE Germantown 1
Smithsonian Astro Observ 6
MIT 150William & Mary 112U. Arizona 99UC San Diego 77UC Berkeley 75U. Illinois U-C 72U. Washington 56UC Irvine 54UCLA 53U. Colorado Boulder 50Princeton Univ 40Caltech 40U. Chicago 37U. Kentucky 36Columbia Univ 32U. Wisc. Madison 32Temple University 28Tulane Univ 28Northwestern Univ 25
U. Penn 25Iowa State 23U. Maryland 22Colorado School Mines 20U. Michigan 19U. Texas Austin 18U. Rochester 17U. Oklahoma 17U. Central Florida 16Cornell Univ 15North Carolina State 13U. South Carolina 13Vanderbilt Univ 13U. Missouri KC 11Harvard Univ 10U. Florida 8Penn State 8Stanford Univ 8UMass Amherst 8
Louisiana State 7UC Santa Barbara 7U. Tennessee 7U. New Hampshire 7George Wash Univ 7U. Delaware 6Rice Univ 6UNC Chapel Hill 6Dartmouth College 5Clemson Univ 5U. New Mexico 5Auburn Univ 5U. Minnesota 5UC Riverside 4Johns Hopkins Univ 4U. Illinois Chicago 4Georgetown Univ 4Oregon State 4Indiana St U 4
UC Santa Cruz 4U. Tulsa 4Virginia Comm Univ 3UC Davis 3West Virginia Univ 3U.Toledo 3Kansas State 3Yale Univ 3Georgia State 3U South Dakota 3Haverford College 2Univ Col London UK 2N. Illinois University 2U. Utah 2U. Houston 2Marquette Univ 2Utah State 2Rensselaer 2Georgia Tech 2
Northeastern 2U. Colorado Denver 2SUNY Stony Brook 2U Texas El Paso 2Purdue Univ 2Mississippi State 2Florida State 2Michigan State 2LA Tech U 2Boston Univ 2Missouri State 2New Jersey IT 2N Dakota State U 1Arkansas State 1U. Notre Dame 1U. Georgia 1U. Guelph Canada 1Other (45) <1
0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150
150 140 130 110 100 80 70 50 40 20 10120 90 60 30 0
Materials ScienceMaterials Science
MIT 150William & Mary 112U. Arizona 99UC San Diego 77UC Berkeley 75U. Illinois U-C 72U. Washington 56UC Irvine 54UCLA 53U. Colorado Boulder 50Princeton Univ 40Caltech 40U. Chicago 37U. Kentucky 36Columbia Univ 32U. Wisc. Madison 32
Temple University 28Tulane Univ 28Northwestern Univ 25U. Penn 25Iowa State 23U. Maryland 22Colorado School Mines 20U. Michigan 19U. Texas Austin 18U. Rochester 17U. Oklahoma 17U. Central Florida 16Cornell Univ 15North Carolina State 13U. South Carolina 13Vanderbilt Univ 13
U. Missouri KC 11Harvard Univ 10U. Florida 8Penn State 8Stanford Univ 8UMass Amherst 8Louisiana State 7UC Santa Barbara 7U. Tennessee 7U. New Hampshire 7George Wash Univ 7U. Delaware 6Rice Univ 6UNC Chapel Hill 6Dartmouth College 5Clemson Univ 5
NERSC User/Usage Demographics 9
2015 NERSC Users by State
California 2265Illinois 368 Tennessee 265 New York 258 Washington 231 Massachusetts 185 Texas 169 Colorado 168 Pennsylvania 151 New Mexico 144 New Jersey 123 Maryland 103 Michigan 89 Connecticut 85 North Carolina 83 Indiana 78 Florida 72 Georgia 69 Wisconsin 69 Virginia 66 Iowa 64 Ohio 43 Minnesota 41 Utah 39 District of Columbia 38 Arizona 30 Louisiana 30 South Carolina 29 Alabama 28 Oregon 28 Delaware 26 Missouri 21 South Dakota 18 Kansas 16 Kentucky 15 Oklahoma 12 New Hampshire 11 North Dakota 10 Rhode Island 10 Arkansas 8 West Virginia 8 Mississippi 6 Hawaii 5 Wyoming 5 Nevada 4 Montana 3 Alaska 2 Nebraska 2 Idaho 1 Maine 0 Vermont 0
MIT 150William & Mary 112U. Arizona 99UC San Diego 77UC Berkeley 75U. Illinois U-C 72U. Washington 56UC Irvine 54UCLA 53U. Colorado Boulder 50Princeton Univ 40Caltech 40U. Chicago 37U. Kentucky 36Columbia Univ 32U. Wisc. Madison 32Temple University 28Tulane Univ 28Northwestern Univ 25
U. Penn 25Iowa State 23U. Maryland 22Colorado School Mines 20U. Michigan 19U. Texas Austin 18U. Rochester 17U. Oklahoma 17U. Central Florida 16Cornell Univ 15North Carolina State 13U. South Carolina 13Vanderbilt Univ 13U. Missouri KC 11Harvard Univ 10U. Florida 8Penn State 8Stanford Univ 8UMass Amherst 8
Louisiana State 7UC Santa Barbara 7U. Tennessee 7U. New Hampshire 7George Wash Univ 7U. Delaware 6Rice Univ 6UNC Chapel Hill 6Dartmouth College 5Clemson Univ 5U. New Mexico 5Auburn Univ 5U. Minnesota 5UC Riverside 4Johns Hopkins Univ 4U. Illinois Chicago 4Georgetown Univ 4Oregon State 4Indiana St U 4
UC Santa Cruz 4U. Tulsa 4Virginia Comm Univ 3UC Davis 3West Virginia Univ 3U.Toledo 3Kansas State 3Yale Univ 3Georgia State 3U South Dakota 3Haverford College 2Univ Col London UK 2N. Illinois University 2U. Utah 2U. Houston 2Marquette Univ 2Utah State 2Rensselaer 2Georgia Tech 2
Northeastern 2U. Colorado Denver 2SUNY Stony Brook 2U Texas El Paso 2Purdue Univ 2Mississippi State 2Florida State 2Michigan State 2LA Tech U 2Boston Univ 2Missouri State 2New Jersey IT 2N Dakota State U 1Arkansas State 1U. Notre Dame 1U. Georgia 1U. Guelph Canada 1Other (45) <1
0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150
150 140 130 110 100 80 70 50 40 20 10120 90 60 30 0
Materials ScienceMaterials Science
U. New Mexico 5Auburn Univ 5U. Minnesota 5UC Riverside 4Johns Hopkins Univ 4U. Illinois Chicago 4Georgetown Univ 4Oregon State 4Indiana St U. 4UC Santa Cruz 4U. Tulsa 4Virginia Comm Univ 3UC Davis 3West Virginia Univ 3U. Toledo 3Kansas State 3
Yale Univ 3Georgia State 3U. South Dakota 3Haverford College 2Univ Col London UK 2N. Illinois University 2U. Utah 2U. Houston 2Marquette Univ 2Utah State 2Rensselaer 2Georgia Tech 2Northeastern 2U. Colorado Denver 2SUNY Stony Brook 2U. Texas El Paso 2
Purdue Univ 2Mississippi State 2Florida State 2Michigan State 2LA Tech U. 2Boston Univ 2Missouri State 2New Jersey IT 2N Dakota State U. 1Arkansas State 1U. Notre Dame 1U. Georgia 1U. Guelph Canada 1Other (45) <1
California 2,265Illinois 368 Tennessee 265 New York 258 Washington 231 Massachusetts 185 Texas 169 Colorado 168 Pennsylvania 151 New Mexico 144
New Jersey 123 Maryland 103 Michigan 89 Connecticut 85 North Carolina 83 Indiana 78 Florida 72 Georgia 69 Wisconsin 69 Virginia 66
Iowa 64 Ohio 43 Minnesota 41 Utah 39 District of Columbia 38 Arizona 30 Louisiana 30 South Carolina 29 Alabama 28 Oregon 28
Delaware 26 Missouri 21 South Dakota 18 Kansas 16 Kentucky 15 Oklahoma 12 New Hampshire 11 North Dakota 10 Rhode Island 10 Arkansas 8
West Virginia 8 Mississippi 6 Hawaii 5 Wyoming 5 Nevada 4 Montana 3 Alaska 2 Nebraska 2 Idaho 1 Maine 0 Vermont 0
501 and over
101 – 500
26 – 100
1 – 25
0
NERSC ANNUAL REPORT 201510
2015 NERSC Usage by Institution Type (MPP Hours in Millions)
2015 NERSC Usage by DOE Program Office(MPP Hours in Millions)
2015 NERSC Usage By Discipline (MPP Hours in Millions)
Fusion Energy585
Materials Science616
Climate333
Lattice QCD341
Chemistry311
Geosciences132
Nuclear Physics208
Astrophysics158
Computer Science
88
Life Sciences
129
High Energy Physics
83Mathematics
55
Combustion45
Accelerator Physics
56
EnvironmentalScience
24
Engineering24
Nuclear Energy
4
Other<1
BES 1,121
BER 551
FES 528
HEP 441
NP 282
ASCR 267
Materials Science 616
Fusion Energy 585
Lattice QCD 341
Climate 333
Chemistry 311
Nuclear Physics 208
AstroPhysics 158
Geosciences 132
Life Sciences 129
Computer Science 88
High Energy Physics 83
Accelerator Physics 56
Mathematics 55
Combustion 45
Environmental Science 24
Engineering 24
Nuclear Energy 4
Other <1
Universities 1,570DOE Labs 1,478Other Government Labs 115Industry 23Nonprofits 4
0
500
1000
1500
2000
2500
3000
35002015 NERSC Users by Country
2015 NERSC Usage By Discipline (MPP HOURS IN MILLIONS)
North America: 5,637United States of America 5,594Canada 38Puerto Rico 5
South America: 29Chile 15Brazil 11Argentina 1Colombia 1Ecuador 1
Africa: 5South Africa 5
Middle East/ Asia Pacific: 287People’s Republic of China 142Republic of Korea (South) 30Japan 24India 24Israel 19Australia 18Taiwan, Province of China 17Singapore 6Saudi Arabia 4Malaysia 1New Zealand 1Pakistan 1
North America 5,637
South America 29
NERSC User/Usage Demographics 11
2015 NERSC Usage by Institution Type (MPP Hours in Millions)
2015 NERSC Usage by DOE Program Office(MPP Hours in Millions)
2015 NERSC Usage By Discipline (MPP Hours in Millions)
Fusion Energy585
Materials Science616
Climate333
Lattice QCD341
Chemistry311
Geosciences132
Nuclear Physics208
Astrophysics158
Computer Science
88
Life Sciences
129
High Energy Physics
83Mathematics
55
Combustion45
Accelerator Physics
56
EnvironmentalScience
24
Engineering24
Nuclear Energy
4
Other<1
BES 1,121
BER 551
FES 528
HEP 441
NP 282
ASCR 267
Materials Science 616
Fusion Energy 585
Lattice QCD 341
Climate 333
Chemistry 311
Nuclear Physics 208
AstroPhysics 158
Geosciences 132
Life Sciences 129
Computer Science 88
High Energy Physics 83
Accelerator Physics 56
Mathematics 55
Combustion 45
Environmental Science 24
Engineering 24
Nuclear Energy 4
Other <1
Universities 1,570DOE Labs 1,478Other Government Labs 115Industry 23Nonprofits 4
0
500
1000
1500
2000
2500
3000
3500
Europe: 607United Kingdom 137 Germany 84France 79Italy 65Switzerland 46Spain 36 Denmark 34Poland 23Czech Republic 22Russian Federation 14Norway 12Sweden 11Netherlands 7Finland 6Greece 6Portugal 6Belgium 4Cyprus 4Ireland 3Turkey 3Slovenia 2Austria 1Serbia 1Ukraine 1
Middle East/ Asia Pacific
287
Europe 607
Africa5
NERSC ANNUAL REPORT 201512
InnovationsNERSC sets the bar high when it comes to providing its users with increased performance and additional services and operational efficiencies. The following pages highlight NERSC innovations from 2015 in our three strategic focus areas: exascale computing, data-intensive science and operational excellence.
Innovations 13
Paving the Way to Exascale Scientific discovery and translational R&D are increasingly data driven as faster, more powerful experimental facilities generate data at unprecedented speed and resolution. In turn, increasingly powerful processing tools are needed to search, transform, analyze and interpret this data. And it needs to be done quickly so that scientists can use the results to guide future experiments and make the most efficient use of facilities. Specific examples of these impending data challenges include:
• Genomics and metagenomics, where data volume will increase to ~100PB by 2021
• High energy physics, such as the Large Hadron Collider, with 50-100 exabytes of data starting in 2025
• Light sources, such as Berkeley Lab’s Advanced Light Source, which will each generate about 1.5PB/month by 2020
• Climate, where annual data from simulations is expected to reach around 100PB by 2020
Toward this end, a principal goal of NERSC’s exascale strategic initiative is to meet the ever-growing computing and data needs of the DOE Office of Science community by providing exascale computing and storage systems to the broad NERSC user community. Over the past year, NERSC made significant progress in developing the tools and techniques that will be used to assess the usability of exascale architectures in the context of the evolving NERSC workload.
Powerful light sources
(such as Berkeley Lab’s
Advanced Light Source,
shown here), telescopes and
accelerators are giving scientists
greater insight into many
phenomena, generating data
at unprecedented speed and
resolution. But getting the
science out of the data requires
powerful processing—such as
exascale—to search, transform,
analyze and interpret the data.
Image: Lawrence Berkeley
National Laboratory
NERSC ANNUAL REPORT 201514
Extreme-Scale and Data-Intensive PlatformsAn increasing number of scientific discoveries are being made based on the analysis of large volumes of data coming from observational science and large-scale simulation. These analyses are beginning to require resources only available on large-scale computing platforms, and data-intensive workloads are expected to become a significant component of the Office of Science’s workload in the coming years.
For example, scientists are using telescopes on Earth—such as the Dark Energy Spectroscopic Instrument (DESI) managed by LBNL—and mounted on satellites to try to “see” the unseen forces, such as dark matter, that can explain the birth, evolution and fate of our universe. Detailed, large-scale simulations of how structures form in the universe will play a key role in advancing our understanding of cosmology and the origins of the universe.
To ensure that future systems will be able to meet the needs of extreme-scale and data-intensive computing in cosmology, particle physics, climate and other areas of science, NERSC has undertaken an extensive effort to characterize the data demands of the scientific workflows run by its users. In 2015 two important image-analysis workflows—astronomy and microtomography—were characterized to determine the feasibility of adapting these workflows to exascale architectures. Although both workflows were found to have poor thread scalability in general, we also determined that adapting these workflows to manycore and exascale node architectures will not be fundamentally difficult; coarse-grained threading and the use of large private buffers were the principal limits to scalability and optimization strategies to address these bottlenecks already exist.
The I/O requirements of these workflows were also characterized, and in general their I/O patterns are good candidates for acceleration on a flash-based storage subsystem. However, these workflows do not re-read data heavily, and it follows that transparent-caching architectures would not deliver the
The Dark Energy Spectroscopic
Instrument (DESI) will measure
the effect of dark energy on the
expansion of the universe by
obtaining optical spectra for
tens of millions of galaxies
and quasars. DESI will be
conducted on the Mayall
4-meter telescope at Kitt Peak
National Observatory in Arizona
(shown here) starting in 2018.
Image: NOAO/AURA/NSF
Innovations 15
full benefit of the flash subsystem. Rather, these workflows would benefit most from the ability to explicitly stage data into the flash tier before they run.
Although this in-depth study revealed scalability issues in today’s workflows, it did not identify any critical roadblocks to developing a single exascale system architecture that can meet the needs of both existing extreme-scale workloads and these data-intensive workflows. Work is ongoing, and the results of the broader study have been published in a whitepaper that is being provided as a part of the center’s next-generation supercomputer (currently known as NERSC-9) technical requirements.
New Benchmarks for Data-Intensive Computing
The increasing volumes of data that are driving the convergence of extreme-scale and data-intensive computing are accompanied by applications that stress aspects of system performance that have not been traditionally represented in benchmark suites. To ensure that these emerging applications will run well on future architectures, NERSC developed two new benchmark codes that are now a requirement in our next-generation supercomputer (currently known as NERSC-9) benchmarking suite.
In collaboration with LBNL’s Computational Research Division, UC Berkeley and the Joint Genome Institute, NERSC developed Meraculous, a mini-app that performs de novo assembly of large genomes. Algorithmically, this mini-app performs parallel graph construction and traversal, and it relies on communication patterns that have not been stressed in traditional extreme-scale applications. However, these lightweight, single-sided communications are expected to be critical for communication performance between the throughput-optimized cores on exascale architectures.
Another bioinformatics application, BLAST, will become a significant source of the I/O load on the NERSC-9 system as it absorbs the high-throughput computing workflows currently running on a
The genome of bread wheat is
five times larger than that of
humans and contains many
repeat sequences, making it
challenging to obtain a
reference sequence for this
important crop. De novo
assembly algorithms developed
for such large genomes can
dramatically speed data
sequencing on supercomputers
such as NERSC’s Edison
system and future exascale
systems. Image: Quer, https://
commons.wikimedia.org/w/
index.php?curid=15601542
NERSC ANNUAL REPORT 201516
dedicated cluster for Joint Genome Institute users. This pattern-searching application matches a series of unknown DNA or protein sequences with known sequences by repeatedly scanning a database that is stored on disk. This process generates significant I/O loads that lie somewhere between the fully sequential and fully random I/O patterns traditionally benchmarked with the standard IOR benchmark. To address this, NERSC has extended the IOR benchmark to include a new mode that exercises aspects of I/O performance, including page cache and single-thread I/O, that are required by BLAST.
These new benchmarks represent the requirements of the broadening workload that is expected to require exascale-era resources. Optimizing for these data-intensive applications may reveal additional avenues for innovation in the I/O subsystem architectures being developed by vendors, and their inclusion in the procurement requirements for pre-exascale systems should provide a smoother on-ramp to exascale for data-intensive applications.
NERSC Introduces Power Measurement LibraryEnergy efficiency is a key driver in exascale architecture design considerations, and exascale nodes are expected to derive their performance from a large number of power-efficient, low-frequency compute cores. To better understand how application performance and power consumption are related, NERSC developed a profiling library to measure the power consumption of individual applications or subroutines. This library builds upon the low-level power-monitoring infrastructure provided by Cray by allowing application developers to measure the power consumed by a node while certain regions of code are being executed.
This library is currently being used by NERSC and LBNL’s Computational Research Division to explore how compute and I/O performance change as CPU power efficiency increases. For example, the clock speed at which applications achieve peak energy efficiency was found to vary widely across applications due to the varying tradeoff between lower clock speeds and increased wall times. Those applications whose performance is limited by memory bandwidth achieve minimum energy consumption at lower frequencies, while codes that do not saturate the memory bandwidth achieve lower overall energy consumption at higher frequencies.
Similarly, applications that can saturate the I/O bandwidth of a node can achieve lower overall power consumption by reducing the CPU frequency during heavy I/O (for more on this, see https://sdm.lbl.gov/~sbyna/research/papers/201504-CUG-Power.pdf). This work found that the energy savings of running at each application’s energy-optimal frequency were found to be small (less than 10 percent savings) relative to the increase in wall time (more than 10 percent longer) on the Cray XC30 platform. However, this library provides a powerful foundation on which the power-performance behavior of applications and kernels can be evaluated for emerging manycore architectures on the path to exascale.
Innovations 17
Data-Intensive Science
Shifter: Enabling Docker-Like Containers for HPCThe explosive growth in data coming out of experiments in cosmology, particle physics, bioinformatics and nuclear physics is pushing computational scientists to design novel software tools to help researchers better access, manage and analyze that data on current and next-generation HPC architectures.
One increasingly popular approach is container-based computing. In this context, a container is an isolated environment, used to run processes, that is constructed from an “image”—that is, a “snapshot” of an operating environment
for an application or a service that typically contains the required operating system software, libraries and binaries in addition to anything specific to the application.
With the growing adoption of container-based computing, new tools to create and manage containers have emerged—notably Docker, an open-source, automated container deployment service. Docker containers wrap up a piece of software in a complete file system that houses everything it needs to run, including code, runtime, system tools and system libraries. This guarantees that the software will always operate the same, regardless of the environment in which it is running.
Shifter is helping ATLAS, one of
two general-purpose detectors
at the Large Hadron Collider,
run production level work more
efficiently on NERSC’s Edison
system. Image: CERN
NERSC ANNUAL REPORT 201518
While Docker has taken the general computing world by storm in recent years, it has yet to be fully recognized in HPC. There are several reasons for this. First, HPC centers typically provide shared resources that are used by hundreds or, in the case of NERSC, thousands of users. Securely offering full Docker functionality is difficult in these shared environments. Second, Docker uses many advanced Linux kernel features that are not always available on large HPC systems. Finally, Docker is designed to run on systems that have local storage and direct access to the Internet, and HPC systems are typically not designed this way due to cost, performance and maintainability reasons.
To overcome these challenges, NERSC developed Shifter, an open-source, scalable software package that leverages container computing to help supercomputer users run a wider range of software more easily and securely. Shifter enables Docker-like Linux containers and user-defined images to be deployed in an HPC environment—something that to date has not been possible. Shifter is specifically focused on the needs and requirements of the HPC workload, delivering the functionality NERSC and its users are looking for while still meeting the overall performance requirements and constraints that a large-scale HPC platform and shared resource providers impose. Shifter provides performance benefits for data-intensive codes that require many different dependencies because of the way it prepares and distributes the images. It also typically provides the fastest startup times and best scaling performance of any alternative we have measured.
Shifter is not a replacement for Docker; rather, it leverages the user interface that Docker makes available to people to create their software images but does not directly use Docker software internally. Through Shifter, users can supply NERSC with a Docker image that can then be converted to a form that is easier and more efficient to distribute to the compute nodes on a supercomputer. Shifter was initially prototyped and demonstrated on NERSC’s Edison system and is now in production on Cori. It was fully open-sourced and released under a modified BSD license in 2015 and is being distributed via bitbucket (https://bitbucket.org/berkeleylab/shifter/).
NERSC collaborated closely with Cray during the development of Shifter and is working with other HPC centers to broaden its adoption outside of NERSC. At SC15, Cray announced plans to make Shifter available to all Cray XC series users; it will also be made available on the Cray CS400, Cray XE and Cray XK platforms in 2016. NERSC also worked closely with the Swiss National Computing Centre (CSCS), which has deployed and tested Shifter on its Cray CS-Storm cluster.
In 2015, two major science groups began working with Shifter: the LCLS experimental facility at SLAC and the high energy physics community at CERN. The LCLS has its own software environment, which has offered some challenges when migrating it to NERSC’s HPC environment. Using Shifter, however, NERSC staff were able to create a Docker image in one day, demonstrating that this tool greatly reduces the staff effort needed to migrate applications into an HPC environment. In addition, the software loads much faster; in the case of the LCLS, before Shifter was implemented, when an image was ported to NERSC it could take up to half an hour for the software to start. With Shifter, it starts in 5-20 seconds.
Scientists from the Large Hadron Collider’s (LHC) ALICE, ATLAS and CMS experiments at CERN have also been testing Shifter in conjunction with the CERN Virtual Machine File System (CVMFS) software package. CVMFS is a network file system that relies on http to deliver the full LHC software stack to local nodes. It requires fuse and root permissions to install, which has made
Innovations 19
installing it on Cray systems untenable. LHC users have been getting around this by untarring pre-made bundles of their software stack onto Cray compute nodes.
Building on work done by LHC scientists, members of NERSC’s Data and Analytics Services group were able to build a file system image of the CVMFS repository. These images, which are uncharacteristically large for Shifter images, were successfully deployed using the Shifter framework. For example, the ATLAS image contains 300 GB (3.5 TB uncompressed) of data and 20 million files. Despite this, the Shifter framework has been used to run production level work on Edison out to scales of 1,000 nodes, with load times that equaled or surpassed the old tarball method (but with the full software framework instead of a small subset). For groups like this with highly specialized software stacks, Shifter allows them to run jobs as efficiently, if not more so, than they do in their current configurations and requires less effort by the user to port and optimize the codes to run on Cori and Edison.
But Shifter is not just for the largest HPC centers; it is also valuable for departmental and university clusters. Many universities and departments operate small- to medium-sized clusters used by a variety of researchers, and Shifter can enable even these smaller systems to be used more flexibly.
New 400Gbps Production Network Speeds ScienceIn 2015 NERSC collaborated with ESnet to build a 400 Gbps superchannel, the first-ever 400G production link to be deployed by a national research and education network. The connection, known as the BayExpress, provided critical support for NERSC’s 6,000 users as the facility moved from its Oakland, Calif. location to the main LBNL campus.
To develop the 400 Gbps production connection as a step toward the next level of networking, ESnet and NERSC joined with Ciena, a global supplier of telecommunications networking equipment, software and services, and Level 3 Communications, a telecommunications and Internet service provider. The production link is deployed as a pair of 200 Gbps-per-wavelength connections and allows for a simple-to-deploy dual subcarrier 400 Gbps solution.
Map showing the 400G
production link between Wang
Hall at the main LBNL site and
the Oakland Scientific Facility.
NERSC(OSF)
4x100G
4x100G
Wang Hall(Bldg 59)
LBNL
JGI
NERSC
SLACSunnyvale
to San Luis Obispo
to Renoto Eugene
Sacramento
Sandia National LabsLawrence Livermore National Lab
Modesto
Tracy
Stockton
GaltSuisun City
NERSC
LBNLPRODUCTIONSYSTEM
Existing 2xWSS+DIA ROADMExisting ESnet IRU �ber Existing Amp Node
NERSC ANNUAL REPORT 201520
As part of this first-of-its-kind project, the team also set up a 400 Gbps research testbed for assessing new tools and technologies without interfering with production data traffic while allowing staff to gain experience operating a 400 Gbps optical infrastructure. For the testbed, the optical frequencies are more narrowly tuned to maximize use of the fiber spectrum, known as “gridless superchannels.” The testbed uses a dark fiber link provided by Level 3 to ESnet and NERSC for six months. The team conducted field trials on the testbed during the latter part of 2015.
The 400 Gbps superchannel was further put to the test when it was used to relocate file systems from the Oakland Scientific Facility to the new Wang Hall at LBNL. Using features of IBM’s Spectrum Scale file system (formerly GPFS), the NERSC Storage Systems Group successfully moved the 4.8 petabyte /project file system to Wang Hall in November 2015. Achieving transfer speeds up to 170 Gbps over the 400 Gbps link, NERSC engineers balanced the need to quickly complete the file system relocation while minimizing impact to users. Users continued to actively use the file system and noticed little impact from the move.
The last leg of the transfer of file system data to Wang Hall presented an additional challenge. The storage system network at Wang Hall is Infiniband, a high-bandwidth, low-latency switched network technology that is widely used in HPC centers, while data from the 400G superchannel uses the more common Ethernet networking technology. A method was needed to route between the Ethernet and Infiniband networks without degradation in transfer rates. To address this, NERSC deployed an array of 14 commodity compute systems running the open source VyOS network operating system. Unlike typical routers that make packet-forwarding decisions using custom-built hardware, VyOS uses software to route packets. With access to the software source code, NERSC modified the VyOS operating system to support Infiniband and provide failover for the connections on the Infiniband network. The 14 routers were deployed in pairs, with each member in the pair able to stand in for the other if a router failed. With tuning, a single router achieved an aggregate transfer rate of 37 Gbs of the 40 Gbs available.
During 2016, the link between the Oakland and LBNL facilities will remain in place until all systems have been moved to Wang Hall.
Cori Gets a Real-Time QueueNERSC is increasingly being coupled with other user facilities to support analysis from instruments and detectors. In many cases, these use cases have a need for rapid turnaround since users may need to see the results of the analysis to determine the next set of experiments, configuration changes, etc.
To help support these workloads and others with similar needs, Cori now supports a real-time queue. Users can request a small number of on-demand nodes if their jobs have special needs that cannot be accommodated through the regular batch system. The real-time queue enables immediate access to a set of nodes for jobs that are under the real-time wallclock limit. It is not intended to be used to guarantee fast throughput for production runs. To support this queue, NERSC heavily leveraged features within the SLURM batch system.
Since the real-time queue could easily be misused, NERSC has given only a select set of projects access to this capability. To create this list, NERSC first sent a call for proposals to NERSC users asking if they had real-time computing needs with a justification during the allocation process.
Innovations 21
NERSC then reviewed the list of projects and evaluated the justifications they provided. From this list, 15 were accepted, and as of January 2016 four projects are enabled and running in the real-time queue. Many of the submissions were associated with experimental facilities such as the ALS and LCLS or instruments such as telescopes and mass spectrometers.
One example of a project that is already making use of the real-time queue is the Palomar Transient Factory (PTF). The PTF survey uses a robotic telescope mounted on the 48-inch Samuel Oschin Telescope at the Palomar Observatory in Southern California to scan the sky nightly. As soon as the observations are taken, the data travels more than 400 miles to NERSC via the National Science Foundation’s High Performance Wireless Research and Education Network and ESnet. At NERSC, computers running machine-learning algorithms in the Real-Time Transient Detection Pipeline scan through the data and identify events to follow up on. The real-time queues on Cori have allowed the project to go from shutter closed at the telescope to new transients classified in the database and distributed to the collaboration in less than 6 minutes for 95 percent of the images.
Other projects making early use of the real-time queue are the Metabolite Atlas and OpenMSI, which are using this capability to accelerate compound identification from mass spectrometry datasets. Identification is achieved by comparing measured spectra to all theoretically possible fragmentation paths for known molecular structures. To date, these projects have computed and stored complete fragmentation trees to a depth of five for greater than 11,000 compounds. These trees are then used to identify new molecules from raw experimental data using the Cori real-time queue, which allows Metabolite Atlas and OpenMSI users to search their raw spectra against these trees and obtain results in minutes. Without systems like Cori, these tasks would take months or would not be performed at all. The real-time queue enables users to make better use of their time, avoiding the variable wait times previously experienced using the standard batch queues.
iPython/JupyterHub Portal DebutsThe Jupyter Notebook is a web application that allows users to create and share documents that contain live code, equations, visualizations and explanatory text. JupyterHub is a multi-user version of the notebook designed for centralized deployments.
NERSC has deployed a JupyterHub service that allows users to spin up interactive iPython notebooks with access to their data and computational capabilities of the large-scale NERSC platforms. By integrating it with NERSC LDAP authentication, each user can start a private notebook on the NERSC JupyterHub server and run interactive computations through a Python shell. This can be used for interactive development and prototyping but is also useful for workflow orchestration and advanced visualization. For example, the OpenMSI project is using iPython notebooks to enable users to analyze and visualize mass spectrometry data.
NERSC’s JupyterHub service also gives users the ability to customize their notebook environments and include their own packages and libraries. This makes it easy to include domain-specific or
The Palomar 48-inch telescope
and time-lapse image of the
night sky. The Palomar
Transient Factory uses the
real-time queue at NERSC to
identify candidate transient
objects for further analysis.
Image: Caltech
NERSC ANNUAL REPORT 201522
community-developed packages for specific forms of analysis or visualization. This powerful tool combines the expressiveness of a language like Python with the ability to perform interactive analyses through a web portal. One can interact with vast datasets stored on NERSC file systems and easily access Python tools for visualizing and analyzing the data.
Despite being only in beta release mode during 2015 and not broadly advertised, the iPython/JupyterHub deployment has already been used by over 100 users in domains spanning biology, cosmology, high-energy physics and climate. NERSC made this a full production service in early 2016 and is already exploring ways to more tightly integrate it with Cori to enable advanced analytics and workflow orchestration, including the ability to direct the execution of parallel jobs via iPython.
Cray-AMPLab-NERSC Partnership Targets Data AnalyticsAs data-centric workloads become increasingly common in scientific and industrial applications, a pressing concern is how to design large-scale data analytics stacks that simplify analysis of the resulting data. A collaboration between Cray, researchers at UC Berkeley’s AMPLab and NERSC is working to address this issue.
The need to build and study increasingly detailed models of physical phenomena has benefited from advancement in HPC, but it has also resulted in an exponential increase in data from simulations as well as real-world experiments. This has fundamental implications for HPC systems design, such as the need for improved algorithmic methods and the ability to exploit deeper memory/storage hierarchies and efficient methods for data interchange and representation in a scientific workflow. The modern HPC platform has to be equally capable of handling both traditional HPC workloads and the emerging class of data-centric workloads and analytics motifs.
In the commercial sector, these challenges have fueled the development of frameworks such as Hadoop and Spark and a rapidly growing body of open-source software for common data analysis and machine learning problems. These technologies are typically designed for and implemented in
Orange highlights show
portions of the Berkeley Data
Analytics Stack that are being
explored as part of the Cray/
AMPLab/NERSC collaboration.
In-house Apps BioImaging Neuroscience Climate
MLBase
MLPipelinesSparkR
Spark Core (JVM, Sockets, Shuffling)
HDFS,NetCDF
HDFS, S3, Ceph
Hadoop YarnMesos
Succint
Tachyon
SampleClean
SparkStreaming
G-OLA
BlinkDB
SparkSQLVelox
GraphX
MLlib
Access andInterfaces
Processing Engine
Storage
Resource Virtualization
Lustre
Innovations 23
distributed data centers consisting of a large number of commodity processing nodes, with an emphasis on scalability, fault tolerance and productivity. In contrast, HPC environments are focused primarily on no-compromise performance of carefully optimized codes at extreme scale.
This scenario prompted the Cray, AMPLab and NERSC team to look at how to derive the greatest value from adapting productivity-oriented analytics tools such as Spark to HPC environments, and how to ensure that a framework like Spark can exploit supercomputing technologies like advanced interconnects and memory hierarchies to improve performance at scale without losing productivity benefits. The team is actively examining research and performance issues in getting Spark up and running on HPC environments, such as NERSC’s Edison and Cori systems, with a long-term goal of bridging the gap between data analytics for commercial applications and high-performance scientific applications.
In addition, since linear algebra algorithms underlie many of NERSC’s most pressing scientific data-analysis problems, the team is developing novel randomized linear algebra algorithms and implementing these algorithms within the AMPLab stack and on Edison and Cori (dubbed the Berkeley Data Analytics Stack). They are also applying these algorithms to some of NERSC’s most pressing scientific data-analysis challenges, including problems in bioimaging, neuroscience and climate science.
Two of the main performance issues are file system I/O and node memory. The performance of Spark applications is dictated by the spark “shuffle,” during which data is sent between nodes through a file system spill. Traditionally, commercial data centers have local disk, so shuffle performance is network-bound. In contrast, HPC systems like Edison and Cori have global file systems, so shuffle performance is dictated primarily by file system metadata latency. In addition, if a node runs out of memory during the shuffle, it will spill intermediately to the file system. HPC systems typically have less memory per node (and, thus, per core) compared to industry data centers. For large memory workloads, such as linear algebra algorithms on large matrices, HPC systems have to spill to the file system more often than data center workloads, which results in more of a hit on performance.
In order to see this difference in performance, we compared the performance of randomized linear algebra algorithm CX on a real scientific use case, a 1TB bioimaging dataset in a paper submitted to the 2016 IEEE International Parallel & Distributed Processing Symposium. To close some of the performance gaps observed in this use case, we tried a variety of techniques in collaboration with Cray and CRD to decrease the number of metadata operations in a Spark application. Testing has shown that these techniques can significantly improve performance, especially for workloads that impose a heavy shuffle load.
Deep Learning Comes to HPCIn 2015 researchers from NERSC and LBNL began applying deep learning software tools developed for HPC environments to a number of “grand challenge” science problems running computations at NERSC and other supercomputing facilities.
Deep learning, a subset of machine learning, is the latest iteration of neural network approaches to machine learning problems. Machine learning algorithms enable a computer to analyze pointed
NERSC ANNUAL REPORT 201524
examples in a given dataset, find patterns in those datasets and make predictions about what other patterns it might find.
Deep learning algorithms are designed to learn hierarchical, nonlinear combinations of input data. They get around the typical machine learning requirements of designing custom features and produce state-of-the-art performance for classification, regression and sequence prediction tasks. While the core concepts were developed over three decades ago, the availability of large data, coupled with data center hardware resources and recent algorithmic innovations, has enabled companies like Google, Baidu and Facebook to make advances in image search and speech recognition problems.
Surprisingly, deep learning has so far not made similar inroads in scientific data analysis, largely because the algorithms have not been designed to work on high-end supercomputing systems such as those found at NERSC. But during 2015 NERSC began enabling the Daya Bay Neutrino Experiment to apply deep learning algorithms to enhance its data analysis efforts. Located in China, Daya Bay is an international neutrino-oscillation experiment designed to determine the last unknown neutrino mixing angle θ13 using anti-neutrinos produced by the Daya Bay and Ling Ao Nuclear Power Plant reactors. NERSC’s Data and Analytics Services Group implemented a deep learning pipeline for Daya Bay at NERSC, using convolutional neural nets to automatically reduce and classify the features in the data. The work produced a supervised classifier with over 97 percent accuracy across different classes, as well as an unsupervised convolutional autoencoder capable of identifying patterns of physics interest without specialized input knowledge.
To our knowledge, this is the first time unsupervised deep learning has been performed on raw particle physics data. NERSC is now collaborating with physicists to investigate in detail the various clusters formed by the representation to determine what interesting physics is captured in them beyond the initial labeling and to incorporate such visualizations into the monitoring pipeline of the experiment.
NERSC's Data and
Analytics Services Group
has implemented a deep
learning pipeline for the
Daya Bay Neutrino
Experiment Facility.
Image: Roy Kaltschmidt,
Lawrence Berkeley
National Laboratory
Innovations 25
Operational Excellence
Thruk: A New Fault-Monitoring Web InterfaceThe move to Wang Hall prompted NERSC’s Operational Technology Group (OTG)—the team that ensures site reliability of NERSC systems, storage, facility environment and ESnet—to implement several new fault-monitoring efficiencies. Foremost among them is Thruk, an enhancement to OTG’s nagios fault-monitoring software that uses the Livestatus API to combine all instances of systems and storage into a single view.
Implementing Thruk provided OTG with an efficient method for correlating various alerts and determining if an issue affects systems throughout Wang Hall or just in certain parts of the facility. The most important advantage of this new tool is that it allows the OTG to monitor multiple aspects of systems in a single view instead of multiple tabs or multiple url interfaces. This has decreased the possibility of missing critical alerts. Other advantages include:
• A single location to view alerts to multiple systems
• A configuration interface that is flexible, allowing the OTG to easily add and remove systems
• The ability to send multiple commands at once without waiting for a response
• Multiple filtering schemes
Cori Phase I installation in the
new Wang Hall computer room.
Image: Kelly Owen, Lawrence
Berkeley National Laboratory
NERSC ANNUAL REPORT 201526
Thruk has also been instrumental in supporting the OTG’s efforts to improve workflow automation. In 2015 OTG staff implemented an automated workflow for timely reporting of user issues that uses program event handlers to manage multiple instances of issues and multiple action items. For example, alerts that called for a staff person to open a new ticket and assign it to a group can now be automated based on a filtering mechanism. The new interface saves the group’s workflow approximately 17 hours per week, in addition to tickets now being opened accurately on a more consistent basis. In the past, the language and the information included on the ticket depended on which staff member opened the ticket.
Another advantage of the automation process involves opening tickets with vendors as a result of an alert. In the past, a staff person would either e-mail or call the vendor call center to report a problem. The phone call could take anywhere from 15 minutes to 45 minutes, depending on the vendor, and the time it took for the vendor tech to respond could be anywhere from one to four hours. In contrast, the automated process sends an e-mail to the vendor call center, which usually puts a priority on clearing out their inbox of requests. Thus a ticket sent to the vendor can be opened within 15 minutes, providing detailed information such as diagnostic logs as an attachment to the e-mail. This has resulted in improving vendor tech response times to under one hour. In addition, with the appropriate information provided with each alert, vendors spend less time on diagnosis and more time on problem resolution.
The event handlers can also create and assign work tickets to staff. For example, an alert that says a drive has failed and needs replacement can activate an event handler that opens a ticket and assigns it to a staff person who can replace that drive, work with vendors for parts under warranty and allow the staff to bring the node back into production sooner. One of the areas that allowed this process to be automated involves centralizing the location of the warrantied parts with their purchase date, serial number, vendor information and warranty expiration date. With the verification process handled automatically, it now takes only seconds to decide whether to exchange or acquire a part. This has resulted in compute nodes coming back to production in a shorter amount of time—in some instances as little as four to eight hours, where previously it could sometimes take twice that long.
Screen shot of the new
Thruk interface.
Innovations 27
Environmental Data Collection Projects With the move to Wang Hall, NERSC is also taking the opportunity to collect environmental data from areas of the machine room where we had not previously done so at the Oakland Scientific Facility. We’ve installed sensors in different areas of the computational floor such as underneath the racks and even on the ceiling. We felt that getting a wide range of data on the environment will assist us in correlating between events that could affect the center and jobs running on the computing resources.
For example, we are collecting data from sensors at the substation, through the different levels of PDUs available in the building down to the node level if possible, including the UPS/generator setup. The data is collected from the subpanels and passed to the central data collection system via a Power Over Ethernet (PoE) network setup, since many of these switches and panels are located far away from the traditional network infrastructure. In one case, the substation might not include normal 110 volt power for the sensors and hardware, which is why PoE is an easy way to get both networking and power to these remote locations. We plan to eventually install more than 3,000 sensors, with each sensor collecting 10 data points every second.
Because Wang Hall is so dependent on the environment, we can leverage all the environmental data to determine optimal times to run computation-intensive jobs or small jobs or potentially predict how the environment affects computation on a hot summer day.
Centralizing Sensor Data Collection In addition to environmental sensors, NERSC's move to Wang Hall also prompted opportunities for the OTG to install other types of sensors in areas where this had not been done previously. These include temperature sensors from all parts of the computer floor and from all levels (floor to ceiling), power sensors from the substations to the PDUs in the rack, water flow and temperature sensors for the building and large systems and host and application measurement from all running systems. The data collected from these sensors gives us more ways to correlate causes to issues than was previously possible. We believe that this data should be able to answer questions such as power consumption of a job, power efficiency of systems for processing jobs or how we can leverage a schedule based on size of job to time of day or even the current weather.
However, the additional sensors generate a large amount of data that needs to be stored, queried, analyzed and displayed, and our existing infrastructure needed to be updated to meet this challenge. Thus, during 2015 the OTG created a novel approach to collecting system and environmental data and implemented a centralized infrastructure based on the open source ELK (Elasticsearch, Logstash and Kibana) stack. Elasticsearch is a scalable framework that allows for fast search queries of data; Logstash is a tool for managing events and parsing, storing and collecting logs; and Kibana is a data visualization plugin that we use on top of the indexed content collected by Elasticsearch.
Using the ELK stack allows us to store, analyze and query the sensor data, going beyond system administration and adding the ability to extend the analysis to improve the operational efficiency of an
NERSC ANNUAL REPORT 201528
entire facility. This process involves collecting information on hosts or sensors across the center. This data is then passed into the collection environment via RabbitMQ—an open-source message broker software—which sends and stores the data to the Elasticsearch framework. RabbitMQ is capable of passing data to many targets, and we leveraged this feature to send the same data to Redis (an open-source, in-memory data structure store) for real-time access to some data, and to Freeboard (an open-source dashboard) to display real-time data without any storage ability. We anticipate using this method to send data to multiple targets to address future collection issues as they arise.
All of the acquired sensor data is stored in a commonly accessible format—few changes need to be made when collecting new data or analyzing the data—that is centrally accessible to most staff without transferring the data from one storage method to another or re-collecting the data in a different configuration. This new method is already providing potential data analysis opportunities not previously possible.
Compact Cori: An Educational Tool for HPCIn 2015 NERSC developed an HPC educational tool designed to inspire students of all ages, including middle and high-school levels, and demonstrate the value of HPC to the general public. A team of NERSC staff and student interns (two undergraduates and one high school student) created “Compact Cori,” a micro-HPC system based on Intel NUC (next unit of computing) hardware and a modern software stack based on python, MPI4PY, REST APIs, WebGL, Meteor and other modern web technologies.
The hardware design was inspired by the “Tiny Titan” micro-HPC developed at Oak Ridge National Laboratory. The hardware and software stack were chosen to be accessible and to represent modern programming techniques and paradigms with richer, more accessible educational value than traditional HPC languages like C and FORTRAN. The software stack allows students to explore several novel design points for HPC.
Elastic cluster
Multiple Logstash
Rabbitmq cluster
ELK Stack
Closed shards that are archived and can be loaded later by other elastic cluster instances. This included Mac and PC stacks.
custom methods(text or json)
Archive syslog data to disk.Possible custom methods
logstash followerfor text filesgenerated byother programs
rsyslog collectd
Other output methods can be addedsuch as mysql, mongoDB, etc.This can be a subset of the data collectedor the full data.This data can also be shipped to othercenters either over ssl or not.
Archive collectd and other timeseries data to hdf5 format
Per system forwarding logstash
logstash follower fortext files generated byother programs
Power and environmentaldata. Direct from sensorsor a collector.text and json format
Kibana graphical interfacefor data exploration and display
Event and time series datastored here. Master for fullCollectd, power and correlation or data.
json inputs sentto logstash toenter into Elastic
Redis DB for Lustre and other datafor immediate access
Redis DB for power and coolinglimited to under 5 minutes
LDMSProcmonamqp inputs
Collectd, power and environmental datasent to rabbitmq forother consumers
The new centralized data
collection infrastructure
at NERSC.
Innovations 29
Despite being targeted for an educational mini-HPC, the software infrastructure represents a realistic model that can be applied to larger computing problems on production HPC resources. In particular, we developed a web-based, real-time, immersive 3D visualization technique for parallel simulations on Compact Cori. The user interface uses a web-based video game engine, a visualization concept new to HPC that allows for a rich human-computer interaction and realistic modeling of scientific data and is a compelling educational device. The web-based, real-time visualization framework represents a tangible advancement in the ability to engage users via science gateways as opposed to the traditional terminal/ssh type interaction.
Compact Cori is easy to transport and set up and carries a configurable price tag in the range of a few thousand dollars. Compact Cori is available for interactive demonstrations at NERSC.
Left: Compact Cori uses 16 Intel
NUC boards and a mobile
Broadwell CPU. The NUC boards
are linked via Gigabit Ethernet.
Right: An immersive 3D
visualization technique
demonstrated on Compact Cori.
Science HighlightsThis section presents a selection of research highlights from 2015 based on computations and simulations run at NERSC, illustrating the breadth of science supported by the center for all six DOE program offices.
NERSC ANNUAL REPORT 201530
Science Highlights
Publication:
S.W. French, B. Romanowicz, “Broad
plumes rooted at the base of the Earth’s
mantle beneath major hotspots,” Nature
525, 95-99, September 3, 2015,
doi:10.1038/nature14876
Full Story:
http://bit.ly/NERSCarMantlePlumes
A 3D rendering of shear wave speed
anomalies beneath the Central Pacific
between the core-mantle boundary
(2891 km depth) and 400 km depth.
Green cones and labels highlight
locations of key hotspot volcanoes in
this region. Image: Scott French
Simulations Link Mantle Plumes with Volcanic Hotspots
OBJECTIVE
To demonstrate the relationship between plumes of hot rock rising through the Earth’s mantle with surface hotspots that generate volcanic island chains.
FINDINGS/ACCOMPLISHMENTS
Numerical simulations run at NERSC helped UC Berkeley seismologists produce for the first time a 3D scan of Earth’s interior that conclusively connects plumes of hot rock rising through the mantle with surface hotspots that generate volcanic island chains like Hawaii, Samoa and Iceland. Until this study, evidence for the plume and hotspot theory had been circumstantial, and some seismologists argued instead that hotspots are very shallow pools of hot rock feeding magma chambers under volcanoes.
Previous attempts to image mantle plumes have detected pockets of hot rock rising in areas where plumes have been proposed to exist. However, it was unclear whether they were connected to volcanic hotspots at the surface or at the roots of the plumes at the core mantle boundary 2,900 kilometers below the surface. The new, high-resolution map of the mantle showed the plumes were connected to many volcanic hotspots at the surface of the earth. It also revealed that below about 1,000 km under the surface of the Earth the plumes are between 600 and 1,000 km across, up to five times wider than geophysicists thought and at least 400 degrees Celsius hotter than surrounding rock.
RESEARCH DETAILS
To create a high-resolution computed tomography image of Earth, the researchers used very accurate numerical simulations of how seismic waves travel through the mantle. They then compared their predictions to the ground motion measured by detectors around the globe. They mapped mantle plumes by analyzing the paths of seismic waves bouncing around Earth’s interior after 273 strong earthquakes that shook the globe over the past 20 years.
Earlier attempts by other research groups often approximated the physics of wave propagation and focused mainly on the arrival times of only certain types of seismic waves, such as the P (pressure) and S (shear) waves, which travel at different speeds. In this case, the research team computed all components of the seismic waves, such as their scattering and diffraction, then tweaked the model repeatedly to fit recorded data using a method similar to statistical regression. The final computation required 3 million CPU hours on NERSC's Edison system.
Principal Investigator:
Barbara Romanowicz, UC Berkeley
Imaging and Calibration of Mantle Structure at Global and Regional Scales Using Full-Waveform Seismic Tomography
BES—Geosciences
31
What the Blank Makes Quantum Dots Blink?
NERSC ANNUAL REPORT 201532
OBJECTIVE
Quantum dots are nanoparticles of semiconductor that can be tuned to glow in a rainbow of colors. Since their discovery in the 1980s, these nanoparticles have held out tantalizing prospects for all kinds of new technologies, from paint-on lighting materials and solar cells to quantum computer chips, biological markers and even lasers and communications technologies.
Unfortunately, quantum dots often “blink,” and this “fluorescence intermittency” has put a damper on many potential applications; for example, lasers and logic gates won’t work very well with iffy light sources. Quantum dots can also absorb specific colors of light, but using them to harvest sunlight in photovoltaics is not yet very efficient, due in part to the mechanisms behind blinking.
FINDINGS/ACCOMPLISHMENTS
University of Chicago scientists computing at NERSC used ab initio simulations to probe the mysterious blinking process in silicon quantum dots. Their findings could help scientists fix the problem and pave the way for a number of new applications for this technology. The results are the first reported ab initio calculations showing that dangling bonds on the surface of oxidized silicon nanoparticles can act as efficient non-radiative recombination centers. These techniques could also be used to tackle the effects of trapping in solar cells, which can limit the efficiency of solar cells.
RESEARCH DETAILS
The team used simulated silicon (Si) nanoparticles configured with various defects and coated with silicon dioxide. Starting with three different possible defect states, they used the Hopper supercomputer to calculate the optical and electronic properties of the oxidized silicon nanoparticle with the scientific package called Quantum Espresso.
To perform their calculations, which took 100,000 processor hours on Hopper, the team first constructed virtual models. They computationally carved virtual holes out of a crystalline silicon oxide (SiO2) matrix and inserted silicon quantum dots of various sizes, computing cycles of annealing and cooling to create a more realistic interface between the quantum dots and the SiO2 matrix. Finally, dangling bond defects were introduced at the surface of quantum dots by removing a few selected atoms.
By computing the electronic properties and the rate at which electrons release energy, they found that trapped states do indeed cause quantum dot dimming. Dangling bonds on the surface of silicon nanoparticles trapped electrons where they recombined “non-radiatively” by releasing heat. That is, the electrons shed excess energy without radiating light. Dimming also depended on the overall charge of the entire quantum dot, the team found.
Principal Investigator:
Márton Vörös, University of Chicago
Large-Scale Calculations on Nanostructured Heterogeneous Interfaces
BES—Materials Sciences
Publication:
N.P. Brawand, M. Vörös, G. Galli,
“Surface dangling bonds are a cause of
B-type blinking in Si nanoparticles,”
Nanoscale, 2015, 7, 3737-3744, doi:
10.1039/C4NR06376G
Full Story:
http://bit.ly/NERSCarQuantumDots
Silicon quantum dots are shown in
various states of “blinking.” The “on”
crystals emit light (represented by a
white dot) as an excited electron sheds
excess energy as a photon. The “off”
crystals are dark because their electrons
(yellow) are trapped in surface defects
and siphon off energy through other
paths, such as heat or lattice vibrations.
Image: Peter Allen, University
of Chicago
Science Highlights 33
33
Stabilizing a Tokamak Plasma
OBJECTIVE
To use 3D simulations to gain new insights into the behavior of fusion plasmas and find improved methods for ensuring plasma stability in future tokamak reactors.
FINDINGS/ACCOMPLISHMENTS
A team of physicists from Princeton Plasma Physics Laboratory, General Atomics and the Max Planck Institute for Plasma Physics used NERSC’s Edison computer to discover a mechanism that prevents the electrical current flowing through fusion plasma from repeatedly peaking and crashing. This behavior is known as a “sawtooth cycle” and can cause instabilities within the plasma’s core.
Learning how to model and study the behavior of fusion plasmas has important implications for ITER, the multinational fusion facility being constructed in France to demonstrate the practicality of fusion power. Instabilities in the plasma could destabilize and halt the fusion process. Preventing the destabilizing cycle from starting would therefore be highly beneficial for the ITER experiment and future tokamak reactors.
RESEARCH DETAILS
Running M3D-C1—a program that creates 3D simulations of fusion plasmas—on Edison, researchers found that under certain conditions a helix-shaped whirlpool of plasma forms around the center of the tokamak. The swirling plasma acts like a dynamo—a moving fluid that creates electric and magnetic fields. Together these fields prevent the current flowing through plasma from peaking and crashing.
During the simulations the scientists were able to virtually add new diagnostics, or probes, to the computer code, which measure the helical velocity fields, electric potential, and magnetic fields to clarify how the dynamo forms and persists. The persistence produces the voltage in the center of the discharge that keeps the plasma current from peaking.
Principal Investigator:
Stephen Jardin, Princeton Plasma
Physics Laboratory
Publication:
S. C. Jardin, N. Ferraro, and I. Krebs,
“Self-Organized Stationary States of
Tokamaks,” Phys. Rev. Lett. 115, 215001,
November 17, 2015; DOI: http://dx.doi.
org/10.1103/PhysRevLett.115.215001
Full Story:
http://bit.ly/NERSCarTokamakPlasma
A cross-section of the virtual plasma
showing where the magnetic field lines
intersect the plane. The central section
has field lines that rotate exactly once.
Image: Stephen Jardin, Princeton
Plasma Physics Laboratory
3D Extended MHD Simulation of Fusion Plasmas
FES—Fusion SciDAC
Experiments + Simulations = Better Nuclear Power Research
NERSC ANNUAL REPORT 201534
OBJECTIVE
An international collaboration of physicists is working to improve the safety and economics of nuclear power by studying how various cladding materials and fuels used in reactors respond to radiation damage.
FINDINGS/ACCOMPLISHMENTS
The research team—which includes representatives from Los Alamos, Pacific Northwest and Idaho national laboratories as well as institutions in Germany and Belgium—used molecular dynamics simulations run at NERSC to identify atomic-level details of early-stage damage production in cerium dioxide (CeO2), a surrogate material used in nuclear research to understand the performance of uranium dioxide (UO2).
RESEARCH DETAILS
The researchers coupled swift heavy ion experiments and electron microscopy analysis with parallel simulations—one of the few instances of combining physical experiments and simulations for this kind of research. The physical experiments involved implanting CeO2 with 940 MeV gold ions and performing very careful high-resolution electron microscopy studies to understand the structure of the damaged material. Because the evolution of radiation damage is controlled by events that take place on the nanosecond scale, the simulations—run on NERSC’s Hopper system—helped fill gaps in experimental understanding by simulating these transient events that take place on very small time and length scales.
While the molecular dynamics code (DL_POLY) the researchers used is fairly common, the simulations themselves were very computationally intensive; the team simulated a system of 11 million atoms for a fairly long time, something that couldn’t have been done without a massively parallel computing resource.
.
Principal Investigator:
Ram Devanathan, Pacific Northwest
National Laboratory
Radiation Effects on the Mechanical Behavior of Ceramics
BES—Materials Science
Publication:
C.A. Yablinsky, R. Devanathan,
J. Pakarinen, J. Gan, D. Severin,
C. Trautmann, T.R. Allen,
“Characterization of swift heavy ion
irradiation damage in ceria,” Journal of
Materials Research 30, 9, May 14, 2015.
doi: 10.1557/jmr.2015.43
Full Story:
http://bit.ly/NERSCarNuclearPower
35
Computer simulations run at NERSC
show estimates of ice retreat in the
Amundsen Sea Embayment by 2154.
Image: Stephen Cornford, University
of Bristol
BISICLES Ice Sheet Model Quantifies Climate Change
OBJECTIVE
West Antarctica is one of the fastest warming regions on Earth, and its ice sheet has seen dramatic thinning in recent years. The ice sheet is losing significant amounts of ice to the ocean, with the losses not being offset by snowfall. The acceleration of West Antarctic ice streams in response to ocean warming could result in a major contribution to sea-level rise, but previous models were unable to quantify this.
FINDINGS/ACCOMPLISHMENTS
Using the high-resolution, large-scale computer model known as Berkeley-ISICLES (BISICLES), researchers were able to estimate how much ice the West Antarctic Ice Sheet could lose over the next two centuries and how that could impact global sea-level rise. These comprehensive, high-resolution simulations are a significant improvement from previous calculations (which were lower in resolution or scale) and will enable climate scientists to make more accurate predictions about West Antarctica’s future.
RESEARCH DETAILS
The BISICLES adaptive mesh ice sheet model was used to carry out one, two, and three century simulations of the fast-flowing ice streams of the West Antarctic Ice Sheet, deploying sub-kilometer resolution around the grounding line since coarser resolution results in substantial underestimation of the response. The research team subjected the BISICLES model to a range of ocean and atmospheric change: no change at all, future changes projected by ocean and atmosphere models and extreme changes intended to study the upper reaches of future sea-level rise. The results of global climate models were fed into regional models of the Antarctic atmosphere and ocean, whose results were in turn used to force the ice-sheet model in this study. Some of the BISICLES simulations were run on NERSC's Hopper and Edison systems.
Principal Investigators:
Dan Martin, Esmond Ng, Lawrence
Berkeley National Laboratory
Predicting Ice Sheet and Climate Evolution at Extreme Scales
ASCR—Applied Mathematical Sciences
Publication:
S.L. Cornford, D.F. Martin, et al,
“Century-scale simulations of the
response of the West Antarctic Ice Sheet
to a warming climate,” The Cryosphere,
9, 1579-1600, 2015, doi: 10.5194/
tc-9-1579-2015
Full Story:
http://bit.ly/NERSCarIceSheetModel
Science Highlights
Tall Clouds from Tiny Raindrops Grow
NERSC ANNUAL REPORT 201536
OBJECTIVE
“Big clouds get bigger while small clouds shrink” may seem like a simple concept, but the mechanisms behind how clouds are born, grow and die are surprisingly complex. So researchers from Pacific Northwest National Laboratory (PNNL) used supercomputing
resources at NERSC to simulate tropical clouds and their interaction with the warm ocean surface and compare the simulations to real-world observations. Scientists need to understand clouds to effectively predict weather patterns, including the potential for droughts and floods.
FINDINGS/ACCOMPLISHMENTS
By running simulations on NERSC’s Hopper Cray XE6 system, the PNNL researchers found that factors as small as how sizes of raindrops were represented in a computer model made a big difference in the accuracy of the results. The simulations clearly showed that larger clouds tend to grow larger because they capture less dry air, while smaller clouds dwindle away.
RESEARCH DETAILS
The PNNL team collaborated with researchers from NASA Goddard Space Flight Center and the National Center for Atmospheric Research to run high-resolution regional models that simulated cloud lifecycles with different ranges of raindrop size. They compared those results to observational data collected from the Atmospheric Radiation Measurement Climate Research Facility’s Madden-Julian Oscillation Investigation Experiment and the Earth Observing Laboratory’s Dynamics of the Madden-Julian Oscillation research campaign, which collected data in the South Pacific from October 2011 to March 2012 to support studies on the birth, growth and evolution of certain types of clouds.
The scientists used satellite and ground-based radar measurements from the campaign to examine how well the Weather Research and Forecasting Model simulated tropical clouds. They looked specifically at four approaches to handling the physics of tiny raindrops and found that two key factors were rain rates at ground level and the modeling of “cold pools”—the cold, dry air flowing out from deep thunderstorms.
Principal Investigator:
Samson Hagos, Pacific Northwest
National Laboratory
Evolution in Cloud Population Statistics of the Madden-Julian Oscillation
BER—Climate and Environmental Sciences
Publication:
S. Hagos, Z. Feng, C. Burleyson, K-S.
Lim, C. Long, D. Wu, G. Thompson,
“Evaluation of Convection-Permitting
Model Simulations of Cloud Populations
Associated with the Madden-Julian
Oscillation using Data Collected during
the AMIE/DYNAMO Field Campaign,”
Journal of Geophysical Research:
Atmospheres 119(21):12,052-12,068.
doi:10.1002/2014JD022143
Full Story:
http://bit.ly/NERSCarCloudScience
PNNL scientists used real-world
observations to simulate how small
clouds are likely to stay shallow, while
larger clouds grow deeper because
they mix with less dry air. Image: NASA
Johnson Space Center
Science Highlights
Supernova Hunting with Supercomputers
OBJECTIVE
Because the relative brightness of Type Ia supernovae can be measured so well no matter where they are located in the Universe, they make excellent distance markers. In fact, they were instrumental to measuring the accelerating expansion of the Universe in the 1990s—a finding that netted three scientists the 2011 Nobel Prize in Physics. Yet, astronomers still do not fully understand where they come from.
FINDINGS/ACCOMPLISHMENTS
Using a “roadmap” of theoretical calculations and supercomputer simulations performed at NERSC, astronomers observed for the first time a flash of light caused by a supernova slamming into a nearby star, allowing them to determine the stellar system—IC831—from which the supernova—iPTF14atg—was born. This finding confirms one of two competing theories about the birth of Type Ia supernovae: the single-degenerate model.
The intermediate Palomar Transient Factory (iPTF) analysis pipeline found the light signal from the supernova just hours after it ignited in a galaxy about 300 million light years away from Earth. iPTF depends on NERSC computational and global storage resources.
RESEARCH DETAILS
On May 3, 2014, the iPTF took images of IC831 and transmitted the data for analysis to computers at NERSC, where a machine-learning algorithm analyzed the images and prioritized real celestial objects over digital artifacts, leading to the discovery of the young supernova.
Principal Investigators:
Peter Nugent, Lawrence Berkeley
National Laboratory; Stan Woosley,
University of California, Santa Cruz
Palomar Transient Factory; Type Ia Supernovae
HEP—Cosmic Frontier
Publication:
Y. Cao, S.R. Kulkarni, D. Andrew Howell,
et al, “A strong ultraviolet pulse from a
newborn type 1a supernova,” Nature
421, 328-331, May 21, 2015, doi:
10.1038/nature14440
Full Story:
http://bit.ly/NERSCarSupernovaHunting
Simulation of the expanding debris from
a supernova explosion (shown in red)
running over and shredding a nearby star
(shown in blue). Image: Daniel Kasen,
Lawrence Berkeley National Laboratory/
UC Berkeley
37
NERSC ANNUAL REPORT 201538
Petascale Pattern Recognition Enhances Climate Science
OBJECTIVE
To demonstrate how a new data analytics tool developed at Berkeley Lab uses pattern recognition to help climate researchers better detect extreme weather events, such as
cyclones and atmospheric rivers, in large datasets.
FINDINGS/ACCOMPLISHMENTS
Modern climate simulations produce massive amounts of data, requiring sophisticated pattern recognition algorithms to be run on terabyte- to petabyte-sized datasets. Using the new Toolkit for Extreme Climate Analysis (TECA), researchers from Berkeley Lab and Argonne National Laboratory demonstrated how TECA, running at full scale on NERSC’s Hopper system and Argonne’s Mira system, reduced the runtime for pattern detection tasks from years to hours.
“TECA: Petascale Pattern Recognition for Climate Science,” a paper presented by the team at the 16th International Conference on Computer Analysis of Images and Patterns (CAIP) in September 2015, was awarded CAIP’s Juelich Supercomputing Center prize for the best application of HPC technology in solving a pattern recognition problem.
RESEARCH DETAILS
For this project, the researchers downloaded 56TB of climate data from the fifth phase of the Coupled Model Intercomparison Project (CMIP5) to NERSC. Their goal was to access subsets of those datasets to identify three different classes of storms: tropical cyclones, atmospheric rivers and extra-tropical cyclones. All of the datasets were accessed through a portal created by the Earth Systems Grid Federation to facilitate the sharing of data; in this case, the data consisted of atmospheric model data sampled at six-hour increments running out to the year 2100. The data was stored at 21 sites around the world, including Norway, the United Kingdom, France, Japan and the U.S. NERSC’s Hopper system was used to preprocess the data, which took about two weeks and resulted in a final 15TB dataset.
Principal Investigator:
Travis O’Brien, Lawrence Berkeley
National Laboratory
Calibrated and Systematic Characterization, Attribution and Detection of Extremes
BER—Climate and Environmental Sciences
Publication:
“TECA: Petascale Pattern Recognition
for Climate Science,” 16th International
Conference, CAIP 2015, Valletta, Malta,
September 2-4, 2015
Full Story:
http://bit.ly/NERSCarPatternRecognition
TECA implements multivariate threshold
conditions to detect and track extreme
weather events in large climate datasets.
This visualization depicts tropical cyclone
tracks overlaid on atmospheric flow
patterns. Image: Prabhat, Lawrence
Berkeley National Laboratory
Science Highlights
Celeste: A New Model for Cataloging the Universe
OBJECTIVE
To demonstrate how Celeste—a new statistical analysis model developed by a collaboration of astrophysicists, statisticians and computer scientists from UC Berkeley, Harvard, Carnegie Mellon and Berkeley Lab—can automate digital sky surveys and make it possible to create a single catalog of the entire universe. Celeste is designed to be a better model for identifying the astrophysical sources in the sky and the calibration parameters of each telescope. It also has the potential to significantly reduce the time and effort that astronomers currently spend working with image data.
FINDINGS/ACCOMPLISHMENTS
The research team spent much of 2015 developing and refining Celeste, a hierarchical model designed to catalog stars, galaxies and other light sources in the universe visible through the next generation of telescopes. It also enables astronomers to identify promising galaxies for spectrograph targeting, define galaxies they may want to explore further and help them better understand dark energy and the geometry of the universe. Celeste uses analytical techniques common in machine learning and applied statistics but not so much in astronomy, implementing a statistical inference to build a fully generative model that can mathematically locate and characterize light sources in the sky.
RESEARCH DETAILS
The researchers have so far used Celeste to analyze pieces of Sloan Digital Sky Survey (SDSS) images, whole SDSS images and sets of SDSS images on NERSC’s Edison supercomputer. These initial runs helped them refine and improve the model and validate its ability to exceed the performance of current state-of-the-art methods for locating celestial bodies and measuring their colors. The next milestone will be to run an analysis of the entire SDSS dataset all at once at NERSC. The researchers will then begin adding other datasets and building the catalog—which, like the SDSS data, will be housed on a science gateway at NERSC. In all, the Celeste team expects the catalog to collect and process some 500TB of data, or about 1 trillion pixels.
Principal Investigator:
Prabhat, Lawrence Berkeley National
Laboratory
MANTISSA: Massive Acceleration of New Techniques in Science with Scalable Algorithms
ASCR—Applied Math
Publication:
J. Regier, A. Miller, J. McAuliffe, R. Adams, M. Hoffman, D. Lang, D. Schlegel, Prabhat,
“Celeste: Variational inference for a generative model of astronomical images,” July 2015,
32nd International Conference on Machine Learning, Lille, France, JMLR: W&CP Volume 37.
J. Regier, J. McAuliffe, Prabhat, “A deep generative model for astronomical images of
galaxies,” Neural Informational Processing Systems (NIPS) Workshop: Advances in
Approximate Bayesian Inference. 2015
The Celeste graphical model. Shaded
vertices represent observed random
variables. Empty vertices represent latent
random variables. Black dots represent
constants.
39
Full Story:
http://bit.ly/NERSCarCeleste
Ψ
ϒ
Φ
Ω
Λμ
θσ
ρ
Ξ
τξα
ε
ι
νη
xnbm
nb
ss
s
s
s
i
i
nb
nb
nb
nb
csrs
as ks
celestialbody
galaxyprofile
fieldimage
pixel point spreadfunction
S
I
NB
M
Supercomputers Speed Search for New Subatomic Particles
NERSC ANNUAL REPORT 201540
OBJECTIVE
After being produced in a collision, subatomic particles spontaneously decay into other particles, following one of many possible decay paths. Out of 1 billion B mesons detected in a collider, only about 20 decay through this particular process. With the discovery of the Higgs boson, the last missing piece, the Standard Model of particle physics now accounts for all known subatomic particles and correctly describes their interactions. But scientists know that the Standard Model doesn’t tell the whole story, and they are searching for evidence of physics beyond the Standard Model.
FINDINGS/ACCOMPLISHMENTS
A team of physicists in the Fermilab Lattice and MILC Collaboration has developed a new, high-precision calculation that could significantly advance the search for physics beyond the Standard Model.
RESEARCH DETAILS
The new calculation, which employs lattice quantum chromodynamics (QCD), applies to a particularly rare decay of the B meson (a subatomic particle), sometimes also called a “penguin decay” process. The team ran the calculation on several supercomputers, including NERSC’s Edison and Hopper systems.
Generating the lattices and characterizing them is a time-intensive effort, noted Doug Toussaint, professor of physics at the University of Arizona and the PI who oversees this project’s allocation at NERSC. The research team used Monte Carlo simulations to generate sample configurations of fields for QCD, then calculate numerous things by averaging the desired quantity over these sample configurations, or lattices, and use these stored lattices to calculate many different strong interaction masses and decay rates.
Principal Investigator:
Douglas Toussaint, University of Arizona
Quantum Chromodynamics with Four Flavors of Dynamical Quarks
HEP—Lattice Gauge Theory
Publication:
J.A. Bailey, A. Bazavovo, C. Bernard,
C.M. Bouchard, et al, “B→πll Form
Factors for New Physics Searches from
Lattice QCD,” Physics Review Letters
115, 152002, October 7, 2015, doi:
http://dx.doi.org/10.1103/
PhysRevLett.115.152002
Full Story:
http://bit.ly/NERSCarPenguinDecay
Artist's rendering of a rare B-meson
“penguin” showing the quark-level decay
process. Image: Daping Du, Syracuse
University
Science Highlights
What Causes Electron Heat Loss in Fusion Plasma?
OBJECTIVE
Creating controlled fusion energy entails many challenges, but one of the most basic is heating plasma to extremely high temperatures and then maintaining those temperatures. Researchers at Princeton Plasma Physics Laboratory (PPPL) proposed an explanation for why the hot plasma within fusion facilities called tokamaks sometimes fails to reach the required temperature, even as beams of fast-moving neutral atoms are pumped into the plasma in an effort to make it hotter.
FINDINGS/ACCOMPLISHMENTS
3D simulations run at NERSC yielded new insights into why plasma in a tokamak fusion reactor sometimes fails to reach required temperatures. The findings enhance our understanding of electron energy transport in fusion experiments and could lead to improved temperature control in fusion devices such as ITER, the international fusion facility currently under construction in France.
RESEARCH DETAILS
The PPPL team focused on the puzzling tendency of electron heat to leak from the core of the plasma to the plasma’s edge. While performing 3D simulations of past NSTX (National Spherical Tokamak Experiment) plasmas on computers at NERSC, they saw that two kinds of waves found in fusion plasmas appear to form a chain that transfers the neutral-beam energy from the core of the plasma to the edge, where the heat dissipates. While physicists have long known that the coupling between the two kinds of waves –known as compressional Alfvén waves and kinetic Alfvén waves–can lead to energy dissipation in plasmas, the PPPL team’s findings were the first to demonstrate the process for beam-excited compressional Alfvén eigenmodes in tokamaks.
Principal Investigators:
Ronald Davidson, Elena Belova,
Princeton Plasma Physics Laboratory
Simulations of Field-Reversed Configuration and Other Compact Tori Plasmas
FES—Fusion Base Program
Publication:
E.V. Belova, N.N. Gorelenkov, E.D.
Fredrickson, K. Tritz, N.A. Crocker,
“Coupling of neutral-beam-driven
compressional Alfven Eigenmodes to
kinetic Alfven waves in NSTX tokamak
and energy channeling,” Physics Review
Letters 115, 015001, June 29, 2015, doi:
http://dx.doi.org/10.1103/
PhysRevLett.115.015001
Full Story:
http://bit.ly/NERSCarFusionPlasma
NSTXU interior vacuum vessel. Image:
Elle Starkman/ PPPL Communications
41
Documenting CO2’s Increasing Greenhouse Effect at Earth’s Surface
NERSC ANNUAL REPORT 201542
OBJECTIVE
The influence of atmospheric CO2 on the balance between incoming energy from the Sun and outgoing heat from the Earth is well established. But this effect had not been experimentally confirmed outside the laboratory.
FINDINGS/ACCOMPLISHMENTS
By measuring atmospheric CO2’s increasing capacity to absorb thermal radiation emitted from the Earth’s surface over an 11-year period at two locations in North America, researchers for the first time observed an increase in CO2’s greenhouse effect at the Earth’s surface. The team—which included researchers from Berkeley Lab, the University of Wisconsin-Madison and PNNL—attributed this upward trend to rising CO2 levels from fossil fuel emissions. The measurements also enabled the scientists to detect, for the first time, the influence of photosynthesis on the balance of energy at the surface. They found that CO2-attributed radiative forcing dipped in the spring as flourishing photosynthetic activity pulled more of the greenhouse gas from the air.
RESEARCH DETAILS
The research collaboration measured atmospheric CO2’s contribution to radiative forcing at two sites, one in Oklahoma and one on the North Slope of Alaska, from 2000 to the end of 2010. Using NERSC’s Hopper and Edison systems and the Community Earth System Model, the team used about 5,000 CPU hours to perform tens of thousands of calculations to analyze the data. They ran a large number of radiative transfer calculations that turned out to be very parallel, so they then used the taskfarmer utility for some 500 calculations at a time, with each lasting about 10 minutes on a single core.
Principal Investigator:
William Collins, Lawrence Berkeley
National Laboratory
Center at LBNL for Integrative Modeling of the Earth System (CLIMES)
BER—Climate and Environmental Sciences
Publication:
D.R. Feldman, W.D. Collins, P.J. Gero,
M.S. Torn, E.J. Mlawer, T.R. Shippert,
“Observational determination of surface
radiative forcing by CO2 from 2000 to
2010,” Nature, February 25, 2015, doi:
10.1038/nature14240
Full Story:
http://bit.ly/NERSCarGreenhouseEffect
Data for this study was collected
using spectroscopic instruments at two
sites operated by DOE’s Atmospheric
Radiation Measurement Climate
Research Facility, including this
research site on the North Slope of
Alaska near the town of Barrow.
Image: Jonathan Gero
Science Highlights
What Ignites a Neutron Star?
OBJECTIVE
To use multi-dimensional simulations to gain new insights into the physics inside a neutron star and the behavior of dense nuclear matter by studying convection in Type I X-ray bursts—the thermonuclear “runaway” in a thin hydrogen/helium layer on the star’s surface—thereby enhancing our understanding of how dense nuclear matter and neutron stars behave.
FINDINGS/ACCOMPLISHMENTS
Astrophysicists from Stony Brook University, Los Alamos National Laboratory and Berkeley Lab used supercomputing resources at NERSC to set the stage for the first detailed 3D simulations of convective burning in an X-ray burst. The research builds on the team’s previous 2D X-ray burst studies at NERSC using MAESTRO, a hydrodynamics code developed at Berkeley Lab. One-dimensional hydrodynamic studies have been able to reproduce many of the observable features of X-ray bursts, such as burst energies and recurrence times. But multi-dimensional simulations are necessary to more fully understand the convection process in the extreme conditions found in a neutron star.
RESEARCH DETAILS
The research team used 5 million CPU hours on NERSC’s Edison system as an initial testbed to explore multiple models and parameter space. They then used this information to run 3D simulations of the X-ray bursts at the Oak Ridge Leadership Computing Facility, comparing the simulations to previously created 2D simulations. Using the MAESTRO code, they modeled a small box on the surface of a neutron star, not the entire star; specifically, they modeled tens of meters on a side and 10 meters of depth on the surface of the star, which was enough to resolve features only a few centimeters in length, to see the burning and to see that it drives the convection. As a follow-on, the team is running a new class of simple model problems at NERSC that uses a more realistic initial model and reaction network.
Principal Investigator:
Michael Zingale, Stony Brook University
Three-dimensional Studies of Convection in X-ray Bursts
NP—Astrophysics
Publication:
M. Zingale, C.M. Malone, A. Nonaka,
A.S. Almgren, J.B. Bell, “Comparisons of
two- and three-dimensional convection
in Type I x-ray bursts,” Astrophysical
Journal 807:60, July 1, 2015, doi:
10.1088/0004-637X/807/1/60
Full Story:
http://bit.ly/NERSCarXrayBursts
Volume renderings of the vertical velocity
field at t=0.01 s (top) and 0.02 s (bottom) for
the wide calculation. Upward moving fluid is
in red and downward moving is blue. Image:
Michael Zingale, Stony Brook University
43
‘Data Deluge’ Pushes MSI Analysis to New Heights
NERSC ANNUAL REPORT 201544
OBJECTIVE
Researchers are combining mathematical theory and scalable algorithms with computational resources at NERSC to address growing data-management challenges in climate research, high-energy physics and the life sciences.
FINDINGS/ACCOMPLISHMENTS
To develop better methods for analyzing large datasets, in 2014 Berkeley Lab launched a project dubbed MANTISSA (Massive Acceleration of New Technologies in Science with Scalable Algorithms). MANTISSA’s first order of business was to improve data analysis in mass spectrometry imaging (MSI). Datasets from MSI provide unprecedented characterization of the amount and spatial distribution of molecules in a sample, leading to detailed investigations of metabolic and microbial processes at subcellular to centimeter resolutions. But the raw datasets range from gigabytes to terabytes in size, which can be a challenge for even large computing systems to wade through efficiently.
Studies applying MANTISSA’s algorithms to MSI datasets have demonstrated how a mathematical theory such as randomized linear algebra (RLA)—commonly used in engineering, computational science and machine learning for image processing and data mining—can be combined with one of the most sophisticated imaging modalities in life sciences to enhance scientific progress.
RESEARCH DETAILS
Researchers from UC Berkeley and Berkeley Lab ran computations involving two RLA algorithms—CX and CUR—on NERSC’s Edison system using two MSI datasets: mammalian brain and lung sections. They found that using these algorithms streamlined the data analysis process and yielded results that were easier to interpret because they prioritized specific elements in the data.
Principal Investigators:
Prabhat, Lawrence Berkeley National
Laboratory; Ben Bowen, University of
California, Berkeley
MANTISSA: Massive Acceleration of New Techniques in Science with Scalable Algorithms
ASCR—Applied Math, BER—Biological Systems Science
Publication:
J. Yang, O. Rubel, Prabhat, M. Mahoney,
B.P. Bowen, “Identifying important ions
and positions in mass spectrometry
imaging data using CUR matrix
decompositions,” Analytical Chemistry,
87 (9), 4658-4666, March 31, 2015, doi:
10.1021/ac5040264
Full Story:
http://bit.ly/NERSCarMSIDatasets
Ion-intensity visualization of the 20 most
important ions in a mouse brain segment
selected by the CX/CUR algorithm. Of
the 20 ions, little redundancy is present,
pointing to the effectiveness of the CX
approach for information prioritization.
Image: Ben Bowen, Lawrence Berkeley
National Laboratory
Science Highlights
Could Material Defects Actually Improve Solar Cells?
OBJECTIVE
Scientists at the National Renewable Energy Laboratory (NREL) are using supercomputers to study what may seem paradoxical: that certain defects in silicon solar cells may actually improve their performance.
FINDINGS/ACCOMPLISHMENTS
Deep-level defects frequently hamper the efficiency of solar cells, but this theoretical research suggests that defects with properly engineered energy levels can improve carrier collection out of the cell or improve surface passivation of the absorber layer. For solar cells and photoanodes, engineered defects could possibly allow thicker, more robust carrier-selective tunneling transport layers or corrosion protection layers that might be easier to fabricate.
RESEARCH DETAILS
Researchers at NREL ran simulations to add impurities to layers adjacent to the silicon wafer in a solar cell. Namely, they introduced defects within a thin tunneling silicon dioxide layer that forms part of the “passivated contact” for carrier collection and within the aluminum oxide surface passivation layer next to the silicon (Si) cell wafer. In both cases, specific defects were identified to be beneficial. NERSC’s Hopper system was used to calculate various defect levels; the researchers ran a total of 100 calculations on Hopper, with each calculation taking approximately eight hours on 192 cores.
Principal Investigator:
Suhuai Wei, National Renewable
Energy Laboratory
Doping bottleneck, van der Waals Interaction and Novel Ordered Alloy
BES—Materials Science
Publication:
Y. Liu, P. Stradins, H. Deng, J. Luo, S.
Wei, “Suppress carrier recombination by
introducing defects: The case of Si solar
cell,” Applied Physics Letters 108,
022101, January 11, 2016.
Full Story:
http://bit.ly/NERSCarMaterialDefects
Deep-level defects frequently hamper
the efficiency of solar cells, but NREL
research suggests that defects with
properly engineered energy levels can
improve carrier collection out of the cell.
Image: Roy Kaltschmidt, Lawrence
Berkeley National Laboratory
45
NERSC ANNUAL REPORT 201546
User Support and Outreach
The Network Operations
Center (aka “The Bridge”),
where NERSC and ESnet staff
interact with users and
systems in the new Wang Hall
at Berkeley Lab.
User Support and Outreach 47
NERSC has long been well regarded for the depth and breadth of its user support, and 2015 was no exception: in our annual user survey, the Overall Satisfaction with NERSC score was tied with our highest ever score in this category. One reason for this is that NERSC consultants and account support staff are available to users around the clock via email and an online web interface. Users worldwide can access basic account support (password resets, resetting login failures) online or via the NERSC operations staff 24 x 7, 365 days a year. In 2015, NERSC users hailed from 48 U.S. states and 47 countries.
With the center’s reorganization in 2015, a cross-group set of consultants now provide the first line of support between NERSC and its active user community, which numbers about 6,000 and includes representatives from universities, national laboratories and industry. These staff are responsible for problem management and consulting, helping with user code optimization and debugging, strategic project support, web documentation and training, third-party applications and library support, running Office of Science-wide requirements reviews and coordinating NERSC’s active users group.
NESAP Begins to BloomThroughout 2015, NERSC expanded its efforts to help users prepare for the Cori Phase 2 system coming online in summer 2016. NERSC established the NERSC Exascale Scientific Applications Program (NESAP), a collaborative effort designed to give code teams and library and tool developers a unique opportunity to prepare for Cori’s manycore architecture. NERSC selected 20 projects to collaborate with NERSC, Cray and Intel and access early hardware, special training and preparation sessions with Intel and Cray staff. In addition, another 24 projects, as well as library and tool developers, are participating in NESAP via NERSC training sessions and early access to prototype and production hardware.
One of the main goals of NESAP is to facilitate a venue for DOE Office of Science application teams to work productively with Intel and Cray engineers who specialize in hardware architecture, compiler and tool design. Through the NERSC-8 Cray Center of Excellence, Cray engineers are paired with NESAP code teams to profile and address a large number of application performance issues requiring code refactoring to improve data locality (such as tiling and cache blocking), thread scaling and vectorization.
After working closely with Cray engineers for months, the NESAP code teams are invited to attend a 2.5-day optimization “dungeon session” at Intel. These sessions generally focus on deep optimization efforts related specifically to the Intel Xeon Phi hardware. Hardware, compiler and tools experts from Intel are present to work closely with the code teams. In 2015 we completed four intensive optimization dungeon sessions at Intel with an average of three code teams at each; most teams achieved optimization of 1.5x to 2.5x in code speed-up. NESAP members have already gained a tremendous amount of experience optimizing codes for the Xeon Phi architecture as well as using Cray and Intel performance tools. NERSC has passed this knowledge onto NERSC users through regular training sessions and special events such as the Xeon Phi “hack-a-thons.”
NERSC is in the process of hiring up to eight postdocs to assist in the NESAP project. To date, six postdocs have been hired and placed with NESAP teams. The first three postdocs arrived in mid 2015; the other three arrived in early 2016.
Code speedups observed during
four NESAP “dungeon sessions”
in 2015.
NESAP Dungeon Speedups
BerkeleyGW FF Kernel
BerkeleyGW GPP Kernel
XGCI Collision Kernel
Castro
WARP Current Field Gathering
WARP Current Deposition
EMGEO IDR Solver
EMGEO QMR Solver
0.5 1 1.5 2 2.5
Speedup Factor
NERSC ANNUAL REPORT 201548
NERSC Sponsors Trainings, Workshops, HackathonsNERSC holds training events and workshops aimed at improving the skills of users of NERSC resources. Events are aimed at users of all skill levels, from novice to advanced.
In 2015 NERSC held two targeted trainings for users new to NERSC resources, one in conjunction with the annual NERSC User Group meeting, the other coinciding with the beginning of the school year. We also offered training for NERSC users in particular application areas. Users of the materials science applications VASP and Quantum Espresso had a training focused on those two applications in June, while users of BerkeleyGW participated in a 2.5-day workshop in November designed to help them use that code more efficiently.
In addition, NERSC held workshops to train users on specific code-development tools. A training specialist from Allinea presented a session on using the DDT debugger and the MAP profiler, a presenter from Rogue Wave software taught users to use the TotalView debugger and Intel taught classes on VTune and other tools.
The center also offered opportunities for hands-on development, beginning in February with the NUG Hackathon, where users were able to work with real tools under the guidance of NERSC and vendor staff. Working in teams, attendees learned how to add OpenMP to an application and how to vectorize code. Inspired by the success of this event, in July NERSC held a similar event at the Computational Science Graduate Fellowship Annual Review Workshop and at the IXPUG conference in September, drawing more than 30 participants to the latter event.
During the hackathons, NERSC staff were on hand to provide optimization strategies, help attendees use coding tools and answer questions. Attendees did the work of refactoring code to make it run faster, and remote viewers were able to follow along and pose questions via a chat interface. For example, many attendees had their first look at Intel’s VTune Amplifier, a performance analysis tool targeted for users developing serial and multithreaded applications. One attendee used VTune to identify a hot loop in his code and also to determine that he is not memory-bandwidth bound. The code contained a large number of flops and instructions but was not being vectorized due to a loop dependence. Experts from NERSC and Intel helped him restructure the code so that the flop-intensive loop no longer contained the variable dependence. This change enabled the flop-heavy loop to vectorize and sped up his entire application run by 30 percent.
NAME NESAP PROJECT START DATE
Brian Friesen BoxLib May, 2015
Andrey Ovsyannikov Chombo Crunch July 2015
Taylor Barnes Quantum Espresso July 2015
Tuomas Koskela GTC January 2016
Tareq Malas EMGeo January 2016
Mathieu Lobet WARP February 2016
NERSC’s NESAP Post-Docs, as of March 2016
User Support and Outreach 49
Other NERSC training events in 2015 covered programming concepts of particular utility for Cori. Ruud van der Pas, one of the leading OpenMP experts, taught a course on OpenMP, and Cray’s Nathan Wichmann, an expert in application performance, provided guidance on optimizing for the Knights Landing architecture in a webinar. In total, NERSC hosted 14 training sessions during the year, some of which were full-day events.
NERSC also hosted a series of developer and user community workshops as part of our NESAP program. For example, in conjunction with IXPUG, we hosted a “Density Functional Theory (DFT) for Exascale” day targeted at the developer community. The meeting attracted speakers and attendees representing the top DFT codes in the world, including VASP, Quantum ESPRESSO, SIESTA, QBOX, PETOT, PARATEC, BerkeleyGW, WEST, Yambo and DG-DFT. Attendees came from all over the U.S. and Europe to attend the workshop and discuss advancements in DFT methods targeting exascale architectures, areas of commonality for collaboration and the development and use of optimized shared libraries.
In parallel with the DFT developer community meeting, NERSC hosted an accelerator modeling code developer meeting targeted at porting particle accelerator codes to Xeon Phi. The workshop had 32 attendees from LBNL, LLNL, SLAC, Fermilab, UCLA and industry.
NERSC also hosted the third annual BerkeleyGW workshop, bringing developers and approximately 50 users together for a 2.5-day intensive training session on advanced materials science simulation using BerkeleyGW and NERSC HPC resources.
IXPUG Annual Meeting a HitNERSC’s newest supercomputer, Cori, is a Cray XC system based on Intel’s Xeon Phi Knights Landing processor. Most applications currently being used at NERSC are expected to require additional code development to fully take advantage of the computational power of the new manycore Knights Landing processor. Cori Phase 2 is expected to arrive in the summer of 2016, and other DOE national laboratories, including Argonne, Los Alamos and Sandia, will also be installing Knights Landing-based systems in the next year or so.
Because of the challenges and potential of the Xeon Phi processors, representatives from research institutions that will be deploying systems using the processor formed the Intel Xeon Phi Users Group (IXPUG), an independent users group that provides a forum for exchanging ideas and information.
NERSC hosted the IXPUG annual meeting September 28 - October 2, 2015. The event drew more than 100 members of the HPC community from around the world, who spent four days preparing for the Knights Landing architecture.
The first day consisted of workshops in which there were hands-on opportunities to optimize a code kernel or use advanced MPI and OpenMP. On the second day, there were talks about the Xeon Phi roadmap, the preparations for manycore architectures being taken by many HPC facilities and
NERSC ANNUAL REPORT 201550
experiences with developing for the Xeon Phi. The third day was a workshop on tools and techniques for optimizing for the Xeon Phi, and the fourth day featured workshops on libraries and programming models for the Xeon Phi.
Application Portability Efforts ProgressHaving an extremely broad set of applications running on our systems, NERSC is very aware that our users, as well as the applications that run on our systems, also run on systems at other large-scale computing facilities, such as the Leadership Computing Facilities at Oak Ridge or Argonne, NCAR, National Science Foundation computing facilities such as NCSA and TACC, as well as at international computing centers. Thus it is important that applications run well across multiple systems and architectures. The present challenge is that there are two architectural paths on the exascale trajectory, which, while converging slowly, nevertheless present portability challenges. For example, NERSC and the ALCF are both procuring Cray-Intel-based manycore architectures while the OLCF is procuring an IBM-NVIDIA based system.
The Office of Science ASCR facilities have been collaborating and discussing best practices for porting applications to advanced architectures since March 2014. The first meeting included ALCF, OLCF and NERSC exchanging information on each other’s systems and tools. By the second meeting, participants had expanded to include the NNSA labs and vendors providing briefings on their toolsets. Finally, in the latter half of 2015, two larger workshops were held: the HPC Operational Review (with the topic of application performance portability) and an HPC Portability Workshop at the SC15 conference.
Going forward, NERSC is partnering with ALCF and OLCF to examine two real applications to assess how easy or difficult it is to achieve application performance portability across architectures. The goal will be to use the same code base on two different architectures. We expect this effort to expand in 2016 and 2017 as prototype hardware becomes available.
More than 100 people
attended the first IXPUG
meeting, held September 28 –
October 2, 2015 in Berkeley
Lab’s Wang Hall.
User Support and Outreach 51
Preliminary Findings from the HPC Operational Review
Data and Analytics SupportOPENMSI + IPYTHON/JUPYTER = INTERACTIVE COMPUTING
In 2015 NERSC provided user support to a number of project teams focused on data-intensive science. Some involve tools accessible to the broad NERSC community, while others were tight collaborations with particular projects.
Toward this end, members of NERSC’s Data and Analytics Support (DAS) group worked with Berkeley Lab’s OpenMSI team to use the Jupyter/iPython notebook interface to enable interactive computing for the OpenMSI project. The DAS team helped create functionality to enable OpenMSI
Current Practices
Applications truly running on both GPU and CPU architectures today create a code structure that allows a “plug-and-play” of architecture-specific modules
No applications currently use the exact code base and run with performance across GPU and CPU systems
Libraries that have been optimized for a particular architecture should be used when possible
Memory management should be abstracted in code so it can easily be replaced
Write modular, layered code without vendor-specific compiler optimizations
Pay careful attention to data structure layout(s), which improves data locality and can improve portability on many architectures
Emerging Practices
Use frameworks like RAJA, KOKKOS or TIDA to enable portability
Consider using domain-specific language
Use LLVM as a backend code generator/compiler/translator
Use autotuning to find best parameters for different architectures
Architecting code in a tasking model
Opportunities
Train early career staff on multi-disciplinary computational science areas (performance, applications, architecture and algorithms)
Use libraries to encapsulate architecture specific optimizations
Use OpenMP 4.x as a standard that is portable across architectures
Adopt OpenACC features into OpenMP and focus investments into one standard
Use DSLs, compilers, translators and code generators to achieve architecture tuned performance
NERSC ANNUAL REPORT 201552
users to automatically open iPython notebooks on the NERSC JupyterHub server (https://jupyter.nersc.gov) with runnable code and analyses customized by the user from the OpenMSI portal. DAS also enabled users to bring their own Python libraries and environments into the notebook via a pluggable mechanism.
Other features of the Jupyter/iPython notebook interface include:
• Customizable template notebooks
• Programmatic and graphical user interface to easily customize notebooks—for example, change analysis parameters via a query string or user interface fields—without requiring users to write specific code to do this
• Ability to customize and integrate external JupyterHub service with a science gateway
• Ability to save notebooks directly to external repositories (such as github) without requiring users to know git
• Integration of user-level data access controls with Jupyter
• Shared notebooks: Sharing code and methods between expert programmers and beginners alike
openmsi.nersc.gov jupyter.nersc.gov
Launch, run and integrate notebooks
Share analysis results• Manage data, analyses, users and resources• Customize templates
Contribute new analysis templates
OpenMSI + JupyterHub: A Simple Vision
Template analysis notebook repository User notebook repository
The Jupyter/iPython
Notebook Interface.
User Support and Outreach 53
Supporting LUX with SciDBThe Large Underground Xenon (LUX) dark matter experiment operates nearly a mile underground at the Sanford Underground Research Facility in South Dakota. The LUX instrument was built to directly detect the faint interactions from galactic dark matter in the form of weakly interacting massive particles (WIMPS). In its initial data run, LUX collected 83 million events, containing 600 million pulses, and wrote 32TB of raw data. Of this data, only 160 events survived basic analysis cuts for potential WIMP candidates. The data rate for the new dark matter run, which started in 2014, exceeds 250TB/year.
As a case study for high energy physics, NERSC’s DAS group worked with the LUX team to load the raw data collected from the instrument’s inaugural run in 2013 into NERSC’s SciDB testbed—an open source database system designed to store and analyze extremely large array-structured data. In this case, analysis of this kind of data typically requires a researcher to search through about 1 million 10MB files within a 10TB dataset. If a WIMP candidate is spotted, the researcher would save the candidate to another file—a process that is slow and cumbersome, taking at least a day and involving analysis steps that are difficult to share. But the DAS group found that by using the SciDB testbed, the same search took just 1.5 minutes from start to finish. Since the SciDB testbed was implemented at NERSC in 2013, there has been so much interest in the use cases that the center deployed a dedicated set of nodes for this kind of analysis. These nodes will be one of the main analysis platforms for the LUX experiment.
Burst Buffer Early User Program Moves ForwardIn August 2015, NERSC put out a Burst Buffer Early User Program call for proposals, asking NERSC’s nearly 6,000 users to describe use cases and workflows that could benefit from accelerated I/O. NERSC received over 30 responses from the user community. These proposals spanned the DOE Office of Science and represented a breadth of use cases. NERSC staff evaluated each proposal using the following criteria:
• Representation among all six DOE Office of Science programs
• Ability for application to produce scientific advancements
A view inside the LUX detector.
Photomultiplier tubes can pick
up the tiniest bursts of lights
when a particle interacts with
xenon atoms. Image: Matthew
Kapust, Sanford Underground
Research Facility
NERSC ANNUAL REPORT 201554
• Ability for application to benefit significantly from the burst buffer
• Resources available from the application team to match NERSC/vendor resources
NERSC originally intended to select five teams to partner with on early burst buffer use cases. However, because of the large response, NERSC chose to support 13 application teams, meaning that each application team would have a point of contact for support from NERSC’s staff. In addition, NERSC chose to give an additional 16 teams early access to the burst buffer hardware, but without dedicated support from a NERSC staff member.
Initially, the burst buffer was conceived for the checkpoint and restart use case; however, with NERSC’s broad workload, the use cases turned out to be much more diverse, ranging from coupling together complex workflows and staging intermediate files to database applications requiring a large number of I/O operations. Ongoing bi-weekly meetings are held to update users on the latest burst buffer software patches, to offer example use cases and best practices and to allow users to share their experiences.
The Burst Buffer Early User Program Early Access Projects
PROJECT DOE OFFICE BB DATA FEATURES
Nyx/Boxlib cosmology simulations (Ann Almgren, LBNL)
HEP I/O bandwidth with BB; checkpointing; workflow application coupling; in situ analysis
Phoenix: 3D atmosphere simulator for supernovae (Eddie Baron, U. Oklahoma)
HEP I/O bandwidth with BB; staging intermediate files; workflow application coupling; checkpointing
Chombo-Crunch + VisIt for carbon sequestration (David Trebotich, LBNL)
BES I/O bandwidth with BB; in situ analysis/visualization using BB; workflow application coupling
Sigma/UniFam/Sipros Bioinformatics codes (Chongle Pan, ORNL)
BER Staging intermediate files; high IOPs; checkpointing; fast reads
XGC1 for plasma simulation (Scott Klasky, ORNL)
FES I/O bandwidth with BB; intermediate file I/O; checkpointing
PSANA for LCLS (Amadeo Perazzo, SLAC) BES/BER Staging data with BB; workflow management; in-transit analysis
ALICE data analysis (Jeff Porter, LBNL) NP I/O bandwidth with BB; read-intensive I/O
Tractor: cosmological data analysis (DESI) (Peter Nugent, LBNL)
HEP Intermediate file I/O using BB; high IOPs
VPIC-IO performance (Suren Byna, LBNL) HEP/ACSR I/O bandwidth with BB; in situ data analysis; BB to stage data
YODA: Geant4 sims for ATLAS detector (Vakhtang Tsulaia, LBNL)
HEP BB for high IOPs; stage small intermediate files
ALS SPOT Suite (Craig Tull, LBNL) BES/BER BB as fast cache; workflow management; visualization
TomoPy for ALS image reconstruction (Craig Tull, LBNL)
BES/BER I/O throughput with BB; workflow management; read-intensive I/O
kitware: VPIC/Catalyst/ParaView (Berk Geveci, kitware)
ASCR in situ analysis/visualization with BB; multi-stage workflow
User Support and Outreach 55
NERSC has identified a broad range of applications and workflows with a variety of use cases for the burst buffer, beyond the checkpoint-restart use case. NERSC burst buffer early users are now enabled on the Cori system and beginning to test their workflows on the burst buffer. In 2016 we will be examining application performance and comparisons to the file system.
APPLICATION I/O BANDWIDTH: READS
I/O BANDWIDTH: WRITES (CHECK- POINTING)
HIGH IOPS WORKFLOW COUPLING
IN SITU/ IN-TRANSIT ANALYSIS AND VISUALIZATION
STAGING INTERMEDIATE FILES/PRE-LOADING DATA
Nyx/Boxlib X X X
Phoenix 3D X X X
Chomo/Crunch + Visit
X X X
Sigma/UniFam/Sipros
X X X X
XGC1 X X X
PSANA X X X
ALICE X
Tractor X X X
VPIC/IO X X
YODA X X
ALS SPOT/TomoPy
X X X X
kitware X X
Image processing in cryo-microscopy/structural biology, Sam Li, UCSF (BER)
Htslib for bioinformatics, Joel Martin, LBNL (BER)
Falcon genome assembler, William Andreopoulos, LBNL (BER)
Ray/HipMer genome assembly, Rob Egan, LBNL (BER)
HipMer, Steven Hofmeyr, LBNL (BER/ASCR)
CESM Earth System model, John Dennis, UCAR (BER)
ACME/UV-CDAT for climate, Dean N. Williams, LLNL (BER)
Global View Resilience with AMR, neutron transport, molecular dynamics, Andrew Chien, University of Chicago (ASCR/BES/BER)
XRootD for Open Science Grid, Frank Wuerthwein, UCSD (HEP/ASCR)
OpenSpeedShop/component-based tool framework, Jim Galarowicz, Krell Institute (ASCR)
DL-POLY for material science, Eva Zarkadoula, ORNL (BES)
CP2K for geoscience/physical chemistry, Chris Mundy, PNNL (BES)
ATLAS simulation of ITK with Geant4, Swagato Banerjee, University of Louisville (HEP)
ATLAS data analysis, Steve Farrell, LBNL (HEP)
Spark, Costin Iancu, LBNL (ASCR)
Additional Burst Buffer Early Access Projects
Use Cases for 12 Burst Buffer Early User Program Projects
NERSC ANNUAL REPORT 201556
Operational ExcellenceHPSS DISK CACHE GETS AN UPGRADE
NERSC users today are benefiting from a business decision made three years ago to upgrade the High Performance Storage System (HPSS) disk cache. For archiving purposes, NERSC uses HPSS, which implements hierarchical storage management, comprising a disk farm front end (referred to as a disk cache) for short-term storage and a tape back end for long-term storage. When new files land on the disk cache they are immediately copied to tape so that they exist in both places. When the disk cache starts to get full, an algorithm selects and removes files that have already been migrated to tape.
An I/O analysis conducted by NERSC’s Storage Systems Group in 2013 changed the way HPSS would be architected going forward. The group found that the vast majority of the data that comes into NERSC is accessed within the first 90 days. Previously five days of peak I/O was NERSC’s metric for sizing its disk cache.
With the new insights into when users most often access their data, NERSC chose to go in a new direction with its hardware choices for the cache upgrade, which was rolled out in early 2015. Rather than buying SAS drives or fiber channel drives designed to support high performance bandwidth needs, they opted to go with a more capacity-oriented technology involving Nearline-SAS drives, providing 2PB of disk cache for the HPSS Archive system.
With the new arrays, the HPSS disk cache can now retain 30 days’ worth of data—roughly 1.5PB—thus reducing I/O bottlenecks and improving users’ productivity. Another key benefit is improved performance. There has also been a decrease in consult tickets since deploying the new disk cache.
STREAMLINING DATA AND ANALYTICS SOFTWARE SUPPORTIn 2015, staff from NERSC, in collaboration with LBNL’s Computational Research Division, organized a series of working groups to streamline the data and analytics portfolio supported by NERSC. NERSC has been deploying a range of data capabilities, and there has been a strong need to gain a better understanding of the rapidly evolving data technology landscape. The objective of these working groups was to identify, evaluate and recommend technologies for official adoption and deployment on NERSC systems.
Working groups were convened in the following areas:
• Workflows
• Data management (databases and storage formats)
• Data analytics
The sessions consisted of an initial inventory-gathering phase that included a broad survey of relevant technologies in each specific area. Based on the initial survey, the working groups organized deep-dive sessions, with presentations from multiple vendors, software developers and users of a given technology. Finally, the group members did some additional evaluation and offered further recommendations for the data technology stack to be deployed and supported at NERSC.
The High Performance Storage
System, which has been used
for archival storage at NERSC
since 1998, received a major
upgrade in 2015. Image:
Roy Kaltschmidt, Lawrence
Berkeley National Laboratory
User Support and Outreach 57
The conclusions of these working groups, along with existing user needs, resulted in an officially supported set of data technologies for science—NERSC’s Big Data Software Portfolio. These technologies provide a deployment vehicle for several research projects and a path for productive engagement with data science collaborators.
THE BIG MOVE: MINIMIZING DISRUPTIONS, IMPROVING OPERATIONS
NERSC’s move to Wang Hall was a complicated process, and part of the planning involved ensuring that there would be minimal disruption to users. One example of this was staggering the shutdown and move of the Cray systems, ensuring that the Cori Phase 1 system was in place and available for computation prior to shutting down Hopper and Edison.
In terms of Hopper, while computation on it was stopped in December, the login nodes and data remained available to users for several weeks after decommissioning the system. Edison’s move from the Oakland Scientific Facility to Wang Hall began on November 30, 2015; the system came back to production on January 4, 2016, almost a week earlier than anticipated. NERSC staff worked over the holiday break to bring the system back up for users. The NERSC global file systems suffered no downtime during the move as data were synced over a 400G network between the Oakland and Berkeley facilities. Users only experienced lower performing file systems for a few days.
NERSC’s move to the new building also provided opportunities for operational improvements. For example, NERSC’s Operations Technology Group includes matrixed staff funded by ESnet to support the Network Operations Center (NOC) for ESnet. With the transition of the NOC to Wang Hall, the closer proximity of NERSC and ESnet staff has created new opportunities for technology transfer and collaboration between the two divisions. One example is a new NOC Lecture Series, a 90-minute session held once a month on wide-area networking topics that provides a venue for ESnet and NERSC engineers to update operational knowledge on current implementations, new software and new configurations. The series has provided a systematic method for operations staff to improve procedures and diagnostic capabilities and to learn how to use new tools. These sessions also give NERSC staff expanded opportunities to learn about work in another facility.
CAPABILITIES TECHNOLOGY AREAS TOOLS, LIBRARIES
Data Transfer + Access Globus, Grid Stack, Authentication
Globus Online, Grid FTP
Portals, Gateways, RESTful APIs
Drupal/Django, NEWT
Data Processing Workflows Swift, Fireworks
Data Management Formats, Models Databases
HDF5, NetCDF, ROOT MongoDB, SciDB, PostgreSQL, MySQL
Storage, Archiving Lustre/GPFS, HPSS, SRM
Data Analytics Statistics, Machine Learning
python, R, ROOT BDAS/Spark
Imaging OMERO, Fiji, MATLAB
Data Visualization SCiVis InfoVis VisIt, Paraview Tableau
NERSC’s Big Data Software Portfolio
NERSC ANNUAL REPORT 201558
Center News
Shyh Wang Hall features a
number of conference rooms
and “just in time” meeting
rooms for private discussions
and gatherings–many
with spectacular views
of UC Berkeley and the
San Francisco Bay.
Center News 59
NERSC’s New Home: Shyh Wang HallOne of the biggest things to happen at NERSC in 2015 was the center’s move back to the main Berkeley Lab campus and into Shyh Wang Hall, a brand new, state-of-the-art facility for computational science. Wang Hall is named in honor of Shyh Wang, a professor at UC Berkeley for 34 years who was known for his research in semiconductors, magnetic resonances and semiconductor lasers, which laid the foundation for optoelectronics.
The $143 million structure, financed by the University of California, provides an open, collaborative environment bringing together nearly 300 staff members from NERSC, ESnet and the Computational Research Division, plus colleagues from nearby UC Berkeley, to encourage new ideas and approaches to solving some of the nation’s biggest scientific challenges.
The 149,000 square foot facility, built on a hillside overlooking the UC Berkeley campus and San Francisco Bay, houses one of the most energy-efficient computing centers anywhere, tapping into the region’s mild climate to cool NERSC’s supercomputers and eliminating the need for mechanical cooling. The building features large, open bays on the lowest level, facing west toward the Pacific Ocean, that draw in natural air conditioning for the computing systems. In addition, heat captured from those systems is recirculated to heat the building.
The new building was also designed to be flexible to meet the needs of next-generation HPC systems. The building has 12.5 megawatts of power (upgradable to over 40 megawatts) and is
Wang Hall is named in honor
of Shyh Wang, a professor
at UC Berkeley for 34 years
who was known for his
research in semiconductors,
magnetic resonances and
semiconductor lasers.
NERSC ANNUAL REPORT 201560
designed to operate at a PUE (power usage effectiveness) of less than 1.1. The machine room measures 20,000 square feet (expandable to 30,000) and features a novel seismic isolation floor to protect the center’s many resources in the event of an earthquake.
Wang Hall’s innovate design can be seen throughout the building. For example, in addition to the views, the west-facing windows were carefully placed to ensure both optimum light and energy efficiency. The architects also made sure that areas near the windows on all sides of the building offer open space and casual drop-in seating so that all staff can enjoy the space and be inspired by the views.
The building’s design is also intended to promote and encourage staff interaction across the three Computing Sciences divisions that now call Wang Hall home: NERSC, ESnet and CRD. Toward this end, there’s a mix of space in the top two floors of the building: offices, cubicles, open seating areas, “bullpens,” just-in-time meeting spaces and conference rooms of all sizes. The top two floors also feature breakrooms with floor-to-ceiling windows looking out at San Francisco and the Bay Area, an open kitchen and comfortable seating.
Because of the dramatic views and open environment, Wang Hall has already become a popular venue for meetings and conferences for groups within and outside of Computing Sciences, with additional opportunities for computer-room tours and “photo ops” with NERSC’s newest supercomputer, Cori.
Changing of the Guard: Hopper and Carver Retire, Cori Moves InAnother big change in 2015 was the retirement of two of NERSC’s legacy systems, Carver and Hopper.
Carver, which was decommissioned on September 30, was named in honor of American scientist George Washington Carver. The IBM iDataPlex cluster featured nodes with relatively high memory per core, a batch system that supported long running jobs and a rich set of third-party software applications. Carver also served as the interactive host for various experimental testbeds, such the Dirac GPU cluster and Hadoop.
Hopper, a Cray XE6 system, was retired on December 15. Named in honor of computing pioneer Admiral Grace Hopper, who developed the first compiler and originated the concept of machine-independent programming languages, Hopper was NERSC’s first petaflop system, with a peak
The symbolic connection of
the 400Gbps optical network
at the Wang Hall dedication
ceremony in November 2015.
Left to right: DOE’s Barbara
Helland, UC President Janet
Napolitano, former Berkeley
Lab director Paul Alivisatos,
Dila Wang (wife of Shyh Wang),
Berkeley Lab Associate Director
Kathy Yelick, UC Berkeley
Chancellor Nicholas Dirks,
Ipitek CEO Michael Salour and
Berkeley Lab Deputy Director
Horst Simon. Image: Kelly Owen,
Lawrence Berkeley National
Laboratory
Center News 61
performance of 1.28 Petaflops/sec, 153,216 compute cores, 212TB of memory and 2PB of online disk storage. In fact, Hopper was the fifth fastest supercomputer in the world in 2010, according to the TOP500 list of the world’s supercomputers.
The phasing out of Hopper and Carver was done in conjunction with the deployment of the first phase of Cori in Wang Hall and the move of Edison from the Oakland Scientific Facility to Berkeley Lab. Cori Phase 1 is being configured as a data partition and will be joined with the larger Cori Phase 2 system arriving in the fall of 2016. The Cori data partition has a number of unique features designed to support data-intensive workloads, including a real-time queue; a burst buffer that provides high bandwidth, low latency I/O; a large amount of memory per node (128GB/node); several login/interactive nodes to support applications with advanced workflows; and high throughput and serial queues.
A key element of Cori Phase 1 is Cray’s new DataWarp burst buffer technology, which accelerates application I/O and addresses the growing performance gap between compute resources and disk-based storage. The burst buffer is a layer of NVRAM designed to move data more quickly between processor memory and disk and allow users to make the most efficient use of the system. To ensure that they can take full advantage of DataWarp’s capabilities, NERSC also launched the Burst Buffer Early User Program, which enables us to work closely with a set of users who have diverse use cases for the burst buffer.
APEX: A Joint Effort for Designing Next-Generation HPC Systems
2015 also saw the formation of the Alliance for Application
Performance at Extreme Scale (APEX), a collaboration between Lawrence Berkeley, Los Alamos and Sandia national laboratories that is focused on the design, acquisition and deployment of future advanced technology high performance computing systems.
NERSC and ACES first began collaborating in 2010 with the procurement and deployment of two Cray XE-6 systems: Hopper at NERSC and Cielo at LANL. The partnership was formalized and strengthened with the Trinity/NERSC-8 project for the procurement of the Trinity system at LANL and the Cori system at NERSC.
NERSC ANNUAL REPORT 201562
APEX is designed to build on this partnership. Over the years, the labs have independently deployed world-leading supercomputers to support their respective scientific missions. In joining together, they aim to work even more closely with vendors to shape the future of supercomputer design to deliver ever-more capable systems to solve problems of national importance.
Initially, APEX will focus on future computing technologies via two new advanced computing systems: Crossroads for the New Mexico Alliance for Computing at Extreme Scale (ACES, a Los Alamos and Sandia partnership), and NERSC-9, NERSC’s next-generation system. In addition to providing more computing capability to NERSC users, the NERSC-9 system will advance new technologies on the path to exascale and demonstrate unique features to support data-intensive workloads.
APEX will also allow the three labs to pursue areas of collaboration with vendors, providing substantial value to the Crossroads and NERSC-9 systems in the areas of increased application performance, increased workflow efficiency and enhanced reliability of the systems. The APEX partners have released a workflows document that describes the way in which users perform their science end-to-end, including their data motion needs and accesses. A workload analysis and the workflows document can be found at http://www.nersc.gov/research-and-development/apex/apex-benchmarks-and-workflows/.
Industry Partnership Program ExpandsNERSC has been working with industry partners through the DOE’s Small Business Innovation Research (SBIR) grant program for many years. Since 2011, industry researchers have been awarded 40 million hours at NERSC via the SBIR program, and NERSC currently has about 130 industry users from 50 companies. In 2015, through its Private Sector Partnership program, the center added nine new projects in energy-specific fields ranging from semiconductor manufacturing to geothermal modeling:
Among the nine new partners is Melior Innovations, which is developing and optimizing software for accurate modeling of geothermal power plants. Subsurface networks containing thousands of fractures present complex dynamic flow and diffusion problems that must be solved for thousands of time steps. Melior Innovations has been collaborating with the open source FEniCS Project to develop simulation capabilities for fluid flow problems in large geophysical discrete fracture networks. Using NERSC’S Edison system, Melior has already improved the scalability of a solver code for geothermal energy so problems can be studied in greater detail.
Here’s a look at how some of NERSC’s other industry partners are using the center’s super- computing resources:
• Electric utilities are keenly interested in evaluating the effects of climate change and extreme weather on wind energy. Toward this end, Vertum Partners—a company specializing in applying atmospheric simulations to renewable energies, operational weather forecasts and
COMPANY R&D COMMUNIITY AWARD TYPE
Chemical Semantics Inc. Private sector materials science SBIR
Civil Maps Monitoring of energy infrastructure DDR*
Cymer Inc. Improving EUV Lithography Sources with HPC CRADA
Exabyte Web based HPC for materials science BES
Global Foundries Semiconductor device design, chip fab DDR
Kinectrics Metal alloys for pipe networks DDR
Melior Innovations Geothermal reservoir modeling DDR
Nano Precision Products Optical interconnect engineering SBIR
Quanturmscape Solid-state batteries for electric cars DDR
*Directors Discretionary Reserve
Center News 63
long-term climate dataset construction—has been working with NERSC to create 4D next-generation mesoscale numerical weather simulations that model complex atmospheric data resolved to local wind farm scales. Using WRF, a code designed to solve 4D models of the atmosphere, on NERSC supercomputers, Vertum was able to evaluate the sensitivity of the model results to different land surface types, atmospheric datasets and other forcings both internal and external to the model. This will allow for optimal turbine site selection and evaluating the true climate change effects to wind energy.
• DynaFlow, a provider of research and development services and products in fluid dynamics and material sciences, used a NERSC allocation to run a numerical model that can accurately capture flow dynamics of microbubbles that can help and/or hinder industrial machinery. The findings, published in the Journal of Fluids Engineering (April 2015) and Chemical Engineering Science (May 2015), verified the importance of “bubbly flow” in many industrial applications; for example, it can cause cavitation leading to efficiency loss in turbomachinery, but it can also be exploited to enhance cutting, drilling, cleaning and chemical manufacturing. In particular, the simulations run at NERSC showed that bubble cloud characteristics and initial pressure could be “tuned” to generate very high cavitation forces that could be used for beneficial purposes.
• GlobalFoundries, one of the largest semiconductor foundries in the world, provides a unique combination of advanced technology and manufacturing to more than 250 customers. In 2015, the company began using NERSC supercomputers to improve product development time and costs and facilitate the discovery of new materials for next-generation low-power, high-performance semiconductors through atomic scale simulations. GlobalFoundries is using these computational findings to guide development of its new 22-nm FDSOI chips.
• Civil Maps, which uses artificial intelligence and HPC to create comprehensive 3D maps, is using LIDAR monitoring data to help utility companies achieve more reliable power grids. LIDAR monitoring of power grid assets is crucial to the reliably engineered operation of distributed energy assets. For example, Pacific Gas & Electric Co. in California is using a combination of aerial surveys and mobile LIDAR mapping to monitor the growth of trees near power lines, eliminating the need for tedious land surveying using handheld GPS receivers. Geographic information specialists can then use visualizations of the data to explore the infrastructure on a computer. However, the raw, unorganized 3D LIDAR data gathered from the mobile sensors is very large in size. Civil Maps is tapping into NERSC’s resources to test and optimize algorithms and process the enormous data sets generated by LiDAR scans.
On a broader scale, NERSC is participating in DOE’s new High Performance Computing for Manufacturing Program (HPC4Mfg), announced in September 2015. HPC4Mfg couples U.S. manufacturers with world-class computational research and development expertise at Lawrence Berkeley, Lawrence Livermore and Oak Ridge national laboratories to fund and foster public-private R&D projects that enhance U.S. competitiveness in clean energy manufacturing and address key challenges in U.S. manufacturing.
In February 2016, DOE announced $3 million for the first 10 HPC4Mfg projects. Each of the projects will receive approximately $300,000 to fund the national labs to partner closely with each chosen company to provide expertise and access to HPC systems aimed at high-impact challenges. GlobalFoundries was one of the projects selected; the company will collaborate with Berkeley Lab and NERSC to optimize the design of transistors under a project entitled “Computation Design and Optimization of Ultra-Low Power Device Architectures.”
NERSC ANNUAL REPORT 201564
‘Superfacilities’ Concept Showcases Interfacility CollaborationsThroughout 2015 NERSC saw increased interest
in the coupled operation of experimental and computational facilities throughout the DOE Office of Science complex—the so-called “superfacility.”
A superfacility is two or more facilities interconnected by means of a high performance network specifically engineered for science (such as ESnet) and linked together using workflow and data management software such that the scientific output of the connected facilities is greater than it otherwise could be. This represents a new
experimental paradigm that allows HPC resources to be brought to bear on experimental work being done at, for example, a light source beamline within the context of the experimental run itself.
Adding HPC to facility workflows is a growing trend, and NERSC is leveraging its experience with the Joint Genome Institute, the Advanced Light Source, the Large Hadron Collider, the Palomar Transient Factory, AmeriFlux and other experimental facilities to shape the interfaces, schedules and resourcing models that best fit what our facility users need. Treating facilities as users of HPC requires long-term data planning, real-time resource scheduling and pipelining of workflows.
NERSC’s superfacility strategy focuses on scalable data analysis services that can be prototyped and reused as a new interfacility collaborations emerge. For example, in 2015 a collaborative effort linking the Advanced Light Source (ALS) at Berkeley Lab with supercomputing resources at NERSC and Oak Ridge National Laboratory via ESnet yielded exciting results in organic photovoltaics research. In a series of experiments, the operators of a specialized printer of organic photovoltaics used HPC to observe the structure of the material as it dried. Organic photovoltaics show promise as less expensive, more flexible materials for converting sunlight to electricity.
During the experiments, scientists at the ALS simultaneously fabricated organic photovoltaics using a miniature slot-die printer and acquired beamline data of the samples as they were made. This data was transferred to NERSC via ESnet, where it was put into the SPOT Suite data portal and underwent analysis to translate the image pixels into the correct reciprocal space. Using the Globus data management tool developed by the University of Chicago and Argonne National Laboratory, the output from those jobs was then sent to the Oak Ridge Leadership Computing Facility for analysis, where the analysis code ran on the Titan supercomputer for six hours nonstop on 8,000 GPUs.
A unique feature of this experiment was co-scheduling beamline time at the ALS with time and resources at NERSC and Oak Ridge, plus utilizing ESnet to link it all together. This was the first time a beamline was able to simultaneously take data and perform large-scale analysis on a supercomputer.
NERSC and the 2015 Nobel Prize in PhysicsIn 2015 NERSC was once again fortunate to be part of the story behind a Nobel Prize. This time it was the 2015 Nobel Prize in Physics, awarded to the leaders of two large neutrino experiments: Takaaki Kajita of Tokyo University, who led the Super-Kamiokande experiment, and Arthur B. McDonald of Queen’s University in Kingston, Ontario, Canada, head of the Sudbury Neutrino Observatory (SNO). According to the Nobel Prize committee, Kajita and McDonald were honored “for the discovery of neutrino oscillations, which shows that neutrinos have mass.”
A GISAXS pattern from a
printed organic photovoltaic
sample as it dried in the
beamline, part of a
“superfacility” experiment
based at Berkeley Lab.
Image: Craig Tull, Lawrence
Berkeley National Laboratory
Center News 65
NERSC’s connection to this Nobel Prize is through SNO, a huge neutrino detector based in Sudbury, Ontario. The observatory is located in an old copper mine 2 km underground to shield it as much as possible from the noise background of particle interactions that take place on Earth’s surface.
The SNO detector was turned on in 1999, and from the earliest days SNO data was transferred to NERSC, where the center’s PDSF cluster was used for what became known as the “West Coast Analysis.” When the discovery of neutrino flavor mixing in solar neutrinos was published in 2001 in Physical Review Letters, NERSC’s role was well established and recognized by scientists working on the project; in fact, they presented NERSC’s PDSF team with a signed and framed copy of the journal article.
The SNO experiment wound down in 2006, but the data continues to prove invaluable. To assure its integrity and safekeeping, NERSC was chosen to house the data archive. Archiving and moving all the data to NERSC required a close collaboration between SNO, NERSC and ESnet.
When the Nobel Prize for Physics was announced in October 2015, one of the leads on the SNO project, Alan Poon of Berkeley Lab, asked that the following email be distributed to the entire NERSC staff:
SNO has been blessed by the top-notch support and facility at NERSC. Without NERSC’s support, SNO would not have been able to process and reconstruct the data, simulate the data and run massive jobs for the physics fits so smoothly and successfully.
We appreciate your continual support of our data archive at HPSS. As you can imagine, we want to preserve this precious data set, and once again, NERSC has come to our assistance. Many, many thanks from all of us in SNO.
Best, Alan
To which Dr. McDonald added in a subsequent email:
May I add my thanks to those from Alan. We greatly appreciate your support.
Best regards, Art McDonald
OpenMSI Wins 2015 R&D 100 AwardOpenMSI, the most advanced computational tool for analyzing and visualizing mass spectrometry instruments (MSI) data, was one of seven Berkeley Lab winners of a 2015 R&D 100 award. Oliver Rübel of the Computational Research Division and Ben Bowen of the Environmental Genomics and Systems Biology Division led the development of OpenMSI, with collaboration from NERSC.
MSI technology enables scientists to study tissues, cell cultures and bacterial colonies in unprecedented detail at the molecular level. As the mass and spatial resolution of MS instruments increase, so do the number of pixels in MS images and data size. Nowadays, MSI datasets range from
Art McDonald at Berkeley Lab
in November 2015, where he
outlined NERSC’s and Berkeley
Lab’s contributions to the SNO
Collaboration.
NERSC ANNUAL REPORT 201566
tens of gigabytes to several terabytes. Thus, basic tasks like opening a file or plotting spectra and ion images become insurmountable challenges.
OpenMSI overcomes these obstacles by making highly optimized computing technologies available via a user-friendly interface. Because OpenMSI leverages NERSC’s resources to process, analyze, store and serve massive MSI datasets, users can now work on their data at full resolution and in real time without any special hardware or software. They can also access their data on any device with an Internet connection.
Presented by R&D Magazine, the annual R&D 100 Awards recognize the top 100 technology products from industry, academia and government-sponsored research, ranging from chemistry and materials to biomedical and information technology breakthroughs.
NERSC Staff Honored with Director’s Awards for Exceptional Achievement
Two long-time NERSC staff members—Brent Draney and Lynne Rippe—were among five Berkeley Lab Computing Sciences employees to be honored with 2015 Director’s Awards for Exceptional Achievement.
Draney, head of NERSC’s Networking, Security and Servers Group, and Eli Dart of
ESnet’s Science Engagement Team, were recognized for their work in developing the Science DMZ, a network architecture that allows science data to securely bypass institutional firewalls. The Science DMZ has been endorsed by the National Science Foundation, which has funded Science DMZs at more than 100 universities across the U.S. Draney and Dart, who was working at NERSC when the Science DMZ concept was born, were honored in the area of operations for “achievement in operational effectiveness, process re-engineering or improvement, resource management and efficiency or partnerships across organizational/departmental boundaries.”
Rippe was awarded a Berkeley Lab Citation for her longtime procurement work in support of NERSC. A lead subcontracts administrator with the Lab’s Office of the Chief Financial Officer, Rippe has been the procurement department lead on every major NERSC system dating back to the center’s days at Lawrence Livermore National Laboratory. She was also responsible for the development of the Best Value Source Selection process for large-scale computer procurements, which has been widely adopted within DOE laboratories. The Berkeley Lab Citation recognizes the highest level of service to Berkeley Lab and the network of DOE laboratories.
4th Annual HPC Achievement Awards Highlight Innovative ResearchNERSC announced the winners of the fourth annual High Performance Computing Achievement Awards in March 2016 during the annual NERSC Users Group meeting at Berkeley Lab. The awards recognize NERSC users who have either demonstrated an innovative use of HPC resources to solve a scientific problem or whose work has had an exceptional impact on scientific understanding or society. To encourage younger scientists who are using HPC in their research, NERSC also presents two early career awards.
Brent Draney (left) and Eli Dart
(center) received their Director’s
Award from Berkeley Lab
Deputy Director Horst Simon.
Lynne Rippe
Center News 67
NERSC 2016 AWARD FOR HIGH IMPACT SCIENCE: OPEN
Charles Koven and William Riley of Berkeley Lab’s Climate and Ecosystem Sciences Division and David Lawrence of the National Center for Atmospheric Research were honored for their role in using an Earth system model to demonstrate the atmospheric effect of emissions released from carbon sequestered in melting permafrost soil. Running simulations on NERSC’s Hopper system, the team demonstrated that thawing permafrost releases enormous amounts of long-frozen carbon into the atmosphere—more than carbon taken by plants.
The model is the first to represent permafrost processes as well as the dynamics of carbon and nitrogen in the soil. According to the simulations, by the year 2300—if climate change continues unchecked—the net loss of carbon to the atmosphere from Arctic permafrost would range from between 21 petagrams and 164 petagrams. That’s equivalent to between two years and 16 years of human-induced CO2 emissions.
“We found the rate of permafrost thaw and its effect on the decomposition of deep carbon will have a much bigger impact on the carbon cycle than the availability of deep nitrogen and its ability to spark plant growth,” Koven said. Their findings were reported in the Proceedings of the National Academy of Sciences in March 2015.
NERSC 2016 AWARD FOR HIGH IMPACT SCIENCE: EARLY CAREER
Nathan Howard, a research scientist at MIT’s Plasma Science and Fusion Center, is being honored in this category for his pioneering computational work in plasma turbulence simulations. In particular, Howard—who was a postdoc at the University of California, San Diego until joining MIT in 2015—carried out the most physically comprehensive simulations of tokamak plasma microturbulence to date, according to Christopher Holland, associate research scientist in the Center for Energy Research at UCSD, who nominated Howard for this award.
One roadblock in the quest for fusion is that, to date, computer models have often been unable to predict exactly how turbulence will behave inside the reactor. In fact, there have long been differences between predictions and experimental results in fusion experiments when studying how turbulence contributes to heat loss in the confined plasma. But Howard discovered a solution to this disparity: By performing high-resolution multi-scale simulations, the team was able to simultaneously resolve multiple turbulence instabilities that have previously been treated in separate simulations. A series of these multi-scale simulations run on NERSC’s Edison system found that interactions between turbulence at the tiniest scale (that of electrons) and turbulence at a scale 60 times larger (that of ions) can account for the mysterious mismatch between theoretical predictions and experimental observations of the heat loss.
These groundbreaking simulations used well over 100 million core-hours of computing time, and Howard was the top user of NERSC computer time in 2014 and in the top five for 2015. The results resolved a long-standing model-experiment discrepancy and led to the most significant advance in our understanding of tokamak turbulence in over a decade, according to Holland.
NERSC 2016 AWARD FOR INNOVATIVE USE OF HPC: OPEN
Former Berkeley Lab computational scientist Scott French was honored in this category for his role in helping seismologists create a unique 3D scan of the Earth’s interior that resolved some long-standing questions about mantle plumes and volcanic hotspots. (See science highlight on p. 31.)
The 2016 NERSC HPC
Achievement Award winners
(left to right): Charles Koven,
Nathan Howard, Scott French.
Not pictured: William Riley,
David Lawrence, Min Si. Image:
Margie Wylie, Lawrence
Berkeley National Laboratory
NERSC ANNUAL REPORT 201568
Working with Barbara Romanowicz, a professor of earth and planetary science at UC Berkeley, French—who now works at Google—ran a number of simulations at NERSC, producing for the first time a computed tomography scan that conclusively connects plumes of hot rock rising through the mantle with surface hotspots that generate volcanic island chains like Hawaii, Samoa and Iceland. Until this study, evidence for the plume and hotspot theory had been circumstantial, and some seismologists argued instead that hotspots are very shallow pools of hot rock feeding magma chambers under volcanoes.
To create the high-resolution 3D image of Earth, French ran numerical simulations of how seismic waves travel through the mantle and compared their predictions to ground motion actually measured by detectors around the globe. They mapped mantle plumes by analyzing the paths of seismic waves bouncing around Earth’s interior after 273 strong earthquakes that shook the globe over the past 20 years. The simulations, run on NERSC’s Edison system, computed all components of the seismic waves. French tweaked the model repeatedly to fit recorded data using a method similar to statistical regression. The final computation required 3 million CPU hours on Edison; parallel computing shrank this to a couple of weeks.
NERSC 2016 AWARD FOR INNOVATIVE USE OF HPC: EARLY CAREER
Min Si, a graduate student from the University of Tokyo who is working at Argonne National Laboratory, received this award for her pioneering work in developing novel system software in the context of MPI3 one-sided communication. This software has had a transformative impact on the field of computational chemistry by completely eliminating the need for custom ports of global arrays, according to Jeff Hammond, a research scientist in the Parallel Computing Lab at Intel Labs, who nominated Si for the award.
The tool Si created is Casper, a library that sits between an MPI3 application and any MPI implementation, including proprietary ones such as Cray MPI and Intel MPI. When Casper is used, the application sees an ideal implementation of MPI3: it makes excellent asynchronous progress for all operations, including noncontiguous and accumulating one-sided operations that are rarely, if ever, supported in hardware. This is achieved without the usual overheads of asynchronous progress, which include mutual exclusion and oversubscription resulting from communication helper threads. Casper is a boon to MPI3 users and implementers because it resolves the longstanding tension over asynchronous progress, which torments both groups because it previously was not possible to achieve it without tradeoffs.
Using Casper, NWChem runs as fast using MPI3 as it does using DMAPP on NERSC Cray XC systems without compromising all of the productive features of MPI, such as profiling and debugging tool support, noted Hammond, who previously worked as a computational scientist at Argonne with Si.
Organizational Changes Reflect Evolving HPC Data EnvironmentIn February 2015, NERSC announced several organizational changes to help the Center’s 6,000 users manage their data-intensive research more productively.
Notably, NERSC determined that there was enough momentum around HPC for data-intensive science that it warranted a new department focused on data: the Scientific Computing & Data Services Department, which is being led by Katie Antypas, who was also named Division Deputy for Data Science. The department comprises four groups, two of them new: Data Science Engagement (led by Kjiersten Fagnan), which engages with users of experimental and observational facilities, and Infrastructure Services, which focuses on infrastructure to support NERSC and the center’s data initiative. The existing Storage Systems Group and Data and Analytics Group complete the new department.
Center News 69
In addition to the new Data Services Department, NERSC now features an HPC Department, which comprises the Advanced Technologies, Application Performance, Computational Systems and User Engagement groups; and the Systems Department, which comprises the Network and Security Group and Operations Technology Group.
Personnel TransitionsA number of long-term NERSC veterans retired in 2015, including:
• Iwona Sakrejda: A member of the Technology Integration group at NERSC when she retired, Sakrejda was with Berkeley Lab for 24 years, 15 of those at NERSC. She earned her Ph.D. in particle physics from the Institute of Nuclear Physics in Krakow, Poland.
• Karen Schafer: Schafer, a network engineer in the Networking, Security and Servers group, joined NERSC in 2008 and retired in June 2015.
• David Turner: An HPC consultant in what was formerly known as NERSC’s User Services Group, Turner was with NERSC for 17 years.
• Francesca Verdier: Verdier joined NERSC in 1996 as part of the center’s massive hiring effort after relocating from Lawrence Livermore National Laboratory to Berkeley Lab. After 19 years at NERSC, she retired as Allocations Manager in June 2015.
• Harvey Wasserman: An HPC consultant at NERSC since 2006, Wasserman was involved in workload characterization, benchmarking and system evaluation at NERSC and at Los Alamos National Laboratory for more than 30 years.
Despite the multiple retirements and normal turnover and churn from competition with Bay Area technology companies, NERSC expanded its workforce in 2015 with 18 additional staff members, including postdocs hired specifically for the NERSC Exascale Science Application Program. Overall, NERSC currently has more than 60 FTEs, as well as administrative, communications, research and support staff.
NERSC new hires in 2015 totaled 18. Left to right: Cori Snavely, Glenn Lockwood, Taylor Barnes, Jackson Gor, Rollin Thomas, Debbie Bard,
Evan Racah, Yunjie Liu, Rebecca Hartman-Baker, Daniel Udwary, Brian Friesen and Indira Kassymkhanova. Not pictured: Wahid Bhimiji,
Ankit Bhagatwala, Doug Doerfler, Jialin Liu, Andrey Ovsyannikov and Tavia Stone
NERSC ANNUAL REPORT 201570
Allocation of NERSC Director’s Reserve of Computer Time
The NERSC director receives ten percent of the total allocation of computer time, which amounted to 300 million hours in 2015. The Director’s Reserve is used to support innovative and strategic projects as well as projects in NESAP (the NERSC Exascale Science Applications Program), a collaborative effort in which NERSC is partnering with code teams and library and tools developers to prepare for the Cori manycore architecture. A number of Director’s Discretionary Reserve projects were focused on exploring performance and scaling on emerging applications. An additional part of the Director’s Reserve in 2015 was allocated, in conjunction with OS program managers, to high-impact Office of Science mission computing projects, augmenting their base allocations.
CATEGORYNUMBER OF PROJECTS
ALLOCATED FROM DIRECTOR'S RESERVE (MILLION MPP HOURS)
Director’s Reserve Projects 45 160
Mission Computing Augmentation (in collaboration with SC program managers)
221 140
TOTAL 300
PROJECT PI ORGANIZATION ALLOCATION
APPLIED MATH
Anchored Solar-Driven Vortex for Power Generation Pearlstein, Arne U. Illinois U-C 5,800,000
NERSC Application Readiness for Future Architectures Antypas, Katie Berkeley Lab 1,000,000
High-resolution, Integrated Terrestrial and Lower Atmosphere Simulation of the Contiguous United States
Maxwell, Reed Colorado School Mines 1,000,000
COMPUTER SCIENCE
XPRESS Program Environment Testing at Scale Brightwell, Ron Sandia Labs NM 12,000,000
High Performance Data Analytics Prabhat, Mr NERSC 8,000,000
Performance Analysis, Modeling and Scaling of HPC Applications and Tools Bhatele, Abhinav Livermore Lab 3,000,000
Data Intensive Science R&D Prabhat, Mr Berkeley Lab 3,000,000
Exploring Quantum Optimizers via Classical Supercomputing Hen, Itay U. Southern California 2,000,000
Berkeley Institute for Data Sciences through HPC Perlmutter, Saul UC Berkeley 2,000,000
DEGAS: Dynamic Exascale Global Address Space Yelick, Katherine Berkeley Lab 1,100,000
Parallel FEC Verification and Scaling Prototype Bell, Greg Berkeley Lab 1,000,000
Domain-Specific Language Technology Quinlan, Daniel Livermore Lab 1,000,000
2015 Director’s Reserve Allocations
NERSC Director’s Reserve 71
PROJECT PI ORGANIZATION ALLOCATION
Gamma-Ray Data Cloud Quiter, Brian Berkeley Lab 1,000,000
Traleika Glacier X-Stack Borkar, Shekhar Intel Inc. 500,000
Point Cloud Mapping of Energy Assets and Risks Chraim, Fabien Berkeley Lab 500,000
Application Benchmarking and SSP Development Oppe, Thomas HPCMPO 500,000
Performance Profiling and Assessment of MFIX Using Trilinos Linear Solvers Versus Using Native Linear Solvers
Gel, Aytekin ALPEMI Consulting, LLC 100,000
BIOSCIENCES
Integrated Tools for Next-Generation Bio-Imaging Skinner, David Berkeley Lab 13,500,000
Multi-scale Multi-Compartment Computational Models of the Human Immune Response to Infection with M. tuberculosis
Kirschner, Denise U. Michigan 2,500,000
Collaborative Research in Computational Neuroscience - CRCNS Bouchard, Kristofer Berkeley Lab 1,500,000
Enhanced Sampling Simulations of G-protein Coupled Receptors Miao, Yinglong H. Hughes Med Inst 1,000,000
CLIMATE & ENVIRONMENT
FP-RadRes: High Performance Flux Prediction Towards Radiological Resilience Steefel, Carl Berkeley Lab 10,000,000
Next Generation Global Prediction System (NGGPS) Benchmarking Michalakes, John NOAA 4,000,000
Greening the Grid USAID/India Purkayastha, Avi Berkeley Lab 2,500,000
Database of Energy Efficiency Performance Hong, Tianzhen Berkeley Lab 2,000,000
Integrating Climate Change into Air Quality Modeling Kaduwela, Ajith UC Davis 1,000,000
A Computational Study of the Atmospheric Boundary Layer with Mesoscale-Microscale Coupling
Balakrishnan, Ramesh
Argonne Lab 1,000,000
A Multi-Decadal Reforecast Data Set to Improve Weather Forecasts for Renewable Energy Applications
Hamill, Thomas NOAA 100,000
Investigating Bioenergy with Carbon Capture and Sequestration using the California TIMES Model Projected through 2050
Jenn, Alan UC Davis 100,000
Cyclotron Road - Wave Carpet Wave Energy Converter Lehmann, Marcus Berkeley Lab 100,000
Monte Carlo Simulation of Greenhouse Gas Emissions from Biofuels-Induced Land Use Change
Plevin, Richard UC Berkeley 100,000
GEOSCIENCES
Viscoelastic Relaxation and Delayed Earthquake Triggering Along the North Anatolian Fault, the Japan Subduction Zone and the Himalayan Range Front
DeVries, Phoebe Harvard Univ 500,000
Emergent Fracture Propagation in Porous Solids Edmiston, John Berkeley Lab 250,000
Feresa: HPC Scaling of Realistic Geothermal Reservoir Models Bernstein, David Berkeley Lab 150,000
NERSC ANNUAL REPORT 201572
PROJECT PI ORGANIZATION ALLOCATION
MATERIALS SCIENCE
Exploring Dynamics at Interfaces Relevant to Mg-ion Electrochemistry from First Principles
Prendergast, David Berkeley Lab 14,500,000
Quantum Simulations of Nanoscale Energy Conversion Grossman, Jeffrey MIT 12,000,000
Next Generation Semiconductors for Low Power Electronic Devices Lee, Byounghak Berkeley Lab 12,000,000
Computing Resources for the Center for the Computational Design of Functional Layered Materials
Perdew, John Temple University 10,000,000
Hot Carrier Collection in Thin Film Silicon with Tailored Nanocrystalline/Amorphous Structure
Lusk, Mark Colorado School Mines 8,000,000
Testing the Fundamental Limits on the Number of Inorganic Compounds in Nature
Ceder, Gerbrand MIT 7,500,000
Accelerated Discovery and Design of Complex Materials Kent, Paul Oak Ridge 2,000,000
Data Mining Discovery of Thermoelectrics and Inverse Band Structure Design Principles
Jain, Anubhav Berkeley Lab 1,000,000
Topological Phases in Correlated Systems Vishwanath, Ashvin Berkeley Lab 700,000
ACCELERATOR DESIGN
Atomic-Scale Modeling of Fracture Along Grain Boundaries with He Bubbles Xu, Steven Kinectrics Inc. 1,000,000
ASTROPHYSICS
Radiation Hydrodynamic Simulations of Massive Star Envelope Jiang, YanfeiSmithsonian Astro Observatory
6,169,000
Publications 73
Publications
Staying ahead of the technological curve, anticipating problems and developing proactive solutions are part of NERSC’s culture. Many staff members collaborate on computer science research projects, scientific code development and domain-specific research. They also participate in professional organizations and conferences and contribute to journals and proceedings. The NERSC user community benefits from these activities as they are applied to systems, software and services at NERSC and throughout the HPC community. In 2015, NERSC users reported 2,078 refereed papers based on work performed at the center, reflecting the breadth of scientific activity supported by NERSC. In addition, NERSC staff contributed some 120 papers to scientific journals and conferences in 2015, showcasing our increasing involvement in technology and software development designed to enhance the utilization of HPC across a broad range of science and data-management applications. To see a comprehensive list of publications and presentations by NERSC staff in 2015 go to http://1.usa.gov/1f3uFIv.
NERSC ANNUAL REPORT 201574
Appendix A
NERSC Users Group Executive Committee
OFFICE OF ADVANCED SCIENTIFIC COMPUTING RESEARCHMilan Curcic, University of MiamiJeff Hammond, IntelBrian Van Straalen, Lawrence Berkeley National Laboratory
OFFICE OF BASIC ENERGY SCIENCESEric Bylaska, Pacific Northwest National LaboratoryMonojoy Goswami, Oak Ridge National LaboratoryPaul Kent, Oak Ridge National Laboratory
OFFICE OF BIOLOGICAL AND ENVIRONMENTAL RESEARCHSamson Hagos, Pacific Northwest National LaboratoryGary Strand, National Center for Atmospheric ResearchDavid Wong, U.S. Environmental Protection Agency
OFFICE OF FUSION ENERGY PHYSICSChristopher Holland, University of California, San DiegoNathan Howard, Massachusetts Institute of TechnologyDavid Green, Oak Ridge National Laboratory
OFFICE OF HIGH ENERGY PHYSICSTed Kisner, Lawrence Berkeley National LaboratoryZarija Lukic, Lawrence Berkeley National LaboratoryFrank Tsung, University of California, Los Angeles (Chair)
OFFICE OF NUCLEAR PHYSICSBalint Joo, Jefferson LabNicolas Schunck, Lawrence Livermore National LaboratoryMichael Zingale, Stony Brook University
MEMBERS AT LARGEJames Amundson, FermilabCarlo Benedetti, Lawrence Berkeley National LaboratoryDavid Hatch, University of Texas, Austin
Appendices 75
Appendix B
Office of Advanced Scientific Computing Research
The mission of the Advanced Scientific Computing Research (ASCR) program is to discover, develop and deploy computational and networking capabilities to analyze model, simulate and predict complex phenomena important to the Department of Energy (DOE). A particular challenge of this program is fulfilling the science potential of emerging computing systems and other novel computing architectures, which will require numerous significant modifications to today’s tools and techniques to deliver on the promise of exascale science.
To accomplish its mission and address those challenges, the ASCR program is organized into two subprograms: Mathematical, Computational and Computer Sciences Research; and High Performance Computing and Network Facilities
The Mathematical, Computational and Computer Sciences Research subprogram develops mathematical descriptions, models, methods and algorithms to describe and understand complex systems, often involving processes that span a wide range of time and/or length scales. The subprogram also develops the software to make effective use of advanced networks and computers, many of which contain thousands of multi-core processors with complicated interconnections, and to transform enormous data sets from experiments and simulations into scientific insight.
The High Performance Computing and Network Facilities subprogram delivers forefront computational and networking capabilities and contributes to the development of next-generation capabilities through support of prototypes and testbeds.
Berkeley Lab thanks the program managers with direct responsibility for the NERSC program and the research projects described in this report:
ASCR PROGRAM
Steve Binkley, Associate Director, ASCR
Michael Martin, Fellow
Julie Stambaugh, Financial Management Specialist
Lori Jernigan, Program Support Specialist
FACILITIES DIVISION
Barbara Helland, Director
Vince Dattoria, Computer Scientist, ESnet Program Manager
Betsy Riley, Computer Scientist, ALCF Program Manager
Robinson Pino, Computer Scientist, REP Program Manager
Claire Cramer, Physical Scientist, REP Program Manager
Carolyn Lauzon, Physical Scientist, ALCC Program Manager
Dave Goodwin, Physical Scientist, NERSC Program Manager
Christine Chalk, Physical Scientist, ORLC Program Manager, CSGF Program Manager
Sally McPherson, Program Assistant
RESEARCH DIVISION
William Herrod, Director
Teresa Beachley, Program Assistant
Randall Laviolette, Physical Scientist, SciDAC Application Partnerships
Thomas Ndousse-Fetter, Computer Scientist, Network Research
Ceren Susut, Physical Scientist, SC Program SAPs
Rich Carlson, Computer Scientist, Collaboratories/Middleware
Steven Lee, Physical Scientist, SciDAC Institutes
Lucy Nowell, Computer Scientist, Data and Visualization
Sonia Sachs, Computer Scientist, Extreme Scale
Angie Thevenot, Program Assistant
NERSC ANNUAL REPORT 201576
Appendix C
Acronyms and Abbreviations
ACMAssociation for Computing Machines
ACSAmerican Chemical Society
ALCCASCR Leadership Computing Challenge
ALSAdvanced Light Source, Lawrence Berkeley National Laboratory
ANLArgonne National Laboratory
APIApplication Programming Interface
APSAmerican Physical Society
ASCIIAmerican Standard Code for Information Interchange
ASCROffice of Advanced Scientific Computing Research
BELLABerkeley Lab Laser Accelerator
BEROffice of Biological and Environmental Research
BESOffice of Basic Energy Sciences
BNLBrookhaven National Laboratory
C3
Computational Cosmology Center, Lawrence Berkeley National Laboratory
CALComputer Architecture Laboratory
CARBCalifornia Air Resources Board
CCMCluster Compatibility Mode
CdTeCadmium Telluride
CeO2
Cerium Dioxide
CERNEuropean Organization for Nuclear Research
CESMCommunity Earth Systems Model
CFDComputational Fluid Dynamics
CLECray Linux Environment
CMBCosmic Microwave Background
CMIP5Coupled Model Intercomparison Project, 5th Phase
CO2
Carbon dioxide
CPUCentral Processing Unit
CRDComputational Research Division, Lawrence Berkeley National Laboratory
CRTComputational Research and Theory Facility, Lawrence Berkeley National Laboratory
CSEComputational Science and Engineering
DARPADefense Advanced Research Projects Agency
DESIDark Energy Spectroscopic Instrument
DFTDensity Functional Theory
DMEDimethyl Ether
DNSDirect Numerical Simulation
DOEU.S. Department of Energy
DOIDigital Object Identifier
DSLDynamic Shared Library
DTNData Transfer Node
DVSData Virtualization Service
EFRCDOE Energy Frontier Research Center
EMSLEnvironmental Molecular Science Laboratory, Pacific Northwest National Laboratory
EPSISciDAC Center for Edge Physics Simulations
ERDEarth Sciences Division, Lawrence Berkeley National Laboratory
ERTEmpirical Roofline Toolkit
ESnetEnergy Sciences Network
eVElectron Volts
FDMFinite Difference Method
FESOffice of Fusion Energy Sciences
FLOPSFloating Point Operations
FTPFile Transfer Protocol
GBGigabytes
GbpsGigabits Per Second
GPUGraphics Processing Unit
HEPOffice of High Energy Physics
HPCHigh Performance Computing
HPSSHigh Performance Storage System
HTMLHypertext Markup Language
HTTPHypertext Transfer Protocol
IEEEInstitute of Electrical and Electronics Engineers
Appendices 77
InNIndium Nitride
IPCCIntel Parallel Computing Center; Intergovernmental Panel on Climate Change
iPTFintermediate Palomar Transient Factory
ITERAn international fusion energy experiment in southern France
ITGIon Temperature Gradient
IXPUGIntel Xeon Phi Users Group
JCESRJoint Center for Energy Research Storage
JETJoint European Torus
JGIJoint Genome Institute
LEDLight-emitting Diode
LHCIILight-harvesting Complex II
LANLLos Alamos National Laboratory
LLNLLawrence Livermore National Laboratory
MANTISSAMassive Acceleration of New Technologies in Science with Scalable Algorithms
MeVMega-electron Volts
MITMassachusetts Institute of Technology
MOFMetal Oxide Framework
MPPMassively Parallel Processing
MSIMass Spectrometry Imaging
NCARNational Center for Atmospheric Research
NESAPNERSC Exascale Scientific Application Program
NEXAFSNear Edge X-ray Absorption Fine Structure
NGFNERSC Global Filesystem
NIHNational Institutes of Health
NIMNERSC Information Management
NOAANational Oceanic and Atmospheric Administration
NPOffice of Nuclear Physics
NPLQCDNuclear Physics with Lattice QCD
NSFNational Science Foundation
NUGNERSC Users Group
NVRAMNon-volatile Random Access Memory
NWBNeurodata Without Borders
OLCFOak Ridge Leadership Computing Facility
OpenMSIOpen Mass Spectrometry Imaging
OSFOakland Scientific Facility
PDACSPortal for Data Analysis services for Cosmological Simulations
PDSFParallel Distributed Systems Facility, NERSC
PIPrincipal Investigator
PICParticle-In-Cell Simulations
PBPetabytes
PSIIPhotosystem II
PNNLPacific Northwest National Laboratory
PPPLPrinceton Plasma Physics Laboratory
QCDQuantum Chromodynamics
RIPAR Analytics for Image Analysis
SCDOE Office of Science
SciDACScientific Discovery Through Advanced Computing
SDNSoftware-defined Networking
SDSSSloan Digital Sky Survey
SiSilicon
SiO2
Silicon Oxide
SIAMSociety for Industrial and Applied Mathematics
TACCTexas Advanced Computing Center
TBTerabytes
TECAToolkit for Extreme Climate Analysis
UO2
Uranium Dioxide
URLUniversal Resource Locator
XMASX-ray Microdiffraction Analysis Software
NERSC ANNUAL REPORT 201578
FOR MORE INFORMATION ABOUT NERSC, CONTACT:
Jon Bashor
NERSC Communications
1 Cyclotron Road
Berkeley Lab, MS 59R3022B
Berkeley, CA 94720-8148
Email [email protected]
Phone (510) 486-5849
Fax (510) 486-4300
NERSC’s Web Site
nersc.gov
NERSC Annual Report Editor
Kathy Kincade
Contributing Writers and Editors
Jon Bashor, Kathy Kincade, Linda Vu and Margie Wylie
Berkeley Lab Computing Sciences Communications Group
Katie Antypas, Elizabeth Bautista, Shane Canon, Jack Deslippe,
Richard Gerber, Rebecca Hartman-Baker, Glenn Lockwood,
Jackie Scoggins
NERSC
Dan Krotz
Berkeley Lab Public Affairs Communications and
Media Relations Group
Robert Sanders
University of California, Berkeley
Raphael Rosen
Princeton Plasma Physics Laboratory
Siv Schwink
University of Illinois at Urbana-Champaign
Design
Design, layout, illustraton, photography and printing coordination
Berkeley Lab Public Affairs Creative Services Group
Image: Roy Kaltschmidt, Lawrence Berkeley National Laboratory
DISCLAIMER This document was prepared as an account of work sponsored by the United States Government. While this document is believed to contain correct information, neither the United States Government nor any agency thereof, nor The Regents of the University of California, nor any of their employees, makes any warranty, express or implied, or assumes any legal responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by its trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States Government or any agency thereof, or The Regents of the University of California. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof, or The Regents of the University of California.
Ernest Orlando Lawrence Berkeley National Laboratory is an equal opportunity employer. 16-NERSC-2695
NERSC 2015 Annual Report