+ All Categories
Home > Documents > Green HPC: From Nice to Necessity

Green HPC: From Nice to Necessity

Date post: 23-Sep-2016
Category:
Upload: scott
View: 215 times
Download: 1 times
Share this document with a friend
3
G UEST E DITOR’S I NTRODUCTION 8 COMPUTING IN SCIENCE & ENGINEERING Green HPC From Nice to Necessity T he idea of green high-performance computing (HPC) has been gain- ing traction over the past five years. In November 2008, the Green 500 list (www.green500.org) was introduced to “raise awareness about power consumption, promote alternative total cost of ownership performance metrics, and ensure that supercomputers only simulate climate change and not create it.” The Green 500 lists the world’s most energy- efficient supercomputers, based on a floating point operations per second (flops) per watt metric. Although the list has helped raise awareness of en- ergy efficiency for HPC, since its inception, more dramatic drivers for energy-efficient HPC have arisen. Historically, HPC system power consumption has been of secondary importance, but as the HPC community looks from petascale to exa- scale, power will become a first-order concern. 1521-9615/10/$26.00 © 2010 IEEE COPUBLISHED BY THE IEEE CS AND THE AIP Scott Hemmert Sandia National Laboratories As power consumption of supercomputers grows out of control, energy efficiency will move from desirable to mandatory.
Transcript
Page 1: Green HPC: From Nice to Necessity

G U E S T E D I T O R ’ S I N T R O D U C T I O N

8� Computing in SCienCe & engineering

Green HPCFrom Nice to Necessity

T he idea of green high-performance computing (HPC) has been gain-ing traction over the past fi ve years. In November 2008, the Green 500

list (www.green500.org) was introduced to “raise awareness about power consumption, promote alternative total cost of ownership performance

metrics, and ensure that supercomputers only simulate climate change and not create it.”

The Green 500 lists the world’s most energy-effi cient supercomputers, based on a fl oating point operations per second (fl ops) per watt metric. Although the list has helped raise awareness of en-ergy effi ciency for HPC, since its inception, more dramatic drivers for energy-effi cient HPC have arisen.

Historically, HPC system power consumption has been of secondary importance, but as the HPC community looks from petascale to exa-scale, power will become a fi rst-order concern.

1521-9615/10/$26.00 © 2010 ieee

CopuBLiSHeD BY tHe ieee CS AnD tHe Aip

Scott HemmertSandia National Laboratories

As power consumption of supercomputers grows out of control, energy effi ciency will move from desirable to mandatory.

CISE-12-6-Gei.indd 8 16/10/10 2:29 PM

Page 2: Green HPC: From Nice to Necessity

noVemBer/DeCemBer 2010 9

DARPA’s exascale report forecasts a greater than 100 megawatt (MW) power budget for a 1 exa-fl ops machine if current trends continue.1 This power budget is unsustainable from both an en-vironmental and fi nancial standpoint (a 100-MW machine would result in a $100 million power bill each year). So, while green HPC might have been merely desirable to this point, going forward it will be a requirement.

Exascale Requires Green ComputingThe push for exascale computing will provide a driver for an unprecedented level of energy effi -ciency. DARPA’s Ubiquitous HPC program has as its goal a 2018 rack-level prototype that achieves 50 gfl ops/watt compute effi ciency. By compari-son, the current number one on the Green 500 list has an effi ciency of only 773 mfl ops/watt—a difference of more than 60 times. The actual problem might be even more worrisome: eight of the top 10 machines on the Green 500 list are based on accelerator technologies, which are diffi cult to program and unsuitable for some workloads. In addition, the IBM PowerXCell, which powers six of the top 10, has no public road-map going forward. For more information on ac-celerators, see CiSE’s recent special issue.2

Accelerator architectures tend to be either heavily specialized, such as the PowerXCell and graphics processing units (GPUs), or very generic, such as fi eld programmable gate arrays (FPGAs). Both of these models present real challenges when it comes to programming.

After attending a weeklong class on Compute Unifi ed Device Architecture (CUDA; a popular C-based language for programming graphics pro-cessors), one of my coworkers noted, “how easy it is to turn a GPU into a decelerator instead of an accelerator.” With applications commonly ex-ceeding hundreds of thousands to millions of lines of code, programmability is a real issue. More im-portant, however, is the inability of many applica-tions to map to accelerator architectures. At the small scale, this isn’t a concern; a domain-specifi c platform is perfectly acceptable and possibly even desirable. At the high end, where machine costs can run into the hundreds of millions of dollars, suitability for a wide range of applications is a must.

System Balance is KeyThese two issues call into question accelerator vi-ability in the largest exascale machines. But if not accelerators, then what? This is still very much an open question and tends to dominate exascale computing discussions. However, as challenging

as the node architecture will be, other archi-tecture areas will have an important impact on the effi ciency of real applications. This is where both the traditional Top 500 and the Green 500 show their weakness. Both lists are based on the Linpack benchmark, which is notorious for re-quiring high compute capability, moderate mem-ory performance, and only nominal network performance. The drive to be at the top of these lists will inevitably result in system architectures that favor peak fl ops over system balance.

Research using Sandia National Laboratories’ Red Storm supercomputer suggests that a ma-chine with a higher peak compute per watt might actually lower energy effi ciency for real applica-tions, particularly when the cost of higher peak compute effi ciency is an unbalanced system.3

The study looked at application performance on two nearly identical Red Storm confi gurations. The one difference between the two confi gura-tions was in the interconnect bandwidth: one confi guration used full bandwidth and the other only one-quarter bandwidth. Although the quar-ter bandwidth confi guration was 20 percent more energy effi cient when looking at peak fl ops, it was 10 percent less energy effi cient when running CTH, a shock physics code developed at Sandia. This illustrates the need for understanding ap-plication requirements and making architectural decisions based on appropriate metrics.

To help meet this need, the US Department of Energy (DOE) Offi ce of Science has established several exascale codesign centers focused on the codesign of exascale applications and architec-tures. The codesign centers are part of a broader DOE-wide program; one of the program’s main goals is to enable supercomputing systems that can effi ciently run DOE mission-critical science and national security applications. The results from these codesign centers could provide important information on how we can modify supercomput-ers to dramatically improve energy effi ciency.

In This IssueThe push to exascale and unprecedented levels of energy effi ciency will require Herculean efforts, and it’s impossible to cover all aspects of the topic in a single issue. As such, this issue will look at areas whose importance is often overlooked.

In “Money for Research, Not Energy Bills: Finding Energy and Cost Savings in High-Performance Computer Facility Designs,” Dale Sartor and Mark Wilson discuss how thoughtful facility design can greatly reduce energy demands in the machine room and beyond. Their article

CISE-12-6-Gei.indd 9 16/10/10 2:29 PM

Page 3: Green HPC: From Nice to Necessity

10� Computing in SCienCe & engineering

describes approaches for improving datacenter ef-ficiency and is a prime example of looking beyond the supercomputer itself to its broader context.

David W. Jensen and Arun F. Rodrigues wrote “Embedded Systems and Exascale Computing” from the perspective of both embedded and high-performance computing. Although energy ef-ficiency is a relatively new driver for HPC, the embedded space has long sought the best balance of energy and capability. As the authors discuss, embedded computing is beginning to require compute capacities that are causing it to tackle many issues historically limited to the HPC do-main, while HPC is beginning to have many of the constraints that the embedded world has dealt with for years.

In “Software and Hardware Techniques for Power-Efficient HPC Networking,” Torsten Hoefler looks at opportunities to improve the effi-ciency of high-performance interconnects. Power will inevitably be a limiting factor to interconnect performance at the largest scale; it’s therefore vital to make the network transport as efficient as possible. Hoefler’s article reports on a power study done on modern interconnects that points

to areas of improvement for truly energy-efficient networks.

Finally, in “Advanced Architectures and Ex-ecution Models to Support Green Computing,” Richard Murphy, Thomas Sterling, and Chirag Dekate describe research that will be done as part of the DARPA Ubiquitous HPC project. Specifi-cally, it looks at new execution models that can re-duce system overheads to the lowest possible levels to increase scalability—and, in so doing, improve the energy efficiency of real applications.

M any other issues must be resolved to enable efficient exascale comput-ing. One of the primary areas for improvement is in making memory

systems more energy efficient. Providing sufficient memory performance at reasonable power will re-quire significant changes to both the memory in-terface and core memory technology.

Another aspect of system efficiency is resil-ience, which impacts how much time is spent running the application versus preparing for and recovering from faults. Methods currently used to deal with system faults are generally considered to be insufficient for exascale machines. Indeed, exascale computing will drive what have been second-order concerns into the spotlight, but none so dramatically as energy efficiency.

References1. P.M.Kogge,ed.,ExaScale Computing Study: Technol-

ogy Challenges in Achieving Exascale Systems,tech.

reportTR-2008-13,CSEDept.,Univ.NotreDame,

28Sept.2008;www.cse.nd.edu/Reports/2008/

TR-2008-13.pdf.

2. VolodymyrKindratenkoetal.,eds.,“High-Performance

ComputingwithAccelerators,”specialissue,Comput-

ing in Science & Eng.,vol.12,no.4,2010.

3. R.Brightwelletal.,“ChallengesforHigh-

PerformanceNetworkingforExascaleComputing,”

Proc. Int’l Conf. Computer Comm.,IEEEPress,2010;

doi:10.1109/ICCCN.2010.5560144.

Scott Hemmert is a senior member of the technical staff at Sandia National Laboratories, where he leads the advanced supercomputer interconnect research. He’s also a member of the joint Sandia/Los Alamos National Laboratory Alliance for Computing at Extreme Scale (ACES) design team, where he serves as co-lead for interconnect architectures. His research interests include supercomputer interconnects and exascale architectures. Hemmert has a PhD in elec-trical engineering from Brigham Young University. Contact him at [email protected].

CISE-12-6-Gei.indd 10 16/10/10 2:29 PM


Recommended