[hpc-announce] Energy Eff HPC SOP @ Cluster'20 -- Deadline extension

Siddhartha Jana siddharthajana24 at gmail.com
Sun Jul 5 10:30:19 CDT 2020


The (Energy Eff HPC SOP (EE HPC SOP) submission deadline and author
notification dates have been extended (July 13th and August 7th,
respectively).  To those needing a bit more time, we hope you can take
advantage of this extension. For those of you who were timely and
ready to submit by the 6th of July, thank you.

*** Note ***: Virtual participation will be available; travel is not
required.  We’re looking forward to an engaging and productive

## Call for Short & Long Papers:

*Energy Efficient HPC State of the Practice Workshop (EE HPC SOP 2020) *
*September 14th, 2020*

In conjunction with Cluster2020

Submission Deadline: July 13, 2020
Author Notification: August 07, 2020
Camera Ready Paper:  August 14th, 2020
** All dates are Anywhere on Earth

The facility demands for supercomputing centers (SCs) are
characterized by electrical power demands for computing systems that
scale to tens of megawatts (MW) and millisecond power fluctuations
approaching 30MW for the largest systems. The demand for primary
electrical distribution capabilities to current large-scale facilities
can exceed 60MW, comprising multiple, redundant, and diverse
medium-voltage feeders. Despite significant pressure on both Moore’s
Law and Dennard scaling, the appetite for ever-larger systems and the
subsequent demand for both agile power and effective cooling for these
systems continues to grow. Computing trends, in terms of highly
optimized hardware platforms that may leverage accelerators or other
non-traditional components, scalable and high performing applications,
and the requirements to manage exponentially larger data sets are
driving facility demands not envisioned just a few years ago.

SC facilities must consider multiple elements, including the cost to
extend or fit existing primary distribution capabilities; the cost and
consequence of both trapped and stranded capacity, ever-increasing
heat densities for new systems that may render existing cooling
mechanisms obsolete or ineffective, increased mandatory use of liquid
cooling for portions of the heat load, and wet weights that exceed the
carrying capacities of existing raised floor systems.

Additionally, the operational costs of these facilities must be
balanced versus the demand from the systems owners and users for high
availability, high utilization, and low-impact facility maintenance
and service demands. To achieve this balance, many SCs continue to
innovate their operational design practices and technologies.
Solutions seek improved management of both the electrical and
mechanical systems, and minimizing long-term facility costs through
best practices associated with their design.

Some SCs are early adopters and innovators in operational practices
and technologies that are geared towards improving energy and power
management capabilities.  This workshop will explore these operational
and technological innovations that span HPC computational systems as
well as buildings and building infrastructure.

The purpose of this workshop is to allow for publication of practices,
policies, procedures, and technologies in formal peer-reviewed papers
so the broader community can benefit from these experiences.    It
will expose use cases, lessons learned, and best practices in design,
commissioning, and operations.  The nature of these papers is
generally descriptive with hard experiential data generally gathered
through surveys, case studies, and research for practice.

## Workshop Topics of interest include (but are not limited to):
** Electrical power distribution
  -- large HPC power loads and rapid power swings
  -- electricity service provider relationships with HPC facility
  -- facility system design and commissioning

** Power and energy measurement, monitoring and control
  -- operational data collection, aggregation, and analytics
  -- energy and power-aware job scheduling and resource management
  -- cooling control systems
  -- standards and open interfaces (e.g., Power API, Redfish, GEOPM,
READEX, PowerStack)

** Power and energy procurement considerations
  -- system requirements (e.g., HPC equipment, software, mechanical
systems, facilities)
  -- operational costs in procurement

** Liquid cooling
  -- standards and open interfaces (e.g., OCP, ASHRAE)
  -- facility system design and commissioning

** HPC facility preventative maintenance and management practices for
RAS-M (reliability, availability, serviceability, and maintainability)

** Measuring and evaluating the value of ITUE (IT power usage effectiveness)

## Paper Submission

Abstracts and papers to be submitted via Easy Chair -

Accepted papers must be presented at the workshop whether in person or
virtually.  Long papers should not exceed **10** pages and short
papers should not exceed **4** pages in the IEEE format.
Submissions must be in PDF format. Since the camera-ready version of
accepted papers must be compliant with the IEEE Xplore format for
publication, submitted papers must conform to the following Xplore
layout, page limit, and font size.
Submissions are required to be no more than 10 pages (excluding references).
Submissions must be single-spaced, 2-column numbered pages in IEEE
Xplore format (8.5x11-inch paper, margins in inches – top: 0.75,
bottom: 1.0, sides:0.625, and between columns:0.25, main text: 10pt).
Author names and affiliations should be included in the submitted
paper, and appropriate citations of prior work must be included.
Submissions will be judged on correctness, novel or innovative
approaches to a problem, technical and/or operational strength,
written quality, and interest and relevance to the workshop scope.
The workshop organizers will provide written reviews for all timely
submissions.  Editorial reviews and recommendations may be provided as

**** Accepted papers will be published by IEEE as part of the
Cluster2020 Proceedings. ****

## Committees

Natalie Bates, Energy Efficient HPC Working Group
Fumiyoshi Shoji, RIKEN
Jim Rogers, Oak Ridge National Laboratory
Herbert Huber, Leibniz Supercomputing Center (LRZ)
Siddhartha Jana, Intel
Torsten Wilde, HPE

Anna Maria Bailey, Lawrence Livermore National Laboratory
David Grant, Oak Ridge National Laboratory
David Martinez, Sandia National Laboratory
James H. Laros, Sandia National Laboratory

Greg Koenig, KPMG
Steve Martin, HPE
Grant Stewart, Los Alamos National Laboratory
Gert Svensson, Royal Institute of Technology (KTH)
Daniel Hackenberg, Technical University of Dresden
Bilel Hadri, King Abdullah University of Science and Technology
Jeff Broughton, Lawrence Berkeley National Laboratory
Jason Hick, Los Alamos National Laboratory
Ryan Grant, Sandia National Laboratory
Sridutt Bhalachandra, Lawrence Berkeley National Laboratory
Daniele Cesarini, CINECA
Ayse Coskun, Boston University
Thomas Durbin, Independent
Vadim Elisseev, IBM
Dan Ellsworth, Colorado College
Parks Fields, New Mexico Computing Consortium
Jeff Hanson, HPE
Neena Imam, Oak Ridge National Laboratory
Alexander Moskovsky, RSC
Ram Nagappan, Intel
Jorji Nonaka, RIKEN
Kevin Pedretti, Sandia National Laboratory
Joe Prisco, IBM
Woong Shin, Oak Ridge National Laboratory
Hayk Shoukourian, Leibniz Supercomputing Center (LRZ)
Xingfu Wu, Argonne NL
Andrea Bartolini, University of Bologna
Matthias Maiterth, LMU Munich
Francois Robin, CEA
Michael Ott, Leibniz Supercomputing Center (LRZ)
Tom Scogland, Lawrence Livermore National Laboratory

More information about the hpc-announce mailing list