[hpc-announce] Energy Eff HPC State-of-practice - collocated w/ Cluster 2020 - Japan -- [CFP]

Siddhartha Jana siddharthajana24 at gmail.com
Mon Jun 1 12:19:22 CDT 2020

Call for Papers:
*Energy Efficient HPC State of the Practice Workshop (EE HPC SOP 2020)
* *September 14th, 2020 - Kobe, Japan *

In conjunction with Cluster2020

Submission Deadline:  July 06, 2020
Author Notification:  August 03, 2020
Camera Ready Paper:  August 14th, 2020
** All dates are Anywhere on Earth

The facility demands for supercomputing centers (SCs) are
characterized by electrical power demands for computing systems that
scale to tens of megawatts (MW) and millisecond power fluctuations
approaching 30MW for the largest systems. The demand for primary
electrical distribution capabilities to current large-scale facilities
can exceed 60MW, comprising multiple, redundant, and diverse
medium-voltage feeders.   Despite significant pressure on both Moore’s
Law and Dennard scaling, the appetite for ever-larger systems and the
subsequent demand for both agile power and effective cooling for these
systems continues to grow. Computing trends, in terms of highly
optimized hardware platforms that may leverage accelerators or other
non-traditional components, scalable and high performing applications,
and the requirements to manage exponentially larger data sets are
driving facility demands not envisioned just a few years ago.

SC facilities must consider multiple elements, including the cost to
extend or fit existing primary distribution capabilities; the cost and
consequence of both trapped and stranded capacity, ever-increasing
heat densities for new systems that may render existing cooling
mechanisms obsolete or ineffective, increased mandatory use of liquid
cooling for portions of the heat load, and wet weights that exceed the
carrying capacities of existing raised floor systems.

Additionally, the operational costs of these facilities must be
balanced versus the demand from the systems owners and users for high
availability, high utilization, and low-impact facility maintenance
and service demands. To achieve this balance, many SCs continue to
innovate their operational design practices and technologies.
Solutions seek improved management of both the electrical and
mechanical systems, and minimizing long-term facility costs through
best practices associated with their design.

Some SCs are early adopters and innovators in operational practices
and technologies that are geared towards improving energy and power
management capabilities.  This workshop will explore these operational
and technological innovations that span HPC computational systems as
well as buildings and building infrastructure.

The purpose of this workshop is to allow for publication of practices,
policies, procedures, and technologies in formal peer-reviewed papers
so the broader community can benefit from these experiences.    It
will expose use cases, lessons learned and best practices in design,
commissioning and operations.  The nature of these papers is generally
descriptive with hard experiential data generally gathered through
surveys, case studies and research for practice.

## Workshop Topics of interest include (but are not limited to):
** Electrical power distribution
  -- large HPC power loads and rapid power swings
  -- electricity service provider relationships with HPC facility
  -- facility system design and commissioning

** Power and energy measurement, monitoring and control
  -- operational data collection, aggregation and analytics
  -- energy and power-aware job scheduling and resource management
  -- cooling control systems
  -- standards and open interfaces (e.g., Power API, Redfish, GEOPM,
READEX, PowerStack)

** Power and energy procurement considerations
  -- system requirements (e.g., HPC equipment, software, mechanical
systems, facilities)
  -- operational costs in procurement

** Liquid cooling
  -- standards and open interfaces (e.g., OCP, ASHRAE)
  -- facility system design and commissioning

** HPC facility preventative maintenance and management practices for
RAS-M (reliability, availability, serviceability, and maintainability)

** Measuring and evaluating the value of ITUE (IT power usage effectiveness)

## Paper Submission
Abstracts and papers to be submitted via Easy Chair -

Accepted papers must be presented at the workshop whether in person or
virtually.  Long papers should not exceed **10** pages and short
papers should not exceed **4** pages in the IEEE format

Submissions must be in PDF format. Since the camera-ready version of
accepted papers must be compliant with the IEEE Xplore format for
publication, submitted papers must conform to the following Xplore
layout, page limit, and font size.
Submissions are required to be no more than 10 pages (excluding references).
Submissions must be single-spaced, 2-column numbered pages in IEEE
Xplore format (8.5x11-inch paper, margins in inches – top: 0.75,
bottom: 1.0, sides:0.625, and between columns:0.25, main text: 10pt).
Author names and affiliations should be included in the submitted
paper, and appropriate citations of prior work must be included.
Submissions will be judged on correctness, novel or innovative
approaches to a problem, technical and/or operational strength,
written quality, and interest and relevance to the workshop scope.
The workshop organizers will provide written reviews for all timely
submissions.  Editorial review and recommendations may be provided as

Accepted papers will be published by IEEE as part of the Cluster2020

## Committees
Natalie Bates, Energy Efficient HPC Working Group Fumiyoshi Shoji,
RIKEN Jim Rogers, Oak Ridge National Laboratory Herbert Huber, Leibniz
Supercomputing Center (LRZ) Siddhartha Jana, Intel Torsten Wilde, HPE

Anna Maria Bailey, Lawrence Livermore National Laboratory David Grant,
Oak Ridge National Laboratory David Martinez, Sandia National
Laboratory James H. Laros, Sandia National Laboratory

Greg Koenig, KPMG
Steve Martin, HPE
Grant Stewart, Los Alamos National Laboratory
Gert Svensson, Royal Institute of Technology (KTH)
Daniel Hackenberg, Technical University of Dresden
Bilel Hadri, King Abdullah University of Science and Technology
Jeff Broughton, Lawrence Berkeley National Laboratory
Jason Hick, Los Alamos National Laboratory Ryan Grant, Sandia National
Sridutt Bhalachandra, Lawrence Berkeley National Laboratory
Daniele Cesarini, CINECA
Ayse Coskun, Boston University
Thomas Durbin, Independent Vadim Elisseev, IBM
Dan Ellsworth, Colorado College
Parks Fields, New Mexico Computing Consortium
Jeff Hanson, HPE
Neena Imam, Oak Ridge National Laboratory
Alexander Moskovsky, RSC
Ram Nagappan, Intel
Jorji Nonaka, RIKEN
Kevin Pedretti, Sandia National Laboratory
Joe Prisco, IBM
Woong Shin, Oak Ridge National Laboratory
Hayk Shoukourian, Leibniz Supercomputing Center (LRZ)
Xingfu Wu, Argonne NL
Andrea Bartolini, University of Bologna
Matthias Maiterth, LMU Munich
Francois Robin, CEA
Michael Ott, Leibniz Supercomputing Center (LRZ)
Tom Scogland, Lawrence Livermore National Laboratory

More information about the hpc-announce mailing list