[hpc-announce] FTXS 2023 @ SC23: Call for papers (Deadline extended + featured speaker)

Levy, Scott Larson sllevy at sandia.gov
Tue Aug 8 08:53:49 CDT 2023

Two brief updates for FTXS:

* The deadline for FTXS 2023 has been extended until August 16, 2023

* The Featured Speaker for our workshop will be Paolo Rech from Università di Trento (Italy) and UFRGS (Brazil).  In his talk, "Quantum Computing Reliability: Problems, Tools, and Potential Solutions", he'll discuss the work that he and his colleagues have done on understanding and mitigating faults in quantum computing systems.

From: Levy, Scott Larson <sllevy at sandia.gov>
Sent: Thursday, July 6, 2023 9:21 AM
To: hpc-announce
Subject: FTXS 2023 @ SC23: Call for papers

13th Workshop on Fault-Tolerance for HPC at eXtreme Scale (FTXS 2023)

In conjunction with The International Conference for
High Performance Computing, Networking, Storage, and Analysis (SC23)
Denver, Colorado, USA November 12 - 17, 2023

Important Dates
* Submissions open: July 1, 2023 (Submissions are now open!)
* Submission of papers: August 16, 2023 (UPDATED)
* Author notification: September 14, 2023
* Camera-ready papers: October 5, 2023
* Workshop: (TBD) November 12 (1:30-5:00pm), 2023

Authors are invited to submit original papers on the research and practice of fault-tolerance in
extreme-scale distributed systems (primarily HPC systems, but including grid and cloud systems).
Resilience and fault-tolerance remain a major concern for supercomputing and advances in this area
are needed.  Therefore, we are broadly interested in forward-looking papers that seek to
characterize and mitigate the impact of faults.

We are particularly interested in papers that address issues related to the following developments
in extreme-scale systems:

* Storage Devices: The storage hierarchy on HPC systems continues to increase in depth and
  complexity. SSDs and NVMe add high-speed node-local (or rack-local) persistent storage that
  can be used to improve the performance of checkpoint/restart or otherwise facilitate application
  resilience. Continuing to efficiently exploit these devices remains critical for extreme-scale HPC
  systems. Moreover, the recent availability of Non-Volatile Memory Modules (NVMMs) has begun
  to blur the line between memory and storage. The implications of this blurring for fault tolerance
 on extreme-scale systems are still being explored.

* System Heterogeneity: Modern HPC systems increasingly include GPUs, FPGAs, and other types
  of accelerators. New networking devices like Data Processing Units (DPUs) and SmartNICs are
  also starting to be deployed. However, there are many resilience and fault tolerance issues
  associated with these devices that still need to be resolved. Papers at prominent recent
  conferences (including SC20, ICS 2019, and IEEE Cluster 2018) demonstrate that understanding
  the fault tolerance implications of heterogeneous compute devices is an important and active
  area of research.

* Computing Paradigms: Novel non-von Neumann computing paradigms, including quantum and
  neuromorphic computing, have attracted significant research interest. Recent publications
  demonstrate that understanding the fault tolerance implications of these computing paradigms is
  also an area of active research.

* Machine Learning: Algorithms that rely on elements of machine learning are becoming more and
  more prevalent on HPC systems.  Understanding how these algorithms react and respond to the
  frequency and variety of faults that occur on HPC systems is critical to ensuring that they continue to
  provide accurate and timely answers.

Additional topics of interest include, but are not limited to:

*   Algorithmic-Based Fault Tolerance (ABFT) techniques to address undetected (silent) errors
*   Silent data corruption (SDC) detection / correction techniques
*   Novel fault-tolerance techniques and implementations
*   Failure data analysis and field studies
*   Power, performance, resilience (PPR) assessments / tradeoffs
*   Emerging hardware and software technology for resilience
*   Advances in reliability monitoring, analysis, and control of highly complex systems
*   Failure prediction, error preemption, and recovery techniques
*   Fault-tolerant programming models
*   Models for software and hardware reliability
*   Metrics and standards for measuring, improving, and enforcing effective fault-tolerance
*   Scalable Byzantine fault-tolerance and security from single-fault and fail-silent violations
*   Atmospheric evaluations relevant to HPC systems (terrestrial neutrons, temperature, etc.)
*   Near-threshold-voltage implications and evaluations for reliability
*   Benchmarks and experimental environments including fault injection
*   Frameworks and APIs for fault-tolerance and fault management

Submissions are solicited in the following categories:
* Regular papers presenting innovative ideas improving the state of the art or discussing the issues
  seen on existing extreme-scale systems, including some form of analysis and evaluation.   Regular
  papers must be at least six (6) pages should not exceed eleven (11) pages including all text, appendices,
  figures, references.
* Extended abstracts presenting preliminary results, proposing disruptive ideas, or challenging
  assumptions in the field.  The inclusion of some form of preliminary results is encouraged.
  Extended abstract papers should not exceed four (4) pages, including all text, appendices, figures,
  and references.  Extended abstracts will be evaluated separately and given shorter oral

Submissions shall be submitted electronically at submissions.supercomputing.org and must conform
to IEEE conference proceedings style.  IEEE templates are available at:

We do not have an upper limit on the number of papers that we will accept.  We will make every
effort to make sure that every high-quality submission will be included in our workshop.

Subject to publisher constraints, our workshop will publish all submissions accepted for inclusion
in our workshop.  Our workshop has been approved to have our accepted papers included in the SC Workshop Proceedings.

Reproducibility is an important component of extreme-scale system research.  However, the goal of
our workshop is to encourage and facilitate discussion of novel approaches and preliminary results.
As a result, it may not always be feasible to release reproducibility artifacts.  Therefore, while
we encourage authors to make their work as public and reproducible as possible, we do not explicitly
require it.

Scott Levy - Sandia National Laboratories

Keita Teranishi - Sandia National Laboratories
John Daly - Laboratory for Physical Sciences

Questions? Contact Scott Levy (sllevy at sandia.gov).

More information about the hpc-announce mailing list