[hpc-announce] Call for Papers: 2023 IEEE International Symposium on Workload Characterization

Resit Sendag sendag at ele.uri.edu
Mon Apr 10 09:57:07 CDT 2023


Apologies if you have received multiple copies of this email.

2023 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION

CALL FOR PAPERS (HTTPS://IISWC.ORG/IISWC2023/)

IISWC invites manuscripts that present original unpublished research in
all areas related to characterization and analysis of computing system
workloads, including translational research related to
production-oriented commercial systems. Work focusing on emerging
technologies and interdisciplinary work are especially welcome. Topics
of interest include (but are not limited to): Characterization of
applications in traditional and emerging domains, characterization of
system software and middleware, implications of workloads in system
design, benchmarking methodologies and suites, and tools for computer
systems. A detailed list of the topics can be found at the end of this
CFP. 

DEADLINES

 	* Submission Deadline: June 14, 2023
 	* Decision Notification: August 2, 2023
 	* Artifact Submission deadline: August 14, 2023
 	* Camera-ready deadline: September 4, 2023

ARTIFACT EVALUATION

This year, IISWC will continue to include an artifact evaluation process
to promote the reproducibility of experimental results. We will invite
the authors of accepted IISWC papers to submit their supporting
materials to the Artifact Evaluation process, which is to assess how the
artifacts support the work described in the papers. This submission will
be VOLUNTARY and will not influence the final decision regarding
acceptance of the paper. The description of the artifact will not be
included in the page limit. The artifact submission deadline will be
shortly after the notification of the paper's acceptance -- authors
should prepare in advance to ensure sufficient time for artifact
assembly and documentation. More details of artifact evaluation will be
made available to the authors of the accepted paper. 

SUBMISSION GUIDELINES

Submissions to IISWC can be made in one of the following two categories:
(1) regular papers (2) tool and benchmark papers. The primary focus of
_regular papers_ (submission length: 10 pages, excluding references)
should be to describe new research ideas supported by experimental
implementation and evaluation of the proposed research ideas. The
primary focus of _tool and benchmark papers_ should be to describe the
design, development, and evaluation of new open-source tools /
benchmarks suites. Submissions in the _regular papers_ category are also
encouraged to open-source their software or hardware artifacts. 

The authors are required to indicate the category of the paper as a part
of the submitted manuscript's title. On the submission system entry, we
ask the authors to add a prefix to the title indicating the type of the
submission as follows: 1. Regular papers: "Regular-TITLE"; 2.
Tool/Benchmark papers: "Tools-TITLE". 

Papers in the tool and benchmark category with relatively shorter length
(6 pages) are welcome, if the contributions can be well articulated and
substantiated. However, all submissions in the tool and benchmark
category have the flexibility of using all 10 pages (excluding
references). The submissions in both categories will be evaluated to the
same standards in terms of novelty, scientific value, demonstrated
usefulness, and potential impact on the field. The nature of the
contribution differs between the two categories (new research idea vs.
new open-source benchmark-suite / tool) and papers will be evaluated
based on the intended nature of the contribution, as declared by the
chosen paper category at the time of the submission. The chosen category
at the time of the submission can not be changed after the submission
deadline. 

DOUBLE-BLIND submission guidelines apply to the submissions in both
categories. 

Open-source benchmarks and tools that have not been previously published
(but may have been open-sourced) are eligible for submission in the
_tool and benchmark papers_ category. 

When including source code links in their submission, we require the
authors to use new or anonymized code repositories to preserve the
integrity of double-blind review process. All submitted papers should
have obtained legal permission (if applicable) to open-source the
benchmark-suite / tool at the time of submission. 

TOPICS OF INTEREST

Characterization of applications in domains including 

 	* Life sciences, bioinformatics, scientific computing, finance,
forecasting
 	* Machine learning, data analytics, data mining
 	* Cyber-physical systems, pervasive computation and Internet of Things
(IoT)
 	* Security and privacy-preserving computing
 	* High performance computing
 	* Cloud and edge computing
 	* Mobile computing
 	* User behavior and system-user interaction
 	* Search engines, e-commerce, web services, and databases
 	* Embedded, multimedia, real-time, 3D-graphics, gaming
 	* Blockchain services
 	* Augmented reality and virtual reality

Characterization of workloads for emerging workloads and architectures,
such as 

 	* Quantum computations and communication
 	* Serverless computing
 	* Near-threshold computing
 	* Non-volatile memory
 	* Near data processing architectures
 	* Neuromorphic and brain-inspired computing
 	* Transactional memory systems
 	* Biology (e.g., DNA sequencing) and chemistry workloads

Characterization of OS, Virtual Machine, middleware and library
behavior, including 

 	* Virtual machines, .NET, Java VM, databases
 	* Graphics libraries, scientific libraries
 	* Operating system and hypervisor effects and overheads

Implications of workloads in system design, such as 

 	* Power management, reliability, security, privacy, performance
 	* Processors, memory hierarchy, I/O, and networks
 	* Design of accelerators, FPGAs, GPUs, CGRAs, etc.
 	* Large-scale computing infrastructures and facilities

Benchmark methodologies and suites, including 

 	* Representative benchmarks for emerging workloads
 	* Benchmark cloning methods
 	* Profiling, trace collection, synthetic traces
 	* Validation of benchmarks

Measurement tools and techniques, including 

 	* Instrumentation methodologies for workload verification and
characterization
 	* Techniques for accurate analysis/measurement of production systems
 	* Analytical and abstract modeling of program behavior and systems


More information about the hpc-announce mailing list