[hpc-announce] Call for Work In Progress Submission: PDSW' 2024

Gong, Qian gongq at ornl.gov
Mon Aug 19 09:05:20 CDT 2024


[Image]



************ Call for Work In Progress Submission ************



------------------------------------------------------------------------------------------

The 9th International Parallel Data Systems Workshop  (PDSW’ 24)

------------------------------------------------------------------------------------------



PDSW 2024 website: https://urldefense.us/v3/__https://www.pdsw.org/index.shtml__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVVsD_C9-g$ 



Work In Progress (WIP) Submissions due: September 13th, 2024, 11:59PM AoE

Submission website: https://urldefense.us/v3/__https://submissions.supercomputing.org__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVU_T4F5yw$ 

Requirement: an one-page abstract



We are excited to announce the 9th International Parallel Data Systems Workshop (PDSW’24), to be held in conjunction with SC24: The International Conference for High Performance Computing, Networking, Storage, and Analysis, in Atlanta, GA. PDSW’24 builds upon the rich legacy of its predecessor workshops, the Petascale Data Storage Workshop (PDSW, 2006–2015) and the Data Intensive Scalable Computing Systems (DISCS, 2012–2015) workshop.



The increasing importance of efficient data storage and management continues to drive scientific productivity across traditional simulation-based HPC environments and emerging Cloud, AI/ML, and Big Data analysis frameworks. Challenges are compounded by the rapidly expanding volumes of experimental and observational data, the growing disparity between computational and storage hardware performance, and the rise of novel data-driven algorithms in machine learning. This workshop aims to advance research and development by addressing the most pressing challenges in large-scale data storage and processing.



We invite the community to contribute original research manuscripts that introduce and evaluate novel algorithms or architectures, share significant scientific case studies or workloads, or assess the reproducibility of previously published work. We emphasize the importance of community collaboration for problem identification, workload capture, solution interoperability, standardization, and shared tools. Authors are encouraged to provide comprehensive experimental environment details (software versions, benchmark configurations, etc.) to promote transparency and facilitate collaborative progress.



Topics of Interest:



    * Scalable Architectures: Distributed data storage, archival, and virtualization.

    * New Data Processing Models and Algorithms: Application of innovative data processing models and algorithms for parallel computing and analysis.

    * Performance Analysis: Benchmarking, resource management, and workload studies.

    * Cloud and Container-Based Models: Enabling cloud and container-based frameworks for large-scale data analysis.

    * Storage Technologies: Adaptation to emerging hardware and computing models.

    * Data Integrity: Techniques to ensure data integrity, availability, reliability, and fault tolerance.

    * Programming Models and Frameworks: Big data solutions for data-intensive computing.

    * Hybrid Cloud Data Processing: Integration of hybrid cloud and on-premise data processing.

    * Cloud-Specific Opportunities: Data storage and transit opportunities specific to cloud computing.

    * Storage System Programmability: Enhancing programmability in storage systems.

    * Data Reduction Techniques: Filtering, compression, and reduction techniques for large-scale data.

    * File and Metadata Management: Parallel file systems, metadata management at scale.

    * In-Situ and In-Transit Processing: Integrating computation into the memory and storage hierarchy for in-situ and in-transit data processing.

    * Alternative Storage Models: Object stores, key-value stores, and other data storage models.

    * Productivity Tools: Tools for data-intensive computing, data mining, and knowledge discovery.

    * Data Movement: Managing data movement between compute and data-intensive components.

    * Cross-Cloud Data Management: Efficient data management across different cloud environments.

    * AI-enhanced Systems: Storage system optimization and data analytics using machine learning.

    * New Memory and Storage Systems: Innovative techniques and performance evaluation for new memory and storage systems.



The WIP session will showcase brief 5-minute presentations on ongoing work that may not yet be ready for a full paper submission. WIP papers will not be included in the proceedings. A one-page abstract is required for participation.



More details are available at: https://urldefense.us/v3/__https://www.pdsw.org/index.shtml__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVVsD_C9-g$ 





Reproducibility Initiative





Aligned with the SC24 Reproducibility Initiative (https://urldefense.us/v3/__https://sc24.supercomputing.org/program/papers/reproducibility-initiative__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVWF5yfnow$ ), we encourage detailed and structured artifact descriptions (AD) using the SC24 format (https://urldefense.us/v3/__https://github.com/hunsa/sc24-repro__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVX97oe_eg$ ). The AD should include a field for one or more links to data (zenodo, figshare, etc.) and code (Github, GitLab, Bitbucket, etc.) repositories. For the artifacts that will be placed in the code repository, we encourage authors to follow the PDSW 2024 Reproducibility Addendum (https://urldefense.us/v3/__https://www.pdsw.org/pdsw24/Addendum*20for*20the*20PDSW*202024*20Reproducibility*20Initiative.pdf__;JSUlJSUl!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVVxyZxSPw$ ) on how to structure the artifact, as it will make it easier for the reviewing committee and readers of the paper in the future.



Submissions website: https://urldefense.us/v3/__https://submissions.supercomputing.org/__;!!G_uCfscf7eWS!b85ohtHy0_TS6u7mHwphNgiINAE2kmP9-AsZggf1nVVg4c9KXfDD13D6mwuuzNblNBThU04KkyxkSVXJr4XT4g$ 



Organization team:



General Chair:

Bing Xie

Microsoft, USA



Program Co-Chairs:

Suren Byna

The Ohio State University, USA



Anthony Kougkas

Illinois Institute of Technology, USA



Reproducibility Co-Chairs:

Jean Luca Bez

Lawerence Berkeley National Laboratory, USA



Radita Liem

RWTH Aachen University, Germany



Publicity Chair:

Qian Gong

Oak Ridge National Laboratory, USA



Web & Publications Chair:

Joan Digney

Carnegie Mellon University, USA





Sincerely,



Qian Gong

Computer Science and Mathematics Division

Oak Ridge National Laboratory






More information about the hpc-announce mailing list