[hpc-announce] LLM4HPCAsia at SCA/HPC Asia'26 CFP (January 29th, 2026)

Valero Lara, Pedro valerolarap at ornl.gov
Wed Oct 8 20:08:58 CDT 2025


LLM4HPCAsia 2026
The 1st International Workshop on Foundational large Language Models Advances for HPC in Asia:
https://urldefense.us/v3/__https://ornl.github.io/events/llm4hpcasia2026/__;!!G_uCfscf7eWS!ZCCPwleQHWN1e53BgoAveCe8KBRUe0NgzCnH9mqvCOoIdSnaJVDRJMl5kHkUdZ9F8I42SseSdhaWyvFY98KMG9o7Wrhx_A$ 
to be held in conjunction with SCA/HPC Asia 2026:
https://urldefense.us/v3/__https://www.sca-hpcasia2026.jp/__;!!G_uCfscf7eWS!ZCCPwleQHWN1e53BgoAveCe8KBRUe0NgzCnH9mqvCOoIdSnaJVDRJMl5kHkUdZ9F8I42SseSdhaWyvFY98KMG9rJ_Ki5nw$ 

29 January, 2026
Osaka, Japan

-- Introduction

Since their development and release, modern Large Language Models (LLMs), such as the Generative Pre-trained Transformer (GPT) model and the Large Language Model Meta AI (LLaMA), have come to signify a revolution in human-computer interaction spurred on by their high-quality results. LLMs have repaved this landscape thanks to unprecedented investments and enormous training models (hundreds of billions of parameters). The availability of LLMs has led to increasing interest in how they could be applied to a large variety of applications. The HPC community made recent research efforts to evaluate current LLM capabilities for some HPC tasks, including code generation, auto parallelization, performance portability, correctness, among others. All these studies concluded that state-of-the-art LLM capabilities have proven so far insufficient for these targets. Hence, it is necessary to explore novel techniques to further empower LLMs to enrich the HPC mission and its impact.

-- Objectives, scope and topics of the workshop

This workshop objectives are focused on LLMs advances for any HPC major priority and challenge with the aims to define and discuss the fundamentals of LLMs for HPC-specific tasks, including but not limited to hardware design, compilation, parallel programming models and runtimes, application development, enabling LLM technologies to have more autonomous decision-making about the efficient use of HPC. This workshop aims to provide a forum to discuss new and emerging solutions to address these important challenges towards an AI-assisted HPC era. Papers are being sought on many aspects of LLM for HPC targets including (but not limited to):

- LLMs for Programming Environments and Runtime Systems
- LLMs for HPC and Scientific Applications
- LLMs for Hardware design (including non-von Neumann Architectures)
- Reliability/Benchmarking/Measurements for LLMs

-- Important Dates:

Paper (firm) submission deadline : Oct 20, 2025
Notification of acceptance : Nov 26, 2025

-- Organizers (Contact us)

Pedro Valero-Lara (chair)
Oak Ridge National Laboratory, USA
valerolarap at ornl.gov

William F. Godoy (co-chair)
Oak Ridge National Laboratory, USA
godoywf at ornl.gov

Dhabaleswar K. Panda (co-chair)
The Ohio State University, USA
panda at cse.ohio-state.edu

-- Program Committee

- Franz Franchetti, Carnegie Mellon University, USA
-  Min Si, AI Facebook, USA
- Patrick Diehl, Los Alamos National Laboratory, USA
- Diego Andrade Canosa, University of A Coruna, Spain
- Monil Mohammad Alaul Haque, Oak Ridge National Laboratory, USA
- Tze-Meng Low, Carnegie Mellon University, USA
- Keita Teranishi, Oak Ridge National Laboratory, USA
- Hiroyuki Takizawa, Tohoku University, Japan
- Olivier Aumage, INRIA, France
- Upasana Sridhar, Carnegie Mellon University, USA
- Het Mankad, Oak Ridge National Laboratory, USA
- Rabab Alomairy, Massachusetts Institute of Technology, USA
- Jens Domke, RIKEN Center for Computational Science (R-CSS), Japan
- Gokcen Kestor, Barcelona Supercomputing Center, Spain
- Kshitij Srivastava, Aurora, USA
- Simon Garcia De Gonzalo, Sandia National Laboratory, USA
- Shilei Tian, AMD, USA
- Josh Davis, University of Maryland, USA
- Yuning Xia, Rice University, USA
- Erin Carrier, Grand Valley State Unviersity, USA
- Noujoud Nader, Louisiana State University, USA
- Ignacio Laguna, Lawrence Livermore National Laboratory, USA

-- Manuscript submission:

We invite submissions of original, unpublished research and experiential papers. Papers should be at most 12 pages in length (including a bibliography and appendices, with two possible extra pages after the review to address the reviewer’s comments), formatted according to single-column ACM Proceedings Style. More details about the format can be found in the SCA/HPCAsia 2026 submission website. All paper submissions will be managed electronically via EasyChair.

-- Proceedings:

All accepted papers will be published in the SCA/HPCAsia Workshops 2026 proceedings by ACM.

-- Best Paper Award

The Best Paper Award will be selected on the basis of explicit recommendations of the reviewers and their scoring towards the paper’s originality and quality.

-- Keynote (Rio Yokota, Institute of Science Tokio):

- Updates on the Development of Japanese LLMs
Large language models (LLM) are mainly pre-trained on internet data, which is predominantly English. Such models have suboptimal performance when used in non-English languages. Also, LLMs are not mechanical tools that benefit everyone equally. They are rather intellectual tools that disproportionately benefit certain groups of people, depending on what data they are trained on. Furthermore, the interaction with LLMs will influence our local culture in the long term. Sovereign LLMs are crucial for customizing the models to meet the needs of each local culture. In this talk I will give an update of the efforts in Japan to train LLMs. I will cover both the data and training aspects.

- Rio Yokota is a Professor at the Supercomputing Research Center, Institute of Integrated Research, Institute of Science Tokyo. He also leads the AI for Science Foundation Model Research Team at RIKEN Center for Computational Science. His research interests lie at the intersection of high performance computing, machine learning, and linear algebra. He has been optimizing algorithms on GPUs since 2007, and was part of a team that received the Gordon Bell prize in 2009 using the first GPU supercomputer. More recently, he has been leading distributed training efforts on Japanese supercomputers such as ABCI, TSUBAME, and Fugaku. He is the co-developer of the Japanese LLM Swallow, and LLM-jp. He is also involved in the organization of multinational collaborations such as ADAC and TPC.

-- Invited Talk (Min Si, Facebook AI):

- High performance communication library and transport for LLM training at 100K+ Scale
Each successive generation of the LLaMA model has demonstrated substantial growth in both model size and complexity. The largest multimodal mixture-of-experts model within our LLaMA4 series possesses nearly two trillion total parameters, with 288 billion active parameters and 16 experts. To accommodate the computational demands associated with training such a colossal model, we expanded our AI clusters, deploying approximately 100,000 GPUs. GPU-to-GPU communication latency is a critical factor when coordinating such a vast number of GPUs. Even microsecond delays accumulate across thousands of nodes, consequently impacting the time required for training. We engineered the underlying network infrastructure to provide the necessary backbone for high-speed GPU-to-GPU communication, concurrently innovating our communication library stack to enhance overall communication efficiency. In this presentation, we will provide an overview of the network topology deployed within Meta datacenters and introduce a range of communication optimizations and custom features that facilitated LLaMA4 training through cross-layer codesign, encompassing model algorithms, collectives, and extending to the network transport layer.

- Min Si is a Research Scientist at Facebook AI System SW/HW Co-design group. Her role is to investigate and resolve interesting scale-out challenges for Facebook AI workloads. Previously, she was an Assistant Computer Scientist at Argonne National Laboratory and working with the Programming Models and Runtime Systems group. Her research interests include communication runtime in high-performance computing, parallel programming models and runtime systems.

-- Registration

Information about registration at SCA/HPC Asia 2026 website.



More information about the hpc-announce mailing list