Back

Minisymposium

MS4F - Machine Learning Support for the Lifetime of Software (ML4SW)

Fully booked
Tuesday, June 17, 2025
15:00
-
17:00
CEST
Room 5.2D02
Join session

Live streaming

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat. Aenean faucibus nibh et justo cursus id rutrum lorem imperdiet. Nunc ut sem vitae risus tristique posuere.

Session Chair

Description

Software plays a critical role in scientific discovery across computational science domains, including chemistry, climate science, physics, and applied mathematics. As software development advances, Machine Learning (ML) is becoming an essential tool for enhancing developer productivity, optimizing application execution, and replacing computationally expensive simulations with surrogate Neural Network models. However, several challenges hinder the broad adoption of ML in software, particularly in the context of sustainable development. With increasingly complex software stacks, workflows, and heterogeneous systems, novel techniques are needed to support development, execution orchestration, and performance optimization. A promising approach for reducing software development overhead in High Performance Computing (HPC) is program synthesis, where software is automatically generated from high-level specifications. Large Language Models (LLMs) such as GPT-4, Code Llama, and StarCoder provide intelligent code generation capabilities, yet challenges related to correctness, verification, and reliability remain. Understanding these limitations is crucial for improving ML-driven software development. The ML4SW minisymposium serves as a platform for researchers, developers, and industry professionals to explore ML-driven software synthesis, correctness verification, and application optimization. Key discussions will address ML’s role in enhancing software development, ensuring trustworthiness, and integrating ML into real-world applications for sustainable and efficient computing.

Presentations

15:00
-
15:30
CEST
Leveraging AI-Driven Code Generation for Portable and Scalable Simulations

As high performance computing (HPC) applications continue to push the boundaries of complexity and performance, leveraging AI-driven code generation has emerged as a powerful approach to streamline scientific software development processes.This talk explores the role of AI code generation models as a collaborative pair programmer in developing portable and scalable code with the SPH-EXA simulation framework.Specifically, we use AI-based code generation to assist development of SPH-EXA using CUDA, HIP, and SYCL to automate the code implementation and unit test generation and significantly reduce development effort while maintaining high standards of correctness and performance. Our work evaluates the dynamic interplay between human expertise and generative AI capabilities. We will assess productivity gains and discuss the practical challenges of ensuring performance portability and scalability across diverse hardware platforms. Through detailed performance analyses we will highlight both the benefits and the limitations of current AI-assisted approaches for scientific software development.This talk will provide insights into how AI-driven tools can serve as effective collaborative partners in HPC code development.

Osman Seckin Simsek and Florina Ciorba (University of Basel)
15:30
-
16:00
CEST
TADASHI: Enabling ML with Correct Code Transformations

As the landscape of machine learning (ML) continues to evolve, the integration of generative AI has become a focal point for automating code generation. While it is perfectly suitable to generate text for humans such as the abstract you're reading, this approach often falls short in ensuring the correctness of the generated code, leading to potential pitfalls in robust ML applications. Recent cases exemplify this challenge, where the use of AI to accelerate coding processes resulted in faster but incorrect code, something unacceptable in scientific computing. In response to this critical need, we introduce TADASHI—a novel library designed to bridge the gap between speed and correctness in code transformations. TADASHI offers a user-friendly Python interface that can seamlessly integrate into existing ML scripts. It empowers developers by not only expediting code modifications but also enforcing rigorous correctness checks on these transformations. By ensuring that code remains valid and reliable, TADASHI enhances the stability of ML workflows and fosters confidence in automated processes. Join us as we delve into TADASHI's capabilities, showcasing its potential to revolutionize the code transformation landscape in ML, ensuring that efficiency does not compromise correctness.

Emil Vatai and Aleksandr Drozd (RIKEN); Ivan R. Ivanov (Institute of Science Tokyo, RIKEN); Joao E. Batista (RIKEN); Yinghao Ren (SenseTime Research and PowerTensors.AI); and Mohamed Wahib (RIKEN)
16:00
-
16:30
CEST
From Reactive Debugging to Proactive Detection: ML for Performance-Aware Software Development

Software performance evolves over time, yet traditional debugging and profiling remain reactive, costly, and disconnected from development workflows. Performance drift—gradual degradation in execution efficiency due to code modifications—often goes undetected until it causes significant slowdowns, forcing late-stage debugging and costly fixes. This talk presents a vision for AI/ML-driven proactive performance-drift detection, where models continuously monitor software evolution, identifying inefficiencies before they degrade execution. By combining static analysis (abstract syntax trees) with dynamic insights from nightly tests, this framework enables early detection of performance-impacting changes. Traditional ML approaches require full model retraining whenever code changes or new runtime data become available, making them impractical for fast-moving development cycles. Few-shot learning eliminates this overhead by allowing models to update incrementally with minimal new data. Attention-based representation learning further enhances interpretability by prioritizing performance-critical features, enabling more targeted interventions. This framework supports two key decision-making processes where (1) developers can receive automated feedback on whether a code change improves or degrades performance, enabling early intervention; (2) the insights can guide hardware configuration choices and runtime parameter tuning. This approach can be seamlessly integrated into CI/CD pipelines to achieve software that not only remains correct but also maintains efficiency as it evolves.

Tanzima Islam (Texas State University)
16:30
-
17:00
CEST
Design and Use of Energy-Efficient Systems in the Deep Learning Era

Modern GPUs, together with larger datasets, facilitate the exponential growth and adoption of deep learning models. The training and deployment of deep neural networks in widely used large-scale data centers, on the other hand, exhibit low GPU hardware utilization, barely reaching 50%, as shown by studies done on Microsoft and Alibaba clusters. This is a waste of hardware resources, especially for expensive GPUs, and contributes to the unsustainable carbon footprint of AI. In fact, less than 15% of large-language model training can lead to a carbon footprint equivalent to the average yearly energy consumption of a US household.This talk will discuss the reasons, challenges and opportunities for designing energy-efficient computing infrastructures as well as its practical applications.

Pamela Delgado (HES-SO)