Improving access to supercomputers to boost meteorology, motor design and disaster reduction
Europe’s high-performance computing (HPC) capacity will soon reach exascale speeds, with systems more powerful than 1 million of the fastest laptops. This will allow a wide range of next-generation innovations. Today’s data-intensive tasks, such as weather forecasting, are dependent on computationally heavy workflows combining modelling and simulations and new data analytics along with artificial intelligence (AI) large language models. Current workflows and associated tools, designed for use in less powerful supercomputers, will probably need to be adapted for these new machines. “The programming models and tools of the applications underlying many current workflows are different, so we need to simplify how they work together,” explains Rosa Badia from the Barcelona Supercomputing Center and coordinator of the eFlows4HPC project. The project has called their approach ‘HPC Workflow-as-a-Service’ because their software solution, hosted on a user-friendly platform, seamlessly integrates various applications. “Our approach – which was not possible before – has simplified access to HPC systems, opening them up to non-expert users, spreading the scientific, social and industrial opportunities of supercomputing,” says Badia. So far, over 30 scientific publications have been generated using this approach, alongside a policy white paper on natural hazards.
Under the bonnet
The eFlows4HPC project added new functionalities to existing tools, supporting larger and more complex HPC workflows, and cutting-edge technologies such as AI and big data. At the solution’s heart lies a series of software components, collectively known as a software stack, that implements HPC-based workflows, from start to finish. Firstly, the workflow’s various applications – such as simulators, data processing and machine learning predictions – are developed using the PyCOMPSs programming environment. Data pipelines are also set up to coordinate how workflow data is managed, combining these computational and data management aspects of the workflow. Finally, everything needed to run the software is encapsulated in a container file and installed on the HPC system. Users can then run their workflows using a dedicated interface.
Testing with diverse user communities
To demonstrate the potential of their approach, the team conducted three use-case trials with very different user communities. The first use-case trial created is a digital twin of manufacturing processes and assets which simulates experiments with various research and development scenarios cost-effectively. For example, a project partner, CIMNE, collaborated with Siemens to produce a model of an electric motor able to avoid overheating. As this partnership produced encouraging results, CIMNE now intends to create a spin-off company to exploit these outcomes. The second focused on Earth system modelling, using two AI-based workflows to increase the efficiency of both Earth system models and prediction models for North Pacific tropical cyclones. These simulations could ultimately help improve global early warning, adaptation and mitigation systems. With both workflows returning positive results, extensions are now in use within the interTwin project and DestinE initiative. Lastly, workflows to prioritise access to supercomputing resources known as ‘urgent computing’ were developed (likely for the first time in Europe) to predict the impact of natural hazards, particularly earthquakes and tsunamis. “Thanks to the dynamic flexibility of our solution, extensions of these workflows are now being used in the DT-GEO project, developing digital twins of physical systems to monitor and predict natural hazards,” adds Badia.
A springboard for next-generation workflows
The project was carried out with support from the European High Performance Computing Joint Undertaking (EuroHPC JU), an initiative set up to develop a world-class supercomputing ecosystem in Europe. All the open-source software developed by the project, alongside the demonstration workflows, have been made freely available. “We encourage any user community to access the software and start collaborating to extend its capabilities,” notes Jorge Ejarque, eFlows4HPC technical leader. To ensure the transfer of best practices and results gained from the project, training and workshops were conducted in Barcelona, Munich and Helsinki. The team continues to extend the features and capabilities of all the project’s workflows, focused on AI systems and digital twins.
Keywords
eFlows4HPC, EuroHPC JU, exascale, supercomputing, software, AI, digital twins, workflows, hazards, manufacturing, simulators, data processing, machine learning, Earth system, data analytics, HPC