Connecting Large Language Model Agent to High Performance Computing Resource

📅 2025-02-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of deep integration between large language model (LLM) agents and high-performance computing (HPC) resources. We propose a novel architecture that tightly couples the Parsl workflow system with the LangChain/LangGraph tool-calling framework, enabling LLMs to autonomously generate, schedule, and execute parallelized scientific computing tasks—such as multi-condition protein molecular dynamics simulations—directly from natural language instructions. We introduce two parallel execution paradigms tailored to distinct computational scales: queue-based concurrent invocation for local clusters and ensemble-functional supercomputing-optimized invocation for large-scale HPC systems. End-to-end validation is conducted on both a local workstation and the ALCF Polaris supercomputer. Results demonstrate that our approach enables LLMs to autonomously orchestrate heterogeneous HPC resources, manage concurrent workloads, and achieve efficient resource utilization—thereby significantly advancing automation and accessibility in scientific computing.

Technology Category

Application Category

📝 Abstract
The Large Language Model agent workflow enables the LLM to invoke tool functions to increase the performance on specific scientific domain questions. To tackle large scale of scientific research, it requires access to computing resource and parallel computing setup. In this work, we implemented Parsl to the LangChain/LangGraph tool call setup, to bridge the gap between the LLM agent to the computing resource. Two tool call implementations were set up and tested on both local workstation and HPC environment on Polaris/ALCF. The first implementation with Parsl-enabled LangChain tool node queues the tool functions concurrently to the Parsl workers for parallel execution. The second configuration is implemented by converting the tool functions into Parsl ensemble functions, and is more suitable for large task on super computer environment. The LLM agent workflow was prompted to run molecular dynamics simulations, with different protein structure and simulation conditions. These results showed the LLM agent tools were managed and executed concurrently by Parsl on the available computing resource.
Problem

Research questions and friction points this paper is trying to address.

Bridging LLM agents to HPC resources
Enabling parallel scientific computations
Optimizing molecular dynamics simulations
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM agent invokes tool functions
Parsl bridges LLM to HPC
Parallel execution enhances performance
🔎 Similar Papers
No similar papers found.
Heng Ma
Heng Ma
Argonne National Laboratory
Physical ChemistryBiophysicsMachine Learning
A
Alexander Brace
Data Science and Learning Division, Argonne National Laboratory, Lemont, IL, USA; Computer Science Department, University of Chicago, Chicago, IL, USA
C
Carlo Siebenschuh
Data Science and Learning Division, Argonne National Laboratory, Lemont, IL, USA; Computer Science Department, University of Chicago, Chicago, IL, USA
G
Gregory Pauloski
Computer Science Department, University of Chicago, Chicago, IL, USA
Ian T. Foster
Ian T. Foster
University of Chicago and Argonne National Laboratory
Computer sciencecomputational sciencedistributed computingdata science
Arvind Ramanathan
Arvind Ramanathan
Argonne National Laboratory
Machine LearningComputational BiologyMolecular biophysicsenzyme catalysishigher-order statistics