🤖 AI Summary
Solving large-scale, ill-conditioned, and indefinite sparse linear systems is critical in multiphysics simulation, machine learning, and related fields, yet poses significant challenges in computational efficiency and scalability. This work proposes an optimized sparse direct solver framework that integrates task and data parallelism with low-rank approximation and hierarchical matrix compression techniques. Implemented on heterogeneous high-performance computing platforms, the approach substantially reduces communication overhead and computational complexity while preserving the numerical robustness inherent to direct solvers. The resulting method achieves marked improvements in both strong and weak scalability, delivering a highly efficient, reliable, and scalable linear solver toolchain tailored for modern heterogeneous architectures.
📝 Abstract
Efficient solutions of large-scale, ill-conditioned and indefinite algebraic equations are ubiquitously needed in numerous computational fields, including multiphysics simulations, machine learning, and data science. Because of their robustness and accuracy, direct solvers are crucial components in building a scalable solver toolchain. In this article, we will review recent advances of sparse direct solvers along two axes: 1) reducing communication and latency costs in both task- and data-parallel settings, and 2) reducing computational complexity via low-rank and other compression techniques such as hierarchical matrix algebra. In addition to algorithmic principles, we also illustrate the key parallelization challenges and best practices to deliver high speed and reliability on modern heterogeneous parallel machines.