A Proposed Framework for Advanced (Multi)Linear Infrastructure in Engineering and Science (FAMLIES)

📅 2026-04-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the lack of a unified and efficient framework for linear and multilinear (tensor) computations across diverse hardware architectures—from single-node CPUs to large-scale parallel systems with GPU acceleration. To bridge this gap, the authors propose a vertically integrated software stack that systematically combines the FLAME algorithmic derivation methodology with established high-performance libraries and runtime systems, including BLIS, TBLIS, SuperMatrix, and Elemental. This integration yields, for the first time, a cohesive infrastructure capable of delivering high performance for both linear algebra and tensor operations across heterogeneous platforms. The resulting framework supports single-core, multi-core, GPU-accelerated, and distributed-memory environments, offering both flexibility and scalability while providing a robust foundation for scientific computing and machine learning applications.
📝 Abstract
We leverage highly successful prior projects sponsored by multiple NSF grants and gifts from industry: the BLAS-like Library Instantiation Software (BLIS) and the libflame efforts to lay the foundation for a new flexible framework by vertically integrating the dense linear and multi-linear (tensor) software stacks that are important to modern computing. This vertical integration will enable high-performance computations from node-level to massively-parallel, and across both CPU and GPU architectures. The effort builds on decades of experience by the research team turning fundamental research on the systematic derivation of algorithms (the NSF-sponsored FLAME project) into practical software for this domain, targeting single and multi-core (BLIS, TBLIS, and libflame), GPU-accelerated (SuperMatrix), and massively parallel (PLAPACK, Elemental, and ROTE) compute environments. This project will implement key linear algebra and tensor operations which highlight the flexibility and effectiveness of the new framework, and set the stage for further work in broadening functionality and integration into diverse scientific and machine learning software.
Problem

Research questions and friction points this paper is trying to address.

linear algebra
tensor operations
high-performance computing
software framework
multi-linear infrastructure
Innovation

Methods, ideas, or system contributions that make the work stand out.

vertical integration
tensor operations
high-performance computing
linear algebra libraries
multi-architecture support
🔎 Similar Papers
No similar papers found.
D
Devin A. Matthews
Department of Chemistry, Southern Methodist University
Tze Meng Low
Tze Meng Low
Carnegie Mellon University
High performanceFormal methodsProgram Generation
M
Margaret E. Myers
Department of Computer Science and Oden Institute for Computational Engineering and Sciences, The University of Texas at Austin
D
Devangi N. Parikh
Department of Computer Science and Oden Institute for Computational Engineering and Sciences, The University of Texas at Austin
R
Robert A. van de Geijn
Department of Computer Science and Oden Institute for Computational Engineering and Sciences, The University of Texas at Austin