🤖 AI Summary
Existing tensor network architecture search (TN-AS) methods suffer from high computational overhead, reliance on numerous function evaluations, neglect of domain-specific priors, and poor structural interpretability. Method: We propose tnLLM—a novel framework that tightly integrates domain knowledge with large language model (LLM) reasoning. It employs a domain-aware prompting mechanism to directly generate physically interpretable TN architectures aligned with real-world modality relationships; performs structured architecture selection and attribution analysis within a constrained search space; and provides theoretically grounded, high-quality initializations for sampling-based state-of-the-art (SOTA) optimizers. Contribution/Results: Experiments demonstrate that tnLLM reduces function evaluations by several orders of magnitude versus SOTA while preserving optimization performance. Moreover, it significantly enhances architectural transparency and initialization quality, establishing a new paradigm for knowledge-infused, interpretable TN search.
📝 Abstract
Tensor networks (TNs) provide efficient representations of high-dimensional data, yet identification of the optimal TN structures, the so called tensor network structure search (TN-SS) problem, remains a challenge. Current state-of-the-art (SOTA) algorithms are computationally expensive as they require extensive function evaluations, which is prohibitive for real-world applications. In addition, existing methods ignore valuable domain information inherent in real-world tensor data and lack transparency in their identified TN structures. To this end, we propose a novel TN-SS framework, termed the tnLLM, which incorporates domain information about the data and harnesses the reasoning capabilities of large language models (LLMs) to directly predict suitable TN structures. The proposed framework involves a domain-aware prompting pipeline which instructs the LLM to infer suitable TN structures based on the real-world relationships between tensor modes. In this way, our approach is capable of not only iteratively optimizing the objective function, but also generating domain-aware explanations for the identified structures. Experimental results demonstrate that tnLLM achieves comparable TN-SS objective function values with much fewer function evaluations compared to SOTA algorithms. Furthermore, we demonstrate that the LLM-enabled domain information can be used to find good initializations in the search space for sampling-based SOTA methods to accelerate their convergence while preserving theoretical performance guarantees.