Symbolic Representation for Any-to-Any Generative Tasks

📅 2025-04-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multimodal generative tasks typically rely on large-scale training and implicit neural representations, resulting in high computational costs, poor editability, and limited interruptibility. Method: This paper proposes an explicit symbolic task modeling paradigm. It introduces a modality-agnostic symbolic task description language and a structured reasoning engine that zero-shot parses natural language instructions into symbolic streams—comprising functions, parameters, and topological logic—without fine-tuning or task-specific training, leveraging only pre-trained large language models for instruction understanding and workflow orchestration. Contribution/Results: The approach significantly enhances interpretability, real-time editability, and interruption recovery in cross-modal generation. Evaluated on 12 heterogeneous multimodal generation tasks, it achieves or surpasses state-of-the-art performance while reducing computational overhead.

Technology Category

Application Category

📝 Abstract
We propose a symbolic generative task description language and a corresponding inference engine capable of representing arbitrary multimodal tasks as structured symbolic flows. Unlike conventional generative models that rely on large-scale training and implicit neural representations to learn cross-modal mappings, often at high computational cost and with limited flexibility, our framework introduces an explicit symbolic representation comprising three core primitives: functions, parameters, and topological logic. Leveraging a pre-trained language model, our inference engine maps natural language instructions directly to symbolic workflows in a training-free manner. Our framework successfully performs over 12 diverse multimodal generative tasks, demonstrating strong performance and flexibility without the need for task-specific tuning. Experiments show that our method not only matches or outperforms existing state-of-the-art unified models in content quality, but also offers greater efficiency, editability, and interruptibility. We believe that symbolic task representations provide a cost-effective and extensible foundation for advancing the capabilities of generative AI.
Problem

Research questions and friction points this paper is trying to address.

Proposes symbolic language for any-to-any generative tasks
Replaces implicit neural representations with structured symbolic flows
Enables training-free multimodal task execution via natural language
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symbolic representation for multimodal generative tasks
Training-free mapping of instructions to workflows
Efficient, editable, and interruptible generative framework
🔎 Similar Papers
No similar papers found.
J
Jiaqi Chen
Stanford University, Fellou AI, Fudan University
X
Xiaoye Zhu
South China University of Technology
Y
Yue Wang
Cornell University
Tianyang Liu
Tianyang Liu
Ph.D. in Computer Science, UC San Diego
artificial intelligencelarge language modelscode intelligence
X
Xinhui Chen
Fenz.AI, Wuhan University
Y
Ying Chen
University of Illinois at Urbana-Champaign
C
Chak Tou Leong
Hong Kong Polytechnic University
Y
Yifei Ke
Wuhan University
J
Joseph Liu
University of Southern California
Y
Yiwen Yuan
Carnegie Mellon University
Julian McAuley
Julian McAuley
Professor, UC San Diego
Recommender SystemsNatural Language ProcessingPersonalizationComputer Music
L
Li-jia Li
LiveX AI