MoMoE: A Mixture of Expert Agent Model for Financial Sentiment Analysis

📅 2025-11-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitations in fine-grained semantic modeling and domain adaptability in financial sentiment analysis, this paper proposes MoMoE—a novel architecture integrating Mixture of Experts (MoE) with hierarchical multi-agent collaboration. Methodologically, MoMoE unifies expert routing across both neural network layers (embedding MoE in the final attention layer of each agent within LLaMA-3.1 8B) and inter-agent interaction layers, establishing cross-level specialized collaboration pathways. It achieves dual-level (structural and behavioral) coordination via task decomposition, iterative optimization, and hierarchical information fusion. Experimental results demonstrate that MoMoE significantly improves accuracy and stability across multiple financial sentiment analysis benchmarks, outperforming state-of-the-art monolithic models and conventional multi-agent approaches. These findings validate the efficacy of MoE-driven multi-agent collaborative paradigms for domain-specific language understanding.

Technology Category

Application Category

📝 Abstract
We present a novel approach called Mixture of Mixture of Expert (MoMoE) that combines the strengths of Mixture-of-Experts (MoE) architectures with collaborative multi-agent frameworks. By modifying the LLaMA 3.1 8B architecture to incorporate MoE layers in each agent of a layered collaborative structure, we create an ensemble of specialized expert agents that iteratively refine their outputs. Each agent leverages an MoE layer in its final attention block, enabling efficient task decomposition while maintaining computational feasibility. This hybrid approach creates specialized pathways through both the model architecture and the agent collaboration layers. Experimental results demonstrate significant improvements across multiple language understanding and generation benchmarks, highlighting the synergistic benefits of combining expert routing at both the neural and agent levels.
Problem

Research questions and friction points this paper is trying to address.

Proposes MoMoE model combining MoE architectures with multi-agent frameworks
Enhances financial sentiment analysis through specialized expert agent collaboration
Improves language understanding benchmarks via neural and agent-level expert routing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Mixture of Mixture of Expert combines MoE with multi-agent frameworks
Modified LLaMA 3.1 8B architecture incorporates MoE layers
Specialized expert agents iteratively refine outputs through collaboration
🔎 Similar Papers
No similar papers found.
P
Peng Shu
GyriFin Interest Group on Finance Foundation Models
J
Junhao Chen
GyriFin Interest Group on Finance Foundation Models
Zhengliang Liu
Zhengliang Liu
University of Georgia
Natural Language ProcessingMedical NLPMedical Image AnalysisData Visualization
Hanqi Jiang
Hanqi Jiang
University of Georgia
Medical Image AnalysisMulti-modal Large Language Models
Y
Yi Pan
GyriFin Interest Group on Finance Foundation Models
K
Khanh Nhu Nguyen
GyriFin Interest Group on Finance Foundation Models
Zihao Wu
Zihao Wu
University of Georgia
Brain-inspired AIArtificial General IntelligenceNLPMedical Image Analysis
H
Huaqin Zhao
GyriFin Interest Group on Finance Foundation Models
Y
Yiwei Li
GyriFin Interest Group on Finance Foundation Models
E
Enze Shi
GyriFin Interest Group on Finance Foundation Models
S
ShaoChen Xu
GyriFin Interest Group on Finance Foundation Models