Agent-centric Information Access

📅 2025-02-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
As millions of specialized large language models (LLMs) coexist in future AI ecosystems, efficient, scalable, and dynamic selection of domain-expert models—beyond static metadata—is a critical challenge. Method: This paper proposes an agent-centric information access framework that models LLMs as dynamic knowledge agents. It enables online expert model ranking and selection via real-time inference of domain expertise, integrating retrieval-augmented generation (RAG), hierarchical model clustering, dynamic expert ranking, and multi-agent response aggregation—balancing query cost, robustness, and scalability. Contribution/Results: The work introduces the novel “agent-level retrieval” paradigm, enabling coordinated scheduling across millions of heterogeneous LLMs. Experiments across diverse domains demonstrate significant improvements in answer accuracy and resource utilization efficiency. The framework provides a scalable infrastructure for evaluation and orchestration in large-scale specialized model ecosystems.

Technology Category

Application Category

📝 Abstract
As large language models (LLMs) become more specialized, we envision a future where millions of expert LLMs exist, each trained on proprietary data and excelling in specific domains. In such a system, answering a query requires selecting a small subset of relevant models, querying them efficiently, and synthesizing their responses. This paper introduces a framework for agent-centric information access, where LLMs function as knowledge agents that are dynamically ranked and queried based on their demonstrated expertise. Unlike traditional document retrieval, this approach requires inferring expertise on the fly, rather than relying on static metadata or predefined model descriptions. This shift introduces several challenges, including efficient expert selection, cost-effective querying, response aggregation across multiple models, and robustness against adversarial manipulation. To address these issues, we propose a scalable evaluation framework that leverages retrieval-augmented generation and clustering techniques to construct and assess thousands of specialized models, with the potential to scale toward millions.
Problem

Research questions and friction points this paper is trying to address.

Dynamic selection of relevant expert LLMs
Efficient querying and response synthesis
Robustness against adversarial manipulation in model aggregation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic ranking of expert LLMs
Retrieval-augmented generation for scalability
Clustering techniques for model assessment
Evangelos Kanoulas
Evangelos Kanoulas
University of Amsterdam
Information RetrievalNatural Language ProcessingMachine Learning
Panagiotis Eustratiadis
Panagiotis Eustratiadis
University of Amsterdam
AI SafetyInformation RetrievalGenerative AI
Y
Yongkang Li
University of Amsterdam, The Netherlands
Yougang Lyu
Yougang Lyu
University of Amsterdam
Natural Language ProcessingLarge Language ModelsInformation Retrieval
V
Vaishali Pal
University of Amsterdam, The Netherlands
G
Gabrielle Poerwawinata
University of Amsterdam, The Netherlands
J
Jingfen Qiao
University of Amsterdam, The Netherlands
Z
Zihan Wang
University of Amsterdam, The Netherlands