🤖 AI Summary
This paper addresses the tension between ad embedding and revenue optimization in large language model (LLM) text generation. Methodologically, it introduces a paragraph-level ad auction mechanism that integrates retrieval-augmented generation (RAG) with game-theoretic auction design. It formally defines a log-social-welfare maximization objective—uniquely balancing allocative efficiency and fairness—and constructs an incentive-compatible framework for joint multi-ad allocation and dynamic pricing, supporting probabilistic ad retrieval and paragraph-level relevance modeling. Theoretically, the mechanism is proven to satisfy incentive compatibility and achieve log-social-welfare optimality. Empirically, it significantly improves the trade-off between advertising revenue and content relevance across diverse scenarios, uncovering critical insights into flexibility–performance trade-offs and metric balancing.
📝 Abstract
In the field of computational advertising, the integration of ads into the outputs of large language models (LLMs) presents an opportunity to support these services without compromising content integrity. This paper introduces novel auction mechanisms for ad allocation and pricing within the textual outputs of LLMs, leveraging retrieval-augmented generation (RAG). We propose a segment auction where an ad is probabilistically retrieved for each discourse segment (paragraph, section, or entire output) according to its bid and relevance, following the RAG framework, and priced according to competing bids. We show that our auction maximizes logarithmic social welfare, a new notion of welfare that balances allocation efficiency and fairness, and we characterize the associated incentive-compatible pricing rule. These results are extended to multi-ad allocation per segment. An empirical evaluation validates the feasibility and effectiveness of our approach over several ad auction scenarios, and exhibits inherent tradeoffs in metrics as we allow the LLM more flexibility to allocate ads.