🤖 AI Summary
In health technology assessment (HTA), matching-adjusted indirect comparison (MAIC) can yield contradictory efficacy conclusions for the same evidence—termed the “MAIC paradox”—due to sponsor-specific, implicit definitions of the target population. Method: We propose Arbitrated Indirect Comparison (AIC), a novel framework that explicitly defines and focuses on the *inter-trial covariate overlap population* as a unified, objective target population, thereby eliminating subjective population specification bias. AIC integrates individual participant data (IPD) and aggregate data using covariate-balancing reweighting to produce unbiased, cross-trial comparable treatment effect estimates. Contribution/Results: Simulation and empirical analyses demonstrate that AIC substantially improves consistency, robustness, and fairness across multiple stakeholder analyses. It provides a reproducible, transparent, and methodologically grounded benchmark for indirect comparisons in HTA.
📝 Abstract
Matching-adjusted indirect comparison (MAIC) has been increasingly employed in health technology assessments (HTA). By reweighting subjects from a trial with individual participant data (IPD) to match the covariate summary statistics of another trial with only aggregate data (AgD), MAIC facilitates the estimation of a treatment effect defined with respect to the AgD trial population. This manuscript introduces a new class of methods, termed arbitrated indirect treatment comparisons, designed to address the ``MAIC paradox'' -- a phenomenon highlighted by Jiang et al.~(2025). The MAIC paradox arises when different sponsors, analyzing the same data, reach conflicting conclusions regarding which treatment is more effective. The underlying issue is that each sponsor implicitly targets a different population. To resolve this inconsistency, the proposed methods focus on estimating treatment effects in a common target population, specifically chosen to be the overlap population.