ranx: A Blazing-Fast Python Library for Ranking Evaluation and Comparison

πŸ“… 2025-03-04
πŸ›οΈ European Conference on Information Retrieval
πŸ“ˆ Citations: 4
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the dual challenges of accuracy and fairness in expert user identification within community question-answering platforms, this paper introduces ranxβ€”a high-performance ranking evaluation library. Implemented entirely in NumPy and Cython, ranx features a novel sub-millisecond parallel multi-metric computation architecture, enabling pairwise and multi-group statistical significance testing (e.g., Wilcoxon signed-rank test, t-test) across arbitrary numbers of rankers. Through memory mapping and vectorized logic, it achieves extreme computational efficiency. On TREC and MSMARCO benchmarks, ranx outperforms ir-measures and prior versions of ranx by 10–100Γ—, supporting real-time evaluation over million-scale queries. Its core innovation lies in the tight integration of ranking evaluation, statistical inference, and systems-level engineering optimization. As an open-source infrastructure, ranx provides both flexibility and scalability, advancing reproducible and verifiable research in information retrieval and recommender systems.

Technology Category

Application Category

Problem

Research questions and friction points this paper is trying to address.

Improves expert identification in CQA platforms
Enhances robustness using content and social data
Promotes transparent and fair expert finding
Innovation

Methods, ideas, or system contributions that make the work stand out.

TUEF model uses content and social information
Improves expert identification robustness and credibility
Outperforms competitors in reproducible StackOverflow experiments
πŸ”Ž Similar Papers