🤖 AI Summary
To address the lack of comparable, empirically grounded metrics for assessing scholarly effort across subfields of computer science, this paper introduces “ICLR Points”—the first standardized, evidence-based metric for quantifying academic effort. Leveraging DBLP data on top-tier conference publications (ICLR, ICML, NeurIPS) from 2019–2023 and CSRankings faculty data, we employ statistical modeling and cross-domain normalization to estimate the average effort required to publish one paper in each of 27 subfields. Results show that systems-oriented subfields demand significantly higher effort per publication than AI-focused ones—aligning with expert consensus and validating the metric’s construct validity. ICLR Points have been deployed in cross-disciplinary faculty evaluation, enhancing fairness, interpretability, and practical applicability of academic assessment.
📝 Abstract
Scientific publications significantly impact academic-related decisions in computer science, where top-tier conferences are particularly influential. However, efforts required to produce a publication differ drastically across various subfields. While existing citation-based studies compare venues within areas, cross-area comparisons remain challenging due to differing publication volumes and citation practices. To address this gap, we introduce the concept of ICLR points, defined as the average effort required to produce one publication at top-tier machine learning conferences such as ICLR, ICML, and NeurIPS. Leveraging comprehensive publication data from DBLP (2019--2023) and faculty information from CSRankings, we quantitatively measure and compare the average publication effort across 27 computer science sub-areas. Our analysis reveals significant differences in average publication effort, validating anecdotal perceptions: systems conferences generally require more effort per publication than AI conferences. We further demonstrate the utility of the ICLR points metric by evaluating publication records of current faculties and recent faculty candidates. Our findings highlight how using this metric enables more meaningful cross-area comparisons in academic evaluation processes. Lastly, we discuss the metric's limitations and caution against its misuse, emphasizing the necessity of holistic assessment criteria beyond publication metrics alone.