ReSi: A Comprehensive Benchmark for Representational Similarity Measures

πŸ“… 2024-08-01
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 3
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Representation similarity measurement lacks rigorous formal definitions and systematic evaluationβ€”a fundamental open problem in machine learning. This paper introduces ReSi, the first comprehensive benchmark grounded in a formal definition of representation similarity. ReSi encompasses six test categories, 23 similarity metrics (including CKA, CCA, and SVCCA), 11 neural architectures, and six cross-modal datasets spanning vision, language, and graph domains. It enables multi-dimensional, cross-architecture, and cross-dataset comparability of representations. Methodologically, we propose a structured, verifiable, and cross-domain extensible benchmarking framework, establishing a unified testing paradigm and standardized evaluation protocols to fill the longstanding gap in systematic assessment. Empirical results comprehensively expose performance disparities among existing metrics across tasks and architectures. All code, data, and protocols are publicly released to ensure reproducibility and foster community-driven advancement.

Technology Category

Application Category

πŸ“ Abstract
Measuring the similarity of different representations of neural architectures is a fundamental task and an open research challenge for the machine learning community. This paper presents the first comprehensive benchmark for evaluating representational similarity measures based on well-defined groundings of similarity. The representational similarity (ReSi) benchmark consists of (i) six carefully designed tests for similarity measures, (ii) 23 similarity measures, (iii) eleven neural network architectures, and (iv) six datasets, spanning over the graph, language, and vision domains. The benchmark opens up several important avenues of research on representational similarity that enable novel explorations and applications of neural architectures. We demonstrate the utility of the ReSi benchmark by conducting experiments on various neural network architectures, real world datasets and similarity measures. All components of the benchmark are publicly available and thereby facilitate systematic reproduction and production of research results. The benchmark is extensible, future research can build on and further expand it. We believe that the ReSi benchmark can serve as a sound platform catalyzing future research that aims to systematically evaluate existing and explore novel ways of comparing representations of neural architectures.
Problem

Research questions and friction points this paper is trying to address.

Evaluating similarity measures for neural architectures
Providing a comprehensive benchmark for representational similarity
Enabling systematic comparison of neural network representations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Comprehensive benchmark for neural representation similarity
Includes tests, measures, architectures, and datasets
Publicly available and extensible for future research
πŸ”Ž Similar Papers
M
Max Klabunde
University of Passau
Tassilo Wald
Tassilo Wald
PhD Student, Deutsche Krebsforschungszentrum (DKFZ)
representation learningself-supervised learningmedical image analysis
T
Tobias Schumacher
University of Mannheim
K
K. Maier-Hein
Division of Medical Image Computing, German Cancer Research Center (DKFZ), Helmholtz Imaging, DKFZ, University of Heidelberg, Heidelberg University Hospital, National Center for Tumor Diseases (NCT) Heidelberg
Markus Strohmaier
Markus Strohmaier
University of Mannheim, GESIS - Leibniz Institute for the Social Sciences, CSH Vienna
Computational Social ScienceComputational Social SystemsSocial Data Science
Florian Lemmerich
Florian Lemmerich
University of Passau
Data Science(Interpretable) Machine LearningPattern MiningWeb MiningComputational Social Science