PerturBench: Benchmarking Machine Learning Models for Cellular Perturbation Analysis

📅 2024-08-20
🏛️ arXiv.org
📈 Citations: 7
Influential: 1
📄 PDF
🤖 AI Summary
Lack of standardized evaluation protocols for single-cell perturbation effect prediction hinders model comparability and biological interpretability. To address this, we introduce ScPEBench—the first standardized benchmark framework—integrating diverse CRISPR-based and drug perturbation datasets, an accessible benchmarking platform, and a multi-dimensional evaluation suite balancing reconstruction accuracy (e.g., RMSE) and ranking fidelity (e.g., rank-based consistency). Through systematic, reproducible integration testing, we uncover pervasive issues—including mode collapse—and demonstrate that simple models often outperform complex ones, underscoring the critical importance of ranking metrics alongside conventional error measures. ScPEBench establishes a new evaluation paradigm, significantly enhancing model robustness and cross-study comparability. It provides a rigorous, reproducible foundation for genetic and chemical screening–driven target discovery.

Technology Category

Application Category

📝 Abstract
We present a comprehensive framework for predicting the effects of perturbations in single cells, designed to standardize benchmarking in this rapidly evolving field. Our framework, PerturBench, includes a user-friendly platform, diverse datasets, metrics for fair model comparison, and detailed performance analysis. Extensive evaluations of published and baseline models reveal limitations like mode or posterior collapse, and underscore the importance of rank metrics that assess the ordering of perturbations alongside traditional measures like RMSE. Our findings show that simple models can outperform more complex approaches. This benchmarking exercise sets new standards for model evaluation, supports robust model development, and advances the potential of these models to use high-throughput and high-content genetic and chemical screens for disease target discovery.
Problem

Research questions and friction points this paper is trying to address.

Standardizing benchmarking for cellular perturbation response modeling
Comparing model performance fairly using diverse datasets and metrics
Advancing disease target discovery via high-throughput perturbation screens
Innovation

Methods, ideas, or system contributions that make the work stand out.

Modular platform for model development and evaluation
Diverse perturbational datasets for comprehensive benchmarking
Rank metrics complement traditional model fit measures
🔎 Similar Papers
No similar papers found.
Y
Yan Wu
Altos Labs, San Diego, US
E
Esther Wershof
Altos Labs, Cambridge, UK
S
Sebastian M. Schmon
Altos Labs, Cambridge, UK
Marcel Nassar
Marcel Nassar
Altos Labs, San Diego, US
B
Bla.zej Osi'nski
Altos Labs, Cambridge, UK
R
Ridvan Eksi
Altos Labs, San Diego, US
K
Kun Zhang
Altos Labs, San Diego, US
T
T. Graepel
Altos Labs, Cambridge, UK