🤖 AI Summary
This work addresses the lack of standardized adversarial robustness evaluation for genomic foundation models (GFMs) by introducing GERM, the first unified adversarial benchmark for GFMs. Methodologically, it systematically evaluates five state-of-the-art GFMs under four adversarial attack algorithms—including PGD and FGSM—and three defense strategies, integrating quantitative robustness analysis, architectural comparison, and training data provenance to establish a reproducible, fine-grained evaluation framework. Key contributions include: (1) the first empirical demonstration that Transformer-based architectures significantly outperform alternatives such as HyenaDNA in adversarial robustness; (2) the discovery that adversarial perturbations consistently target biologically functional regions—particularly promoters and enhancers—providing evidence that GFMs have learned semantically meaningful genomic representations; and (3) the release of a multidimensional vulnerability attribution toolkit enabling disentangled robustness analysis across architectural design, quantization schemes, and training data composition.
📝 Abstract
We propose the first unified adversarial attack benchmark for Genomic Foundation Models (GFMs), named GERM. Unlike existing GFM benchmarks, GERM offers the first comprehensive evaluation framework to systematically assess the vulnerability of GFMs to adversarial attacks. Methodologically, we evaluate the adversarial robustness of five state-of-the-art GFMs using four widely adopted attack algorithms and three defense strategies. Importantly, our benchmark provides an accessible and comprehensive framework to analyze GFM vulnerabilities with respect to model architecture, quantization schemes, and training datasets. Empirically, transformer-based models exhibit greater robustness to adversarial perturbations compared to HyenaDNA, highlighting the impact of architectural design on vulnerability. Moreover, adversarial attacks frequently target biologically significant genomic regions, suggesting that these models effectively capture meaningful sequence features.