🤖 AI Summary
Existing prototypical networks predominantly rely on Euclidean-space prototypes, which constrain semantic interpretability and structural flexibility. Method: We systematically survey and compare Euclidean versus non-Euclidean prototype representation paradigms, introducing— for the first time—a unified analytical framework that elucidates how prototype geometry governs interpretability. Our approach integrates prototype learning, differentiable attention-based localization, multi-granularity part matching, and cross-dataset generalization evaluation. Contribution/Results: Experiments on three fine-grained benchmarks—CUB-200-2011, Stanford Cars, and Oxford Flowers—demonstrate that non-Euclidean prototypes substantially improve the trade-off between model interpretability and classification accuracy. Specifically, they enhance part-level semantic alignment and out-of-domain generalization robustness, offering greater structural expressivity and principled geometric grounding for prototype-based representation learning.
📝 Abstract
Prototypical part networks offer interpretable alternatives to black-box deep learning models. However, many of these networks rely on Euclidean prototypes, which may limit their flexibility. This work provides a comprehensive overview of various prototype formulations. Experiments conducted on the CUB-200-2011, Stanford Cars, and Oxford Flowers datasets demonstrate the effectiveness and versatility of these different formulations.