🤖 AI Summary
This work addresses the challenge of achieving geometric invariance and representation generalizability in quantitative 2D object contour shape analysis. We propose a self-supervised contrastive learning framework for shape descriptor learning. Methodologically, we encode contour structure via Euclidean distance matrices and design a deep embedding mechanism invariant to translation, scaling, rotation, reflection, and point indexing—without relying on explicit data augmentation or handcrafted priors. Compared with conventional shape descriptors and existing autoencoder-based approaches, our framework preserves strict geometric invariance while significantly improving discriminability and robustness of shape representations. Extensive experiments on multi-source natural and biological image datasets demonstrate that our method achieves state-of-the-art performance in shape classification tasks, validating its strong generalization capability and practical applicability.
📝 Abstract
The shape of objects is an important source of visual information in a wide range of applications. One of the core challenges of shape quantification is to ensure that the extracted measurements remain invariant to transformations that preserve an object's intrinsic geometry, such as changing its size, orientation, and position in the image. In this work, we introduce ShapeEmbed, a self-supervised representation learning framework designed to encode the contour of objects in 2D images, represented as a Euclidean distance matrix, into a shape descriptor that is invariant to translation, scaling, rotation, reflection, and point indexing. Our approach overcomes the limitations of traditional shape descriptors while improving upon existing state-of-the-art autoencoder-based approaches. We demonstrate that the descriptors learned by our framework outperform their competitors in shape classification tasks on natural and biological images. We envision our approach to be of particular relevance to biological imaging applications.