π€ AI Summary
This work proposes PEP U-Net to address the inherent uncertainties in medical image segmentation caused by image blur, noise, and subjective annotation. The method integrates principal component analysis (PCA) and inverse PCA into a conditional variational autoencoder (cVAE) framework, embedding them within the posterior network of a probabilistic U-Net to achieve latent space dimensionality reduction and reconstruction of essential features. This design effectively mitigates redundancy in high-dimensional latent representations while preserving the modelβs capacity to generate diverse segmentation hypotheses. As a result, PEP U-Net achieves a superior balance between segmentation accuracy and predictive uncertainty, outperforming existing generative segmentation models.
π Abstract
Ambiguous Medical Image Segmentation (AMIS) is significant to address the challenges of inherent uncertainties from image ambiguities, noise, and subjective annotations. Existing conditional variational autoencoder (cVAE)-based methods effectively capture uncertainty but face limitations including redundancy in high-dimensional latent spaces and limited expressiveness of single posterior networks. To overcome these issues, we introduce a novel PCA-Enhanced Probabilistic U-Net (\textbf{PEP U-Net}). Our method effectively incorporates Principal Component Analysis (PCA) for dimensionality reduction in the posterior network to mitigate redundancy and improve computational efficiency. Additionally, we further employ an inverse PCA operation to reconstruct critical information, enhancing the latent space's representational capacity. Compared to conventional generative models, our method preserves the ability to generate diverse segmentation hypotheses while achieving a superior balance between segmentation accuracy and predictive variability, thereby advancing the performance of generative modeling in medical image segmentation.