Anatomy-Aware Contrastive Pre-training: Leveraging Spatial Consistency for Label-Efficient Medical Image Diagnosis Across Multi-Modal Imaging

Main Article Content

Mingxuan Han

Abstract

Medical image analysis faces persistent challenges in acquiring expert annotations due to high costs and specialized expertise requirements. Self-supervised learning offers a promising solution by learning representations from unlabeled data. This paper introduces an anatomy-aware contrastive pre-training framework that exploits spatial consistency and anatomical structure priors inherent in medical images. The proposed approach integrates contrastive learning with anatomical constraints, enabling effective knowledge transfer across CT, MRI, and X-ray modalities. Through comprehensive experiments on multiple diagnostic tasks, the framework demonstrates superior label efficiency, achieving competitive performance with only 10% of labeled data compared to fully supervised baselines. The cross-modal evaluation reveals consistent improvements of 8.3% in classification accuracy and 6.7% in segmentation Dice scores. These results validate the effectiveness of incorporating anatomical priors into self-supervised learning pipelines for medical imaging applications.

Article Details

Section

Articles

How to Cite

Anatomy-Aware Contrastive Pre-training: Leveraging Spatial Consistency for Label-Efficient Medical Image Diagnosis Across Multi-Modal Imaging. (2026). Journal of Sustainability, Policy, and Practice, 2(1), 55-70. https://schoalrx.com/index.php/jspp/article/view/77

References

1. S. Azizi, B. Mustafa, F. Ryan, Z. Beaver, J. Freyberg, J. Deaton, and M. Norouzi, "Big self-supervised models advance medical image classification," In Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 3478-3488.

2. F. Haghighi, M. R. H. Taher, M. B. Gotway, and J. Liang, "Dira: Discriminative, restorative, and adversarial learning for self-supervised medical image analysis," In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 20824-20834.

3. S. Azizi, L. Culp, J. Freyberg, B. Mustafa, S. Baur, S. Kornblith, and V. Natarajan, "Robust and efficient medical imaging with self-supervision," arXiv preprint arXiv:2205.09723, 2022.

4. S. C. Huang, A. Pareek, M. Jensen, M. P. Lungren, S. Yeung, and A. S. Chaudhari, "Self-supervised learning for medical image classification: a systematic review and implementation guidelines," NPJ Digital Medicine, vol. 6, no. 1, p. 74, 2023.

5. D. MH Nguyen, H. Nguyen, N. Diep, T. N. Pham, T. Cao, B. Nguyen, and M. Niepert, "Lvm-med: Learning large-scale self-supervised vision models for medical imaging via second-order graph matching," Advances in Neural Information Processing Systems, vol. 36, pp. 27922-27950, 2023.

6. Y. Ye, Y. Xie, J. Zhang, Z. Chen, Q. Wu, and Y. Xia, "Continual self-supervised learning: Towards universal multi-modal medical data representation learning," In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2024, pp. 11114-11124. doi: 10.1109/cvpr52733.2024.01057

7. Z. Huang, R. Jiang, S. Aeron, and M. C. Hughes, "Systematic comparison of semi-supervised and self-supervised learning for medical image classification," In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 22282-22293. doi: 10.1109/cvpr52733.2024.02103

8. W. C. Wang, E. Ahn, D. Feng, and J. Kim, "A review of predictive and contrastive self-supervised learning for medical images," Machine Intelligence Research, vol. 20, no. 4, pp. 483-513, 2023.

9. L. Zhou, H. Liu, J. Bae, J. He, D. Samaras, and P. Prasanna, "Self pre-training with masked autoencoders for medical image classification and segmentation," In 2023 IEEE 20th international symposium on biomedical imaging (ISBI), April, 2023, pp. 1-6. doi: 10.1109/isbi53787.2023.10230477

10. L. Wu, J. Zhuang, and H. Chen, "Voco: A simple-yet-effective volume contrastive learning framework for 3d medical image analysis," In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2024, pp. 22873-22882. doi: 10.1109/cvpr52733.2024.02158

11. W. Huang, C. Li, H. Y. Zhou, H. Yang, J. Liu, Y. Liang, and S. Wang, "Enhancing representation in radiography-reports foundation model: A granular alignment algorithm using masked contrastive learning," Nature Communications, vol. 15, no. 1, p. 7620, 2024.

12. M. Nielsen, L. Wenderoth, T. Sentker, and R. Werner, "Self-supervision for medical image classification: State-of-the-art performance with~ 100 labeled training samples per class," Bioengineering, vol. 10, no. 8, p. 895, 2023. doi: 10.3390/bioengineering10080895

13. Z. Li, L. T. Yang, B. Ren, X. Nie, Z. Gao, C. Tan, and S. Z. Li, "Mlip: Enhancing medical visual representation with divergence encoder and knowledge-guided contrastive learning," In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 11704-11714. doi: 10.1109/cvpr52733.2024.01112

14. J. Mao, S. Guo, X. Yin, Y. Chang, B. Nie, and Y. Wang, "Medical supervised masked autoencoder: Crafting a better masking strategy and efficient fine-tuning schedule for medical image classification," Applied Soft Computing, vol. 169, p. 112536, 2025. doi: 10.1016/j.asoc.2024.112536

15. S. Albelwi, "Survey on self-supervised learning: Auxiliary pretext tasks and contrastive learning methods in imaging," Entropy, vol. 24, no. 4, p. 551, 2022. doi: 10.3390/e24040551