Organ-Agnostic Whole Slide Image Analysis using Self-Supervised Transfer Learning
DOI:
https://doi.org/10.47611/jsrhs.v13i1.6034Keywords:
Pathology Image, Autoencoder, ClassificationAbstract
Traditional methods for pathology image analysis are well-known for their time-consuming and labor-intensive nature, often relying on the expertise of pathologists. In recent years, numerous research studies have been proposed to develop automated systems using machine learning approaches to address these challenges. While these systems have demonstrated promising performance, they often exhibit bias towards specific organs, cells, or tasks, limiting their ability to provide generalized solutions for pathology image analysis. To address this issue, I propose an organ-agnostic pathology image analysis system that leverages a self-supervised transfer learning approach. The proposed system comprises two stages: self-supervised representation learning and transfer learning. In the self-supervised representation learning phase, a machine learning model is trained to consistently extract essential features encapsulating the characteristics of diverse pathological images such as visual patterns of tumors. Subsequently, in the transfer learning phase, these well-pretrained models are utilized to train downstream tasks, such as tumor type classification or cancer area segmentation. The proposed approach outperforms all existing state-of-the-art supervised methods in multiple public pathology image benchmarks.
Downloads
References or Bibliography
Chen, L. C., Papandreou, G., Schroff, F., & Adam, H. (2017). Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587. https://doi.org/10.48550/arXiv.1706.05587
Cruz-Roa, A., Basavanhally, A., González, F., Gilmore, H., Feldman, M., Ganesan, S., ... & Madabhushi, A. (2014, March). Automatic detection of invasive ductal carcinoma in whole slide images with convolutional neural networks. In Medical Imaging 2014: Digital Pathology (Vol. 9041, p. 904103). SPIE. https://doi.org/10.1117/12.2043872
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778). https://doi.org/10.48550/arXiv.1512.03385
Kaczmarzyk, J. R., Gupta, R., Kurc, T. M., Abousamra, S., Saltz, J. H., & Koo, P. K. (2023). ChampKit: a framework for rapid evaluation of deep neural networks for patch-based histopathology classification. Computer Methods and Programs in Biomedicine, 107631. https://doi.org/10.1016/j.cmpb.2023.107631
Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980. https://doi.org/10.48550/arXiv.1412.6980
Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25.
Lin, G., Milan, A., Shen, C., & Reid, I. (2017). Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1925-1934). https://doi.org/10.48550/arXiv.1611.06612
Mahbod, A., Schaefer, G., Bancher, B., Löw, C., Dorffner, G., Ecker, R., & Ellinger, I. (2021). CryoNuSeg: A dataset for nuclei instance segmentation of cryosectioned H&E-stained histological images. Computers in biology and medicine, 132, 104349. https://doi.org/10.1016/j.compbiomed.2021.104349
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. (2018). Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4510-4520). https://doi.org/10.48550/arXiv.1801.04381
Šarić, M., Russo, M., Stella, M., & Sikora, M. (2019, June). CNN-based method for lung cancer detection in whole slide histopathology images. In 2019 4th International Conference on Smart and Sustainable Technologies (SpliTech) (pp. 1-4). IEEE. https://doi.org/10.23919/SpliTech.2019.8783041
Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. https://doi.org/10.48550/arXiv.1409.1556
Tan, M., & Le, Q. (2019, May). Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning (pp. 6105-6114). PMLR. https://doi.org/10.48550/arXiv.1905.11946
Veeling, B. S., Linmans, J., Winkens, J., Cohen, T., & Welling, M. (2018). Rotation equivariant CNNs for digital pathology. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part II 11 (pp. 210-218). Springer International Publishing. https://doi.org/10.48550/arXiv.1806.03962
Wang, G., Luo, P., Lin, L., & Wang, X. (2017). Learning object interactions and descriptions for semantic image segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5859-5867). https://doi.org/10.1109/CVPR.2017.556
Wang, J., Sun, K., Cheng, T., Jiang, B., Deng, C., Zhao, Y., ... & Xiao, B. (2020). Deep high-resolution representation learning for visual recognition. IEEE transactions on pattern analysis and machine intelligence, 43(10), 3349-3364. https://doi.org/10.48550/arXiv.1908.07919
Wang, J., Xu, Z., Pang, Z. F., Huo, Z., & Luo, J. (2021). Tumor detection for whole slide image of liver based on patch-based convolutional neural network. Multimedia Tools and Applications, 80, 17429-17440. https://doi.org/10.1007/s11042-020-09282-x
Wu, Z., Shen, C., & Van Den Hengel, A. (2019). Wider or deeper: Revisiting the resnet model for visual recognition. Pattern Recognition, 90, 119-133. https://doi.org/10.48550/arXiv.1611.10080
Zhao, H., Shi, J., Qi, X., Wang, X., & Jia, J. (2017). Pyramid scene parsing network. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2881-2890). https://doi.org/10.48550/arXiv.1612.01105
Published
How to Cite
Issue
Section
Copyright (c) 2024 Jiwoo Sim; Bruce Kohl
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Copyright holder(s) granted JSR a perpetual, non-exclusive license to distriute & display this article.